1461186147-13597495-f0d0-4b4d-a705-f1615f1bd762

1. A multimode optical fiber for fiber-optic distributed temperature sensing (DTS), comprising:
a core part; and
a cladding part,
wherein the core part has a structure including a plurality of concentric layers in which a refractive index changes stepwise so as to be higher toward a central area, and a quantity of addition of fluorine to silica glass of the core part is adjusted so as to be entirely increased toward an outer layer,
wherein the number M of the layers of the core part is located within a range of 3\u2266M\u226620 and
wherein the refractive index of the entire core is greater than the refractive index of the cladding layer.
2. The multimode optical fiber according to claim 1, wherein in a refractive index profile of the core part, when the central part and an end part in a radial direction of the core part are allowed to satisfy a below-described expression (1) as an ideal refractive index profile, the refractive index of at least one layer except a central layer of the layers corresponds to the ideal refractive index profile of the expression (1):
n(r)=n1{1\u22122\u0394(ra)\u03b1}\xbd, 2.0\u2266\u03b1\u22662.1\u2003\u2003(1)

wherein n designates a refractive index difference, n1 designates a refractive index of the center of the core, \u0394 designates a relative refractive index difference, a designates a radius of the core, r designates a distance from the center of the core and a designates a coefficient of refractive index profile.
3. The multimode optical fiber according to claim 2, wherein the refractive index of all the layers of the core part correspond to the ideal refractive index profile of the expression (1).
4. The multimode optical fiber according to claim 1, wherein a numerical aperture NA is not lower than 0.18.
5. The multimode optical fiber according to claim 1, wherein a bandwidth is not lower than 200 MHz\xb7km.
6. The multimode optical fiber according to claim 1, wherein the increase of attenuation is not higher than 0.2 dBkm in a wavelength area of a range not lower than 800 nm and not higher than 1700 nm after the elapse of 300 hours in an atmosphere of hydrogen of 2 atm and under an environment of 150\xb0 C.
7. The multimode optical fiber according to claim 6, wherein a bandwidth is not lower than 200 MHz\xb7km.
8. The multimode optical fiber according to claim 6, wherein a numerical aperture NA is not lower than 0.18.

The claims below are in addition to those above.
All refrences to claim(s) which appear below refer to the numbering after this setence.

What is claimed is:

1. A wearable display system comprising:
at least one display panel to display a signal;
a waveguide to transmit the signal that is incident thereon from the display panel, the waveguide having first and second ends; and
at least one magnifying lens attached to one of the ends of the waveguide to magnify the signal transmitted via the waveguide,
wherein at least one of the ends of the waveguide is diagonally cut at a predetermined angle so that the signal displayed by the display panel is totally reflected inside the waveguide.
2. The system of claim 1, further comprising at least one reflection plate to reflect the signal displayed by the display panel to the diagonally cut end of the waveguide.
3. A binocular wearable display system comprising:
a plurality of display panels;
a waveguide to transmit a signal emitted from the display panels, the display panels being respectively positioned at first and second ends of the waveguide; and
a plurality of magnifying lenses attached to the waveguide to magnify the signal transmitted via the waveguide,
wherein the first and second ends of the waveguide are diagonally cut at a predetermined angle so that the signal emitted by the display panels is totally reflected in the waveguide.
4. The system of claim 3, further comprising a plurality of reflection plates respectively extending from the first and second ends of the waveguide to reflect the signal emitted by the display panels such that the emitted signal is transmitted into the waveguide through the first and second ends thereof.
5. A wearable display system comprising:
a display panel to display a signal;
a waveguide to transmit the signal incident thereon from the display panel, the waveguide having first and second ends; and
an optical device attached to a surface of the waveguide, the optical device to magnify the signal transmitted via the waveguide,
wherein the first and second ends of the waveguide are diagonally cut at first and second angles, respectively.
6. The system of claim 5, wherein the display panel is attached to the first end of the waveguide and the optical device is attached to the second end of the waveguide.
7. The system of claim 6, wherein the first angle is equal to an angle of total internal reflection of the signal transmitted via the waveguide.
8. The system of claim 7, wherein the second angle is the same as the first angle.
9. The system of claim 5, wherein the optical device is a reflection type optical device.
10. The system of claim 5, wherein the optical device is a transmission type optical device.
11. A wearable display system comprising:
a display panel to display a signal;
a waveguide to transmit the signal incident thereon from the display panel;
a prism attached to a first end of the waveguide, the prism to transmit the signal displayed by the display panel into the waveguide at an angle such that the transmitted signal undergoes total internal reflection inside the waveguide; and
an optical device to magnify the signal transmitted via the waveguide,
wherein a second end of the waveguide, opposite the first end, is cut at a predetermined angle.
12. The system of claim 11, wherein the optical device is a reflection type optical device.
13. The system of claim 11, wherein the optical device is a transmission type optical device.
14. The system of claim 11, wherein the predetermined angle is equal to the angle of total internal reflection.
15. A wearable display system comprising:
a display panel to display a signal;
a waveguide to transmit the signal incident thereon from the display panel; and
a prism attached to a first end of the waveguide to emit the signal transmitted via the waveguide,
wherein a second end of the waveguide, upon which the signal is incident from the display panel, is cut at a predetermined angle.
16. The system of claim 15, wherein the predetermined angle is equal to an angle of total internal reflection of the signal transmitted via the waveguide.
17. The system of claim 15, further comprising an optical device placed on the prism to magnify the signal emitted by the prism.
18. A wearable display system comprising:
a display panel to display a signal; and
a waveguide to receive the signal from the display panel and propagate the signal therein, the waveguide comprising an end to receive the signal and a side, an angle formed between the end and the side allowing total reflection of the signal within the waveguide.
19. The wearable display system according to claim 18, further comprising a lens to magnify the signal propagated through the waveguide.
20. The wearable display system according to claim 19, wherein the angle formed between the end and the side is the same as an angle of total reflection of the signal inside the waveguide.
21. The wearable display system according to claim 20, further comprising:
a frame to mount the waveguide;
a frame arm extending from the frame;
a light source mounted on the frame arm to emit the light; and
a reflection unit to reflect the emitted light to the waveguide.
22. The wearable display system according to claim 20, wherein the display system displays a virtual image of an object to a user.
23. The wearable display system according to claim 22, wherein a diameter D of the lens is calculated according to:
8
D
=
2

L
e
tan
2
+

E
x
wherein Le is a distance between an eye of the user and the lens, 2 is half of a field of view (FOV) of the display system, and Ex is a size of an exit pupil of the eye of the user.
24. The wearable display system according to claim 23, wherein the field of view is calculated according to:
9
F
O
V
=

2
tan


1
Y
i
2
L
wherein Yi is a size of the virtual image, and
L is a distance between the eye of the user and the virtual image.
25. The wearable display system according to claim 19, wherein:
10
>
c
=
sin


1
1

n
(

w
a
v
e
g
u
i
d
e

)
wherein is an angle of incidence of the signal on the lens, c is a critical angle of total reflection within the waveguide, and n(waveguide) is an index of refraction of the waveguide.
26. A binocular wearable display system comprising:
a plurality of display panels to display a signal; and
a waveguide to receive the signal from the display panels and propagate the signal therein, the waveguide comprising first and second ends to receive the signal and a side, an angle formed between the ends and the side allowing total reflection of the signal within the waveguide.
27. The wearable display system according to claim 26, wherein the signal is light, and the display system further comprises:
a frame to mount the waveguide;
a plurality of frame arms extending from the frame;
a plurality of light sources mounted on the frame arms to emit the light; and
a plurality of reflection units to reflect the emitted light to the waveguide.

1461186137-eff7f140-a7ba-48a3-b2ff-e6d5f464b72e

1. Process for preparing a suspension of a particulate, preferably siliceous, filler in a material formed by a silicone oil comprising:
polyorganosiloxanes (POS fluids) of type (I) which carry Si-alkenylpreferably Si-vinylfunctional groups capable of reacting with the Si-H crosslinking functional groups of a POS fluid of type II,
optionally, POS fluids of type (II) which carry Si-H crosslinking functional groups capable of reacting with the Si-alkenyl functional groups of the POS fluids (I),
andor, optionally, POS fluids of type (III) which differ from the POS fluids (I) and (II),
the said suspension being able to be used, in particular, for producing silicone compositions that can be cured by polyaddition, this process being of the kind of those in which the particulate filler is treated with the aid of a compatibilizing agent or compatibilizer (CA),
characterized in that it essentially consists in introducing some compatibilizer (CA) into the preparation mixture:
on the one hand, before andor substantially simultaneously with the contacting of at least part of the silicone oil employed with at least part of the particulate filler used, this CA introduction taking place in one or more steps for a CA fraction corresponding to a proportion of at most 8%, preferably at most 5% and even more preferably at most 3% by dry weight with respect to the total particulate filler;
and, on the other hand, after this POSfiller contacting.
2. Process according to claim 1, characterized in that at essentially consists:
in mixing:
100 parts by weight of silicone oil
0 to 5 parts by weight of water
20 to 80 parts by weight of particulate filler consisting of silica

1 to 20 parts by weight of compatibilizer (CA) selected from silazanes taken by themselves alone or as a mixture thereof, preferably from disilazanes, hexamethyldisilazane which may or may not be combined with divinyltetramethyldisilazane being particularly preferred;
in leaving the above to react, preferably with stirring,
in heating the mixture obtained, choosing a pressuretemperature pair so that at least some of the water and of the volatile elements undergoes devolatilization;
if necessary, in cooling the mixture.
3. Process according to claim 1 or 2, characterized in that:
all or some of the silicone oil, the water and all or some of the particulate siliceous filler are mixed with a first CA fraction of between 1 and 3% by dry weight with respect to the silica,
a second CA fraction, representing between 10 and 15% by dry weight of silica, is incorporated into the mixture,
optionally, the rest of the silicone oil and the rest of the silica are added,
the mixture is allowed to react, preferably by continuing the mixing,
the mixture is devolatilized,
optionally, the devolatilized mixture is allowed to cool
and, optionally, the suspension is completed with the rest of the silicone oil.
4. Process according to claim 3, characterized in that:
1. a mixture comprising the silicone oil, the water and the first CApreferably HMDZfraction is homogenized,
2. the particulate filler, preferably silica, is gradually added to the mixture obtained at 1,
3. the mixing is continued,
4. the second CApreferably HMDZfraction is gradually incorporated into the mixture obtained at 3,
5. the mixing is continued,
6. the mixture is devolatilized, preferably by heating to a temperature 100 C.
5. Process according to claim 3, characterized in that:
1. the silicone oil and the water are homogenized,
2. the particulate fillerpreferably silicaand, at the same time, the first CApreferably HMDZfraction are gradually incorporated into the mixture obtained at 1,
3. the mixing is continued,
4. the second CApreferably HMDZfraction is gradually incorporated into the mixture obtained at 3,
5. the mixing is continued,
6. the mixture is devolatilized, preferably by heating to a temperature 100 C.
6. Process according to claim 3, characterized in that:
1. the silicone oil is introduced,
2. the particulate fillerpreferably silicatogether with the first CApreferably HMDZfraction and the water are gradually and simultaneously incorporated into the oil,
3. the mixing is continued,
4. the second CApreferably HMDZfraction is gradually incorporated into the mixture obtained at 3,
5. the mixing is continued,
6. the mixture is devolatilized, preferably by heating to a temperature 100 C.
7. Process according to any one of claims 1 to 6, characterized
in that the first CA fraction is replaced, completely or partly, with at least one processing aid chosen from molecules and combinations of molecules:
capable of interacting with the particulate filler, particularly with silicon if a siliceous filler is used, to the detriment of the hydrogen bonds that this particulate filler establishes especially between its own atoms andor with those of the silicone oil,
and capable of being removed from the preparation mixture by devolatilization,

and in that actions are taken to ensure that this processing aid is in the presence of water in the preparation mixture.
8. Process according to claim 7, characterized in that the processing aid is chosen from the group comprising:
silazanes, HMDZ being preferred;
difunctional, or preferably monofunctional, hydroxylated siloxanes;
amines, preferably ammonia andor alkylamines, diethylamine being particularly preferred;
organic acids, formic andor acetic acids being preferred;
and mixtures thereof.
9. Process according to any one of claims 1 to 8, characterized in that an alkenylatedpreferably vinylatedsilicone oil comprising at least two Si-alkenyl groups per molecule, each preferably located at one end of the chain, and having a dynamic viscosity at 25 C. not exceeding 250 Pa.s, preferably not exceeding 100 Pa.s and more preferably still not exceeding 10 Pa.s, is employed,
and in that a silica having a BET specific surface area of between 50 and 400 m2g and mixing conditions such that the dynamic viscosity at 25 C. of the suspension does not exceed 300 Pa.s, preferably does not exceed 250 Pa.s, and more preferably still does not exceed 200 Pa.s, are chosen.
10. Process for obtaining a silicone composition that can be cured by polyaddition, characterized in that it consists in mixing the following products:
Aa suspension as prepared according to the process as defined in any one of claims 1 to 9,
Bone or more POS fluids (I), as defined in claim 1,
Cone or more POS fluids (II), as defined in claim 1,
Doptionally, one or more POS fluids (III), as defined in claim 1, useful as diluent(s),
Ea catalytic system comprising a catalyst, preferably of the platinum kind, and, optionally, an inhibitor.
11. Process according to claim 10, characterized
in that the composition is produced in the form of a two-component system comprising parts C1 and C2 which are intended to be brought into contact with each other in order to produce an elastomer crosslinked by polyaddition between the POS fluids (I) and (II)
and in that care is taken to ensure that only one of the parts, C1 or C2, contains some catalyst D and, optionally, one or other of the POS fluids (I) and (II).

The claims below are in addition to those above.
All refrences to claim(s) which appear below refer to the numbering after this setence.

What is claimed is:

1. An image processing apparatus comprising:
an input unit that acquires a RGB signal corresponding to a color image;
a conversion unit that converts the RGB signal into a CMY signal;
an extraction unit that extracts an image attribute from the CMY signal; and
a processing unit that applies, based on the image attribute, an adaptive image processing to the RGB signal.
2. The image processing apparatus according to claim 1, wherein the extraction unit calculates an edge amount of the color image as the image attribute.
3. The image processing apparatus according to claim 1, wherein the extraction unit generates an image area separating signal that is used to separate an image into a plurality of areas as the image attribute.
4. The image processing apparatus according to claim 1, wherein the conversion unit changes a conversion coefficient for converting the RGB signal into the CMY signal based on a type of the color image.
5. The image processing apparatus according to claim 4, wherein the type of the color image is any one of a print image, a photographic printing paper image, and a photocopy image.
6. An image processing apparatus comprising:
an input unit that acquires a RGB signal corresponding to a color image;
a first conversion unit that converts the RGB signal into a CMY signal;
an extraction unit that extracts an image attribute from the CMY signal;
a second conversion unit that generates a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal; and
a processing unit that applies, based on the image attribute, an adaptive image processing to the signal generated by the second conversion unit.
7. The image processing apparatus according to claim 6, wherein the extraction unit calculates an edge amount of the color image as the image attribute.
8. The image processing apparatus according to claim 6, wherein the extraction unit generates an image area separating signal that is used to separate an image into a plurality of areas as the image attribute.
9. The image processing apparatus according to claim 6, wherein the first conversion unit changes a conversion coefficient for converting the RGB signal into the CMY based on a type of the color image.
10. The image processing apparatus according to claim 9, wherein the type of the color image is any one of a print image, a photographic printing paper image, and a photocopy image.
11. An image processing apparatus comprising:
an input unit that acquires a RGB signal corresponding to a color image;
a first extraction unit that extracts a first image attribute from the RGB signal;
a conversion unit that converts the RGB signal into a CMY signal;
a second extraction unit that extracts a second image attribute from the CMY signal; and
a processing unit that applies, based on the first image attribute and the second image attribute, an adaptive image processing to the RGB signal.
12. The image processing apparatus according to claim 11, wherein
the first extraction unit generates an image area separating signal that is used to separate an image into a plurality of areas as the first image attribute, and
the second extraction unit calculates an edge amount of the color image as the second image attribute.
13. The image processing apparatus according to claim 12, wherein the second extraction unit calculates the edge amount from a C signal and an M signal of the CMY signal as the second image attribute.
14. The image processing apparatus according to claim 11, wherein the conversion unit changes a conversion coefficient for converting the RGB signal into the CMY signal based on a type of the color image.
15. The image processing apparatus according to claim 14, wherein the type of the color image is any one of a print image, a photographic printing paper image, and a photocopy image.
16. An image processing apparatus comprising:
an input unit that acquires a RGB signal corresponding to a color image;
a first extraction unit that extracts a first image attribute from the RGB signal;
a first conversion unit that converts the RGB signal into a CMY signal;
a second extraction unit that extracts a second image attribute from the CMY signal;
a second conversion unit that generates a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal; and
a processing unit that applies, based on the first image attribute and the second image attribute, an adaptive image processing to the signal generated by the second conversion unit.
17. The image processing apparatus according to claim 16, wherein
the first extraction unit generates an image area separating signal that is used to separate an image into a plurality of areas as the first image attribute, and
the second extraction unit calculates an edge amount of the color image as the second image attribute.
18. The image processing apparatus according to claim 17, wherein the second extraction unit calculates the edge amount from a C signal and an M signal of the CMY signal as the second image attribute.
19. The image processing apparatus according to claim 16, wherein the first conversion unit changes a conversion coefficient for converting the RGB signal into the CMY based on a type of the color image.
20. The image processing apparatus according to claim 19, wherein the type of the color image is any one of a print image, a photographic printing paper image, and a photocopy image.
21. An image processing apparatus comprising:
an input unit that acquires a RGB signal corresponding to a color image;
a first conversion unit that converts the RGB signal into a CMY signal;
a first extraction unit that extracts a first image attribute from the CMY signal;
a second conversion unit that generates a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
a second extraction unit that extracts a second image attribute from the signal generated by the second conversion unit; and
a processing unit that applies, based on the first image attribute and the second image attribute, an adaptive image processing to the RGB signal.
22. The image processing apparatus according to claim 21, wherein
the first extraction unit calculates an edge amount of the color image as the first image attribute, and
the second extraction unit generates an image area separating signal that is used to separate an image into a plurality of areas as the second image attribute.
23. The image processing apparatus according to claim 22, wherein the first extraction unit calculates the edge amount from a C signal and an M signal of the CMY signal as the second image attribute.
24. The image processing apparatus according to claim 21, wherein the first conversion unit changes a conversion coefficient for converting the RGB signal into the CMY signal based on a type of the color image.
25. The image processing apparatus according to claim 24, wherein the type of the color image is any one of a print image, a photographic printing paper image, and a photocopy image.
26. An image processing apparatus comprising:
an input unit that acquires a RGB signal corresponding to a color image;
a first conversion unit that converts the RGB signal into a CMY signal;
a first extraction unit that extracts a first image attribute from the CMY signal;
a second conversion unit that generates a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
a second extraction unit that extracts a second image attribute from the signal generated by the second conversion unit; and
a processing unit that applies, based on the first image attribute and the second image attribute, an adaptive image processing to the signal generated by the second conversion unit.
27. The image processing apparatus according to claim 26, wherein
the first extraction unit calculates an edge amount of the color image as the first image attribute, and
the second extraction unit generates an image area separating signal that is used to separate an image into a plurality of areas as the second image attribute.
28. The image processing apparatus according to claim 27, wherein the first extraction unit calculates the edge amount from a C signal and an M signal of the CMY signal as the second image attribute.
29. The image processing apparatus according to claim 26, wherein the first conversion unit changes a conversion coefficient for converting the RGB signal into the CMY signal based on a type of the color image.
30. The image processing apparatus according to claim 29, wherein the type of the color image is any one of a print image, a photographic printing paper image, and a photocopy image.
31. An image processing method comprising:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting an image attribute from the CMY signal; and
applying, based on the image attribute, an adaptive image processing to the RGB signal.
32. An image processing method comprising:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting an image attribute from the CMY signal; and
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
applying, based on the image attribute, an adaptive image processing to the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal.
33. An image processing method comprising:
acquiring a RGB signal corresponding to a color image;
extracting a first image attribute from the RGB signal;
converting the RGB signal into a CMY signal;
extracting a second image attribute from the CMY signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the RGB signal.
34. An image processing method comprising:
acquiring a RGB signal corresponding to a color image;
extracting a first image attribute from the RGB signal;
converting the RGB signal into a CMY signal;
extracting a second image attribute from the CMY signal;
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal.
35. An image processing method comprising:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting a first image attribute from the CMY signal;
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
extracting a second image attribute from the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the RGB signal.
36. An image processing method comprising:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting a first image attribute from the CMY signal;
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
extracting a second image attribute from the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal.
37. A computer product that makes a computer execute:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting an image attribute from the CMY signal; and
applying, based on the image attribute, an adaptive image processing to the RGB signal.
38. A computer product that makes a computer execute:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting an image attribute from the CMY signal; and
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
applying, based on the image attribute, an adaptive image processing to the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal.
39. A computer product that makes a computer execute:
acquiring a RGB signal corresponding to a color image;
extracting a first image attribute from the RGB signal;
converting the RGB signal into a CMY signal;
extracting a second image attribute from the CMY signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the RGB signal.
40. A computer product that makes a computer execute:
acquiring a RGB signal corresponding to a color image;
extracting a first image attribute from the RGB signal;
converting the RGB signal into a CMY signal;
extracting a second image attribute from the CMY signal;
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal.
41. A computer product that makes a computer execute:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting a first image attribute from the CMY signal;
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
extracting a second image attribute from the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the RGB signal.
42. A computer product that makes a computer execute:
acquiring a RGB signal corresponding to a color image;
converting the RGB signal into a CMY signal;
extracting a first image attribute from the CMY signal;
generating a signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal from the RGB signal;
extracting a second image attribute from the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal; and
applying, based on the first image attribute and the second image attribute, an adaptive image processing to the signal including either of a luminancechrominance difference signal and a lightnesschromaticity signal.