• Aucun résultat trouvé

Exact Identity Mapping: The Orthoscopic Reality Mediator It is easy to imagine a camera connected to a television screen, and carefully

Dans le document Intelligent Image Processing (Page 88-92)

EFFECTIVELY LOCATING THE CAMERA INSIDE THE EYE

3.3 PRACTICAL EMBODIMENTS OF EYETAP

3.3.3 Exact Identity Mapping: The Orthoscopic Reality Mediator It is easy to imagine a camera connected to a television screen, and carefully

arranged in such a way that the television screen displays exactly what is blocked by the screen so that an illusory transparency results. Moreover it is easy to imagine a portable miniature device that accomplishes this situation, especially given the proliferation of consumer camcorder systems (e.g., portable cameras with built in displays), see Figure 3.7.

We may try to achieve the condition shown in Figure 3.7 with a handheld camcorder, perhaps miniaturized to fit into a helmet-mounted apparatus, but it is impossible to line up the images exactly with what would appear in the absence of the apparatus. We can better understand this problem by referring to Figure 3.8.

In Figure 3.8 we imagine that the objective lens of the camera is much larger than

Hitachi

Figure 3.7 A modern camcorder (denoted by the reference numeral 10 in the figure) could, in principle, have its zoom setting set for unity magnification. Distant objects 23 appear to the eye to be identical in size and position while one looks through the camcorder as they would in the absence of the camcorder. However, nearby subject matter 23 N will be distancede, which is closer to the effective center of projection of the camcorder than distancedeto the effective center of projection of the eye. The eye is denoted by reference numeral 39, while the camera iris denoted 22i defines the center of projection of the camera lens 22. For distant subject matter the difference in location between iris 22i and eye 39 is negligible, but for nearby subject matter it is not. Therefore nearby subject matter will be magnified as denoted by the dotted line figure having reference numeral 23 F. Alternatively, setting the camcorder zoom for unity magnification for nearby subject matter will result in significantly less than unity magnification for distant subject matter. Thus there is no zoom setting that will make both near and far subject matter simultaneously appear as it would in the absence of the camcorder.

it really is. It captures all eyeward bound rays of light, for which we can imagine that it processes these rays in a collinear fashion. However, this reasoning is pure fiction, and breaks down as soon as we consider the scene that has some depth of field, such as is shown in Figure 3.9.

Thus we may regard the apparatus consisting of a camera and display as being modeled by a fictionally large camera opening, but only over subject matter confined to a plane.

Even if the lens of the camera has sufficient depth of focus to form an image of subject matter at various depths, this collinearity criterion will only hold at one such depth, as shown in Figure 3.10. This same argument may be made for the camera being off-axis. Thus, when the subject matter is confined to a single plane, the illusory transparency can be sustained even when the camera is off-axis, as shown in Figure 3.11.

Some real-world examples are shown in Figure 3.12. An important limitation is that the system obviously only works for a particular viewpoint and for

39

Figure 3.8 Suppose that the camera portion of the camcorder, denoted by reference numeral 10C, were fitted with a very large objective lens 22F. This lens would collect eyeward bound rays of light 1E and 2E. It would also collect rays of light coming toward the center of projection of lens 22. Rays of light coming toward this camera center of projection are denoted 1C and 2C. Lens 22 converges rays 1E and 1C to point 24A on the camera sensor element. Likewise rays of light 2C and 2E are focused to point 24B. Ordinarily the image (denoted by reference numeral 24) is upside down in a camera, but cameras and displays are designed so that when the signal from a camera is fed to a display (e.g., a TV set) it shows rightside up. Thus the image appears with point 32A of the display creating rays of light such as denoted 1D. Ray 1D is collinear with eyeward bound ray 1E. Ray 1D is response to, and collinear with ray 1E that would have entered the eye in the absence of the apparatus. Likewise, by similar reasoning, ray 2D is responsive to, and collinear with, eyeward bound ray 2E. It should be noted, however, that the large lens 22F is just an element of fiction. Thus lens 22F is a fictional lens because a true lens should be represented by its center of projection; that is, its behavior should not change other than by depth of focus, diffraction, and amount of light passed when its iris is opened or closed. Therefore we could replace lens 22F with a pinhole lens and simply imagine lens 22 to have captured rays 1E and 2E, when it actually only captures rays 1C and 2C.

subject matter in a particular depth plane. This same setup could obviously be miniaturized and concealed in ordinary looking sunglasses, in which case the limitation to a particular viewpoint is not a problem (since the sunglasses could be anchored to a fixed viewpoint with respect to at least one eye of a user).

However, the other important limitation, that the system only works for subject matter in the same depth plane, remains.

39

Figure 3.9 The small lens 22 shown in solid lines collects rays of light 1C and 2C. Consider, for example, eyeward bound ray of light 1E, which may be imagined to be collected by a large fictional lens 22F (when in fact ray 1C is captured by the actual lens 22), and focused to point 24A. The sensor element collecting light at point 24A is displayed as point 32A on the camcorder viewfinder, which is then viewed by magnifying lens and emerges as ray 1D into eye 39. It should be noted that the top of nearby subject matter 23N also images to point 24A and is displayed at point 32A, emerging as ray 1D as well. Thus nearby subject matter 23N will appear as shown in the dotted line denoted 23F, with the top point appearing as 23FA even though the actual point should appear as 23NA (e.g., would appear as point 23NA in the absence of the apparatus).

Figure 3.10 Camera 10C may therefore be regarded as having a large fictional lens 22F, despite the actual much smaller lens 22, so long as we limit our consideration to a single depth plane and exclude from consideration subject matter 23N not in that same depth plane.

39

Figure 3.11 Subject matter confined to a single plane 23 may be collinearly imaged and displayed by using the same large fictional lens model. Imagine therefore that fictional lens 22F captures eyeward bound rays such as 1E and 2E when in fact rays 1C and 2C are captured.

These rays are then samplings of fictional rays 1F and 2F that are resynthesized by the display (shown here as a television receiver) that produces rays 1D and 2D. Consider, for example, ray 1C, which forms an image at point 24A in the camera denoted as 10C. The image, transmitted by transmitter 40T, is received as 40R and displayed as pixel 32A on the television. Therefore, although this point is responsive to light along ray 1C, we can pretend that it was responsive to light along ray 1E. So the collinearity criterion is modeled by a fictionally large lens 22F.

Obviously subject matter moved closer to the apparatus will show as being not properly lined up. Clearly, a person standing right in front of the camera will not be behind the television yet will appear on the television. Likewise a person standing directly behind the television will not be seen by the camera which is located to the left of the television. Thus subject matter that exists at a variety of different depths, and not confined to a plane, may be impossible to line up in all areas, with its image on the screen. See, for example, Figure 3.13.

Dans le document Intelligent Image Processing (Page 88-92)

Outline

Documents relatifs