US20080316301A1 - System and method for spherical stereoscopic photographing - Google Patents

System and method for spherical stereoscopic photographing Download PDF

Info

Publication number
US20080316301A1
US20080316301A1 US12/107,048 US10704808A US2008316301A1 US 20080316301 A1 US20080316301 A1 US 20080316301A1 US 10704808 A US10704808 A US 10704808A US 2008316301 A1 US2008316301 A1 US 2008316301A1
Authority
US
United States
Prior art keywords
image
images
stereoscopic
viewer
vision
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/107,048
Inventor
Dor Givon
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Micoy Corp
Original Assignee
Micoy Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Micoy Corp filed Critical Micoy Corp
Priority to US12/107,048 priority Critical patent/US20080316301A1/en
Publication of US20080316301A1 publication Critical patent/US20080316301A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B35/00Stereoscopic photography
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/282Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • H04N13/334Displays for viewing with the aid of special glasses or head-mounted displays [HMD] using spectral multiplexing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • H04N13/344Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays

Definitions

  • the present invention relates to stereoscopic photographing system. More specifically, the present invention relates to a photographing system and method for generating stereoscopic images of a full sphere surrounding a viewer 360 degrees both horizontally and vertically, or of a continuous part thereof, comprising a spherical array of lenses having non-parallel optical axes.
  • Stereoscopic and panoramic photographing systems are known in the art.
  • stereo pairs are taken by twin-cameras having parallel optical axes and a fixed distance between their aligned lenses.
  • the pair of the obtained images can then be displayed by any of the known techniques for stereoscopic displaying and viewing. All of these techniques are based, in general, on the principle that the image taken by the right lens is displayed to the viewer's right eye and the image taken by the left lens is displayed to the viewer's left eye.
  • Panoramic photographing is conventionally done either by a very wide-angle lens, such as a “fish-eye” lens, or by “stitching” together slightly overlapping adjacent images to cover a wide, up to a fill circle, field of vision.
  • a very wide-angle lens such as a “fish-eye” lens
  • stitching together slightly overlapping adjacent images to cover a wide, up to a fill circle, field of vision.
  • the same techniques used for panoramic imaging are also exploited for obtaining spherical images.
  • the panoramic or spherical images obtained by using said techniques are not stereoscopic, nor do they give to the viewer a perception of depth.
  • the present invention is based on the finding that a stereoscopic image can be obtained by overlapping images taken by lenses having non-parallel optical axes, providing that said lenses' fields of view are overlapping to a great extent.
  • This finding enables a full spherical stereoscopic photographing by a three dimensional array of cameras pointing out from a common center for covering the whole surroundings, both horizontally and vertically.
  • the present invention provides a novel imaging system for obtaining full stereoscopic spherical images of the visual environment surrounding a viewer, 360 degree both horizontally and vertically. Displaying the images obtained by the present system, by means suitable for stereoscopic displaying, gives the viewers the ability to look everywhere around them, as well as up and down, while having stereoscopic depth perception of the displayed images.
  • One aspect of the present invention is a photographing system for capturing a full stereoscopic spherical image, or a continuous part thereof, comprising a three dimensional array of cameras.
  • the cameras are positioned at an equal distance from a common center and pointing out from said common center, in such a manner that all lenses are situated on a spherical surface.
  • the collective field of vision captured by all of the lenses covers the whole visual scale surrounding said photographing system, or a continuous part thereof, and any point in said collective field of vision is captured by at least two of said lenses.
  • the cameras are arranged in such a way that the field of vision of any of the lenses is overlapped to a great extent by the fields of view of the two adjacent lenses lying on said lens sides, with respect to horizontal axis, and only to a small extent by the fields of vision of adjacent lenses lying above or below said lens with respect to the horizontal axis.
  • the collective field of vision comprises a collection of, fully circular, stereoscopic panoramic images wherein any point within each of said panoramic images is captured by at least two lenses.
  • the cameras are arranged in the specific configuration, wherein the field of vision of any of said lenses is overlapped to a great extent by the fields of vision of all adjacent lenses surrounding said lens and wherein said lenses are optionally equally distributed on the spherical surface.
  • the image information captured by each of said cameras is stored separately.
  • the present invention further comprises a marking means for marking the boundaries of any of the images captured by any of said lenses, or a set of predetermined points within the area of each of said images and a means for image processing for processing the image information captured by any of said cameras and for cropping out redundant data in order to enhance said image quality.
  • a second aspect of the present invention is a multi-lens camera apparatus for generating a full stereoscopic spherical imaging.
  • the apparatus comprises a set of lenses distributed on a spherical surface and a means for recording and optionally storing the image information of each of the images captured by each of said lenses.
  • the lenses are distributed on the spherical surface, wherein the collective field of vision captured by all of said lenses, covers the whole visual environment surrounding the camera apparatus, and any point in said collective field of vision is captured by at least two of said lenses.
  • a third aspect of the present invention is a process for capturing full or partial stereoscopic spherical image of a visual environment to be displayed to a viewer by means allowing for stereoscopic viewing.
  • the process comprises arranging cameras in a three dimensional array, wherein the lenses of all said cameras are situated on a spherical surface and the optical axes of all said lenses meet at the center of said spherical sphere, wherein the collective field of vision captured by all said lenses covers the whole visual environment surrounding array of cameras, or a continuous part thereof, and wherein any point in said collective field of vision is captured by at least two of said lenses.
  • a fourth aspect of the present invention is a process for capturing and displaying a full stereoscopic spherical image of a visual environment, to be displayed to a viewer by a system having means for stereoscopic displaying and viewing.
  • the process comprises: a) arranging a three dimensional array of lenses for capturing a collection of images, wherein the collective field of vision captured by all said lenses covers the whole visual environment surrounding said array of lenses, and wherein any point in said collective field of vision is captured by at least two of said lenses; b) receiving from said system for stereoscopic displaying and viewing, the coordinates of said viewer's field of vision; c) creating a stereoscopic image pair from said collection of images, said stereoscopic image pair covers said viewer's field of vision; d) displaying said stereoscopic image pair to said viewer by means of said system for stereoscopic displaying and viewing.
  • the step of creating a stereoscopic image pair from said collection of images comprises the following sub-steps: a) projecting said viewer's field of vision onto said collection of images; b) selecting from said collection of images a selected group of images overlapping at least partially said viewer's field of vision; c) dividing each of the images in said selected group of images into a left part and a right part by a line perpendicular to viewer's horizon and passing through image center to obtain a group of left image parts and a group of right image parts; d) selecting from said group of left image parts the parts overlapping said viewer's field of vision to obtain a group of selected left parts; e) merging said selected left parts into a uniform two-dimensional first image, said first image is the right image of said stereoscopic pair; f) selecting from said group of right image parts those parts overlapping said viewer's field of vision to obtain a group of selected right parts; g) merging said selected right parts into a uniform two-dimensional second image, said second image is
  • FIGS. 1A and 1B are two perspective views showing the configuration of the lenses on a sphere in accordance with the present invention
  • FIG. 2 is a horizontal cross section plane through the sphere center showing the cameras positions for generating one strap of a full stereoscopic panoramic view;
  • FIG. 3 schematically illustrates the overlapping between fields of view of adjacent cameras for generating a stereoscopic image
  • FIG. 4 is a flow chart describing the main steps for recording image information from each of the lenses comprising the system
  • FIG. 5 is a block diagram of the present invention.
  • FIG. 6 is a flow chart describing the steps for generating a stereoscopic pair in accordance with the viewer's viewing parameters
  • FIG. 7 pictorially describes an example of the process described in FIG. 6 .
  • the present invention provides a system and method for simultaneous capturing of a full spherical stereoscopic imaging or a continuous part thereof.
  • the method of the present invention for obtaining stereoscopic imaging is based on optical principles, and therefore can be adapted to any photographing technique.
  • the preferred embodiment of the present invention comprises digital video cameras, the present system can be adapted as well to analog video cameras and to film cameras.
  • the present system can generate still images or a time sequence of images for motion imaging.
  • the apparatus of the present invention comprises a three dimensional array of cameras, pointing out from common centers, wherein said cameras' principal optical axes meet at said centers.
  • the lenses are situated on a curved surface, allowing for capturing images of the surrounding environment.
  • the cameras are positioned at an equal distance form a common center.
  • any curved surface or object e.g., ellipse
  • FIGS. 1A and 1B are two exterior perspective views showing the position of lenses (designated 1 ) on the so formed sphere.
  • FIG. 2 is a cross section plane through the sphere center showing eight cameras (designated 5 ) equally spaced and radially pointing outward from the center, wherein said cameras' principal optical axis (designating 15 ) meet at the sphere's center.
  • the collection of images captured by lenses 1 covers a fully spherical visual surrounding the camera system, 360 degrees both horizontally and vertically
  • the actual installation of the cameras in order to form such fixed configuration can be obtained by various ways.
  • the cameras are mounted in designated gaps of a spherical object.
  • the cameras can be mounted in a honeycomb net structure, etc.
  • the apparatus may be mounted on a supporting device 8 having means for enabling rotation and translation of the camera system.
  • a supporting device 8 having means for enabling rotation and translation of the camera system.
  • a support can be a tripod, having rolling means for allowing track-in/track-out and track-right/track-left translation and further having means for tilt and pan and for crane-up/crane-down.
  • the apparatus further includes a stabilization system for preventing the system from vibrations and for keeping the cameras at their fixed positions relative to each other.
  • the apparatus may further include internal robotics for allowing rotation, panning and tilting of the sphere, independently of support 8 .
  • the total number of lenses distributed on the sphere is eighteen.
  • the lenses arrangement can be described as follows: eight lenses are situated around the sphere equator for capturing a full 360 degree panoramic view, four lenses are situated around higher/lower latitude on each side of the equator for capturing the view above and below the horizon, and a lens at each pole to complete a full spherical view.
  • the number of lenses is between 18 to 22.
  • the number of cameras can be varied, depending on the desired quality of the image to be obtained, as will be explained below.
  • FIG. 3 shows three adjacent cameras, 5 , 5 ′ and 5 ′′ and their corresponding fields of vision, 11 , 12 and 13 .
  • a main idea of the present invention is to create a cameras' arrangement wherein any point in the collective field of vision is captured by at least two lenses of the cameras, thus providing stereoscopic sphere image.
  • any point is covered by the fields of vision of at least two lenses.
  • the left half of field 12 is overlapped by field 11 while its right half is overlapped by field 13 , thus any point in field 12 is covered by at least one more lens.
  • Distance d is a function of the angle ⁇ between the optical axes of adjacent lenses, the angle of coverage ⁇ of each lens and the sphere radius r.
  • the desired quality of the captured images Another factor to be considered is the desired quality of the captured images.
  • the best image quality is obtained by the central portion of a lens coverage area and as the angle of coverage of a lens increases, so does the relative range of distorted borders. Therefore, the number of the required lenses of the apparatus is a function of the desired quality.
  • a larger number of lenses, each having a smaller field of vision should be used.
  • the angle of coverage of the lenses is in the range of 30 to 65 degrees (lenses of 35 to 70 mm).
  • the lenses are identical, having the same angle of coverage and the same focal length, but it is also possible to have lenses of different angle of coverage wherein lenses of wider angles are used for areas not requiring high image quality.
  • the cameras can be arranged in two basic configurations, resulting two basic distributions of lenses on the sphere.
  • a first configuration the sphere is divided into parallel straps, equivalent to circumferential latitude straps around the globe.
  • the lenses in each such strap are arranged wherein any point in collective field of vision, captured by the lenses in that specific strap, is covered by at least two lenses, but there is only small overlapping between adjacent straps.
  • the number of lenses covering each strap decreases.
  • Additional two lenses at each pole give an upward and downward view in order to complete the fall spherical image.
  • Such a configuration facilitates generating a spherical stereoscopic imaging as if conventional tilt and pan displaying systems are used.
  • the assumption is that the viewer's horizon is substantially parallel with respect to ground, i.e., that his eyes are substantially at the same height with respect to ground.
  • the viewer's horizon is defined as the vector connecting the viewer eyes.
  • the cameras are configured with no preference to latitude or longitude coordinates.
  • this “mosaic like” configuration the field of vision of each lens overlaps to a great extent the fields of view of all adjacent lenses wherein the lenses are optionally equally distributed on the spherical surface.
  • This configuration provides that any point in the collective field of vision is captured by lenses located around said point. The collection of images captured by this configuration facilitates stereoscopic perception from all viewing points, also when the viewer horizon is inclined with respect to ground.
  • the present invention is not limited to the type of the cameras used in the system.
  • the cameras are digital video cameras having an electro-optical sensor such as a CCD sensor.
  • the cameras can be selected from any commercially available digital video cameras
  • the encoded digital image information, captured by each of the lenses is transmitted via communication means to a remote storage device wherein each image is stored separately.
  • the use of digital cameras allows for enhanced manipulation of the captured images and facilitates controlling the image quality prior to storage.
  • the camera system further includes an internal computerized system comprising a suitable software for controlling the internal robotics responsible for the sphere motion and stabilization.
  • the computerized system may further include a central processing unit, a storage device, a memory device and algorithms for processing the captured images prior to storage. Alternatively, the image processing can be done at a remote computer.
  • the digital image information captured by each camera is processed separately by the internal computer to remove distortion from each image separately.
  • the images are cropped into pre-defined sections which are required for keeping an adequate overlap between adjacent images, wherein the image edges are eliminated for sparing the storage space of unnecessary information.
  • the image enhancement may further include other known methods such as stretching and/or compressing certain parts of the image in order to correct distortions and to enhance the image quality.
  • the image processing may also include processing methods as those employed by standard camera control units (CCU) in order to balance the image and achieve uniformity between adjacent images.
  • CCU camera control units
  • the camera system further includes a marking means for marking the borders of the lenses or a set of predetermined points in the rims and/or area of each of the lenses.
  • the marking means can comprise one or more laser sources, such as a small diode laser, or other small radiation sources for generating beams of visible or invisible light. Said beams scan the borders of the lenses or a predetermined set of points in the area of the lenses to form a set of imprinted markers in the captured images.
  • Said set of imprinted markers is identified and used by the computerized system for accurate positioning of the images relative to each other. Although the relative positions of the cameras are fixed, small shifts might occur during operation.
  • the markers system eliminates possible misalignment of the images due to such small shifts.
  • the marked borders or points are used to facilitate image processing in accordance with known image processing methods in order to remove distortion and to enhance the image quality.
  • the present invention is not limited to an array of commercially available cameras but can be implemented by a multi-lens camera apparatus designed in accordance with the spirit of the present invention.
  • a multi-lens camera apparatus comprising a set of lenses distributed on a spherical surface and corresponding sensor means for recording the image information captured by each of said lenses, can be built from basic camera elements.
  • the sensor means can comprise a set of electro-optical sensors such as CCD chips, each corresponding to one lens, for receiving and converting the images captured by the lenses into encoded digital information
  • the number of sensors can be smaller than the number of lenses, wherein along with appropriate optical elements, each sensor receives image information from more than one lens.
  • the sensor means can comprise only one electro-optical sensor wherein the camera apparatus includes suitable optical elements for projecting the images formed by each lens onto said one sensor.
  • the sensor can be designed to be large enough to capture all images simultaneously.
  • all images can be sequentially projected onto the same sensor area wherein a special mechanism controls a system of shutters for switching between lenses. (Similar mechanism can project the image on celluloid film mechanism).
  • the image (or sequence of images) information captured by each of lenses 1 in step 510 is optionally processed in step 520 by image processor 30 in order to enhance the image quality.
  • the images are recorded by recording system 40 and are optionally stored in a suitable storage medium.
  • the image information (whether analog or digital, whether one image or sequence of images) from each of cameras 5 is recorded and stored separately with a lens indicator and a time indicator.
  • the lens indicator includes information about the lens characteristics and position in terms of the sphere coordinates.
  • the image processing in step 520 is optional. Depending on the optical characteristics of the lenses employed and on the desired quality of the image, various image processing techniques if any, may be employed. These techniques may vary from a simple cropping out of the image borders in order to completely remove distorted areas, to more sophisticated image processing methods such as stretching and/or compressing certain parts of the image in order to correct distortions and to enhance the image quality.
  • the image processing may also include processing methods as those employed by standard camera control units (CCU) in order to balance the image and achieve uniformity between adjacent images.
  • CCU camera control units
  • the recorded images are displayed to a viewer via viewing system 50 employing known virtual reality equipment 60 .
  • viewing system 50 receives image information in real time, directly from the camera system.
  • the image processing is an optional step. As explained above, the image processing is performed on each of the images separately in order to maintain higher image quality and with no relation to the stereoscopic effect.
  • the stereoscopic effect according to the present invention, is based solely on optical principles and not requiring any image processing for that purpose, as explained down bellow.
  • FIG. 6 is a flow chart describing the steps for generating a pair of stereoscopic images from the collection of recorded images in order to be displayed to the viewer according to his viewing parameters.
  • the viewer's viewing parameters are detected by employing known means in the art of virtual reality, designated 60 .
  • Such means can be a headset having sensors to detect the viewer's head position or a virtual glove having a sensor to detect the hand position.
  • a pointing device for example: a mouse or a joystick
  • step 610 when a viewer selects a specific view, either by actually turning his head while wearing a virtual reality headset, or by a pointing device coupled to a computer device, the viewing parameters are detected and received by the displaying system.
  • the viewer's viewing parameters include the viewer's viewing direction and viewer's horizon.
  • the viewer's field of vision is determined in terms of the coordinates of the sphere surrounding the viewer and is projected into the collection of stored images.
  • all the images that overlap at least partially the viewer's field of vision are selected from the collection of recorded images.
  • Steps 640 - 680 illustrates the images organization algorithms based on the optical principles of the present invention for creating a pair of merged images wherein said pair of merged images creates the perception of three dimensional panoramic images.
  • This algorithm can be alternatively applied for images captured by a single (or more) cameras wherein the images are taken from different angles and/or location of the same cameras. Such alternative is preferably applied for creating spherical stereoscopic still image.
  • a stereoscopic image pair is created from the group of selected images in the following way: Each of the selected images is divided into a left part and a right part according to the viewer's horizon by a line which is perpendicular the viewer horizon and is passing through the center of the image (step 640 ). In step 650 all the left parts generated in step 640 which are included in the viewer's field of vision are merged into a one uniform two dimensional image that matches the viewer's field of vision. The formed image, is the right image of the stereoscopic pair to be displayed to the viewer's right eye (step 660 ). Following the same lines, a left image is formed by merging together the right parts generated in step 640 (steps 670 , 680 and 690 ).
  • FIG. 7 pictorially describes one example of generating a stereoscopic pair according to the present invention.
  • Frame 100 represents the viewer's field of vision as received by viewing system 60 .
  • the illustrated frame has rectangular shape, however the same stereoscopic principles can be applied to circular frames depending on the cameras technological capabilities.
  • Frames 31 , 32 and 33 represent three adjacent images, taken by the lenses (as viewed from underneath) of the camera system 10 , which overlap viewer's field of vision 100 .
  • the middle frame 32 covers frame 100 completely while frames 31 and 33 partly overlap the left and the right sides of frame 100 , respectively.
  • a new image 40 to be displayed to the viewer's right eye, is formed from images 32 and 33 by merging a left portion of image 32 with a left portion of image 33 .
  • Left image 41 is formed similarly by merging a right portion of image 32 with a right portion of image 31 to form a uniform image.
  • New images 40 and 41 both cover the viewer's field of vision but obviously they are not identical, and are perceived as were taken from two viewpoints, thus giving the viewer a stereoscopic perception. Roughly, this new image pair is equivalent to a pair of images as if were taken by two virtual lenses having their optical axes directed forward in the viewer's viewing direction and having horizontal disparity.
  • the process illustrated in FIG. 7 is for a case where the viewer's field of vision overlaps three recorded images.
  • the viewer's field of vision includes more recorded images (e.g., when it includes upper and lower parts of images from different latitudes or when each image covers only a small part of the viewer's field of vision)
  • the stereoscopic pair is formed along the same lines, by merging together the appropriate parts from each of the recorded images and by removing redundant information.
  • the selection of the recorded images is done by projecting the viewer's field of vision on the collection of recorded images and selecting those images that form with each other the closest angle to the inclining angle.
  • the stereoscopic pair is generated by following steps 640 to 680 of FIG. 6 .
  • the stereoscopic image pair thus formed can be displayed on a flat screen such as a TV or a computer screen or by using a display device for virtual reality such as a virtual reality headset.
  • a flat screen such as a TV or a computer screen
  • the images are displayed as a stereoscopic pair to be viewed by suitable spectacles in accordance with the standard color filtering method for stereoscopic viewing or by any other known in the art techniques for stereoscopic displaying:
  • the part of the image being displayed changes according to the viewer viewpoint as explained above.

Abstract

The present invention provides a novel imaging system for obtaining full stereoscopic spherical images of the visual environment surrounding a viewer, 360 degrees both horizontally and vertically. Displaying the images obtained by the present system, by means suitable for stereoscopic displaying, gives the viewers the ability to look everywhere around them, as well as up and down, while having stereoscopic depth perception of the displayed images. The system according to the present invention comprises an array of cameras, wherein the lenses of said cameras are situated on a curved surface, pointing out from common centers of said curved surface. The captured images of said system are arranged and processed to create sets of stereoscopic image pairs, wherein one image of each pair is designated for the observer's right eye and the second image for his left eye, thus creating a three dimensional perception.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This invention is a division of U.S. patent application Ser. No. 10/416,533 filed May 27, 2003, titled “System and method for spherical stereoscopic photographing”; which is a U.S. National Stage Application under U.S.C. 371 of PCT International Application No. PCT/IL01/01093, which has an international filing date of Nov. 28, 2001; and which claims priority from Israel Patent Application No. IL 139995, filed Nov. 29, 2000; each of which are incorporated herein by reference in their entirety.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to stereoscopic photographing system. More specifically, the present invention relates to a photographing system and method for generating stereoscopic images of a full sphere surrounding a viewer 360 degrees both horizontally and vertically, or of a continuous part thereof, comprising a spherical array of lenses having non-parallel optical axes.
  • 2. Discussion of the Related Art
  • Stereoscopic and panoramic photographing systems are known in the art. In conventional stereoscopic photographing, stereo pairs are taken by twin-cameras having parallel optical axes and a fixed distance between their aligned lenses. The pair of the obtained images can then be displayed by any of the known techniques for stereoscopic displaying and viewing. All of these techniques are based, in general, on the principle that the image taken by the right lens is displayed to the viewer's right eye and the image taken by the left lens is displayed to the viewer's left eye.
  • Panoramic photographing is conventionally done either by a very wide-angle lens, such as a “fish-eye” lens, or by “stitching” together slightly overlapping adjacent images to cover a wide, up to a fill circle, field of vision. Recently, the same techniques used for panoramic imaging are also exploited for obtaining spherical images. However, the panoramic or spherical images obtained by using said techniques are not stereoscopic, nor do they give to the viewer a perception of depth.
  • Moreover, the field of the so called “Virtual Reality” has gained high popularity in recent years. However, imaging systems for virtual reality, which are based on real images and not on computer generated (or other synthetically made) images, are still far from achieving images of a satisfactory quality. This is due mainly to the difficulties in obtaining real images that are full spherical as well as stereoscopic.
  • Accordingly, there is a need for an improved photographing system which will provide fully spherical, and fully stereoscopic real images.
  • The present invention is based on the finding that a stereoscopic image can be obtained by overlapping images taken by lenses having non-parallel optical axes, providing that said lenses' fields of view are overlapping to a great extent. This finding enables a full spherical stereoscopic photographing by a three dimensional array of cameras pointing out from a common center for covering the whole surroundings, both horizontally and vertically.
  • BRIEF SUMMARY OF THE INVENTION
  • The present invention provides a novel imaging system for obtaining full stereoscopic spherical images of the visual environment surrounding a viewer, 360 degree both horizontally and vertically. Displaying the images obtained by the present system, by means suitable for stereoscopic displaying, gives the viewers the ability to look everywhere around them, as well as up and down, while having stereoscopic depth perception of the displayed images.
  • One aspect of the present invention is a photographing system for capturing a full stereoscopic spherical image, or a continuous part thereof, comprising a three dimensional array of cameras. The cameras are positioned at an equal distance from a common center and pointing out from said common center, in such a manner that all lenses are situated on a spherical surface. The collective field of vision captured by all of the lenses covers the whole visual scale surrounding said photographing system, or a continuous part thereof, and any point in said collective field of vision is captured by at least two of said lenses. In one embodiment of the present invention, the cameras are arranged in such a way that the field of vision of any of the lenses is overlapped to a great extent by the fields of view of the two adjacent lenses lying on said lens sides, with respect to horizontal axis, and only to a small extent by the fields of vision of adjacent lenses lying above or below said lens with respect to the horizontal axis. According to this embodiment, the collective field of vision comprises a collection of, fully circular, stereoscopic panoramic images wherein any point within each of said panoramic images is captured by at least two lenses.
  • In a second embodiment, the cameras are arranged in the specific configuration, wherein the field of vision of any of said lenses is overlapped to a great extent by the fields of vision of all adjacent lenses surrounding said lens and wherein said lenses are optionally equally distributed on the spherical surface. The image information captured by each of said cameras is stored separately. Preferably the present invention further comprises a marking means for marking the boundaries of any of the images captured by any of said lenses, or a set of predetermined points within the area of each of said images and a means for image processing for processing the image information captured by any of said cameras and for cropping out redundant data in order to enhance said image quality.
  • A second aspect of the present invention is a multi-lens camera apparatus for generating a full stereoscopic spherical imaging. The apparatus comprises a set of lenses distributed on a spherical surface and a means for recording and optionally storing the image information of each of the images captured by each of said lenses. The lenses are distributed on the spherical surface, wherein the collective field of vision captured by all of said lenses, covers the whole visual environment surrounding the camera apparatus, and any point in said collective field of vision is captured by at least two of said lenses.
  • A third aspect of the present invention is a process for capturing full or partial stereoscopic spherical image of a visual environment to be displayed to a viewer by means allowing for stereoscopic viewing. The process comprises arranging cameras in a three dimensional array, wherein the lenses of all said cameras are situated on a spherical surface and the optical axes of all said lenses meet at the center of said spherical sphere, wherein the collective field of vision captured by all said lenses covers the whole visual environment surrounding array of cameras, or a continuous part thereof, and wherein any point in said collective field of vision is captured by at least two of said lenses.
  • A fourth aspect of the present invention is a process for capturing and displaying a full stereoscopic spherical image of a visual environment, to be displayed to a viewer by a system having means for stereoscopic displaying and viewing. The process comprises: a) arranging a three dimensional array of lenses for capturing a collection of images, wherein the collective field of vision captured by all said lenses covers the whole visual environment surrounding said array of lenses, and wherein any point in said collective field of vision is captured by at least two of said lenses; b) receiving from said system for stereoscopic displaying and viewing, the coordinates of said viewer's field of vision; c) creating a stereoscopic image pair from said collection of images, said stereoscopic image pair covers said viewer's field of vision; d) displaying said stereoscopic image pair to said viewer by means of said system for stereoscopic displaying and viewing. The step of creating a stereoscopic image pair from said collection of images comprises the following sub-steps: a) projecting said viewer's field of vision onto said collection of images; b) selecting from said collection of images a selected group of images overlapping at least partially said viewer's field of vision; c) dividing each of the images in said selected group of images into a left part and a right part by a line perpendicular to viewer's horizon and passing through image center to obtain a group of left image parts and a group of right image parts; d) selecting from said group of left image parts the parts overlapping said viewer's field of vision to obtain a group of selected left parts; e) merging said selected left parts into a uniform two-dimensional first image, said first image is the right image of said stereoscopic pair; f) selecting from said group of right image parts those parts overlapping said viewer's field of vision to obtain a group of selected right parts; g) merging said selected right parts into a uniform two-dimensional second image, said second image is the left image of said stereoscopic pair.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention will be understood and appreciated more fully from the following detailed description taken in conjunction with the drawings in which:
  • FIGS. 1A and 1B are two perspective views showing the configuration of the lenses on a sphere in accordance with the present invention;
  • FIG. 2 is a horizontal cross section plane through the sphere center showing the cameras positions for generating one strap of a full stereoscopic panoramic view;
  • FIG. 3 schematically illustrates the overlapping between fields of view of adjacent cameras for generating a stereoscopic image;
  • FIG. 4 is a flow chart describing the main steps for recording image information from each of the lenses comprising the system;
  • FIG. 5 is a block diagram of the present invention;
  • FIG. 6 is a flow chart describing the steps for generating a stereoscopic pair in accordance with the viewer's viewing parameters;
  • FIG. 7 pictorially describes an example of the process described in FIG. 6.
  • DETAILED DESCRIPTION OF THE INVENTION
  • The present invention provides a system and method for simultaneous capturing of a full spherical stereoscopic imaging or a continuous part thereof. The method of the present invention for obtaining stereoscopic imaging is based on optical principles, and therefore can be adapted to any photographing technique. Although the preferred embodiment of the present invention comprises digital video cameras, the present system can be adapted as well to analog video cameras and to film cameras. The present system can generate still images or a time sequence of images for motion imaging.
  • The apparatus of the present invention comprises a three dimensional array of cameras, pointing out from common centers, wherein said cameras' principal optical axes meet at said centers. In the configuration thus formed, the lenses are situated on a curved surface, allowing for capturing images of the surrounding environment. According to the preferred embodiment of the present invention the cameras are positioned at an equal distance form a common center. However any curved surface or object (e.g., ellipse) can be used, applying the same principles of the present invention.
  • FIGS. 1A and 1B are two exterior perspective views showing the position of lenses (designated 1) on the so formed sphere. FIG. 2 is a cross section plane through the sphere center showing eight cameras (designated 5) equally spaced and radially pointing outward from the center, wherein said cameras' principal optical axis (designating 15) meet at the sphere's center. The collection of images captured by lenses 1 covers a fully spherical visual surrounding the camera system, 360 degrees both horizontally and vertically
  • The actual installation of the cameras in order to form such fixed configuration can be obtained by various ways. For example, the cameras are mounted in designated gaps of a spherical object. Alternatively, the cameras can be mounted in a honeycomb net structure, etc.
  • The apparatus may be mounted on a supporting device 8 having means for enabling rotation and translation of the camera system. Such a support can be a tripod, having rolling means for allowing track-in/track-out and track-right/track-left translation and further having means for tilt and pan and for crane-up/crane-down. The apparatus further includes a stabilization system for preventing the system from vibrations and for keeping the cameras at their fixed positions relative to each other. The apparatus may further include internal robotics for allowing rotation, panning and tilting of the sphere, independently of support 8.
  • In the configuration described in FIGS. 1 and 2, the total number of lenses distributed on the sphere is eighteen. Using analogical terms of a globes terminology, the lenses arrangement can be described as follows: eight lenses are situated around the sphere equator for capturing a full 360 degree panoramic view, four lenses are situated around higher/lower latitude on each side of the equator for capturing the view above and below the horizon, and a lens at each pole to complete a full spherical view.
  • Preferably, the number of lenses is between 18 to 22. However, the number of cameras can be varied, depending on the desired quality of the image to be obtained, as will be explained below.
  • FIG. 3 shows three adjacent cameras, 5, 5′ and 5″ and their corresponding fields of vision, 11, 12 and 13. A main idea of the present invention is to create a cameras' arrangement wherein any point in the collective field of vision is captured by at least two lenses of the cameras, thus providing stereoscopic sphere image. Thus, starting from a predetermined distance d from the sphere surface, any point is covered by the fields of vision of at least two lenses. As shown in FIG. 3., the left half of field 12 is overlapped by field 11 while its right half is overlapped by field 13, thus any point in field 12 is covered by at least one more lens. Distance d is a function of the angle α between the optical axes of adjacent lenses, the angle of coverage β of each lens and the sphere radius r.
  • Another factor to be considered is the desired quality of the captured images. The smaller is a lens angle of coverage (field of vision), the higher is the quality of the image obtained. The best image quality is obtained by the central portion of a lens coverage area and as the angle of coverage of a lens increases, so does the relative range of distorted borders. Therefore, the number of the required lenses of the apparatus is a function of the desired quality. For better image quality, a larger number of lenses, each having a smaller field of vision, should be used. Preferably, the angle of coverage of the lenses is in the range of 30 to 65 degrees (lenses of 35 to 70 mm).
  • Preferably, the lenses are identical, having the same angle of coverage and the same focal length, but it is also possible to have lenses of different angle of coverage wherein lenses of wider angles are used for areas not requiring high image quality.
  • According to the present invention, the cameras can be arranged in two basic configurations, resulting two basic distributions of lenses on the sphere. According to a first configuration, the sphere is divided into parallel straps, equivalent to circumferential latitude straps around the globe. The lenses in each such strap are arranged wherein any point in collective field of vision, captured by the lenses in that specific strap, is covered by at least two lenses, but there is only small overlapping between adjacent straps. As the length of the straps decrease towards the poles with increasing latitude, the number of lenses covering each strap decreases. Optionally Additional two lenses at each pole, give an upward and downward view in order to complete the fall spherical image. Such a configuration facilitates generating a spherical stereoscopic imaging as if conventional tilt and pan displaying systems are used. The assumption is that the viewer's horizon is substantially parallel with respect to ground, i.e., that his eyes are substantially at the same height with respect to ground. In the context of the present invention, the viewer's horizon is defined as the vector connecting the viewer eyes.
  • According to a second configuration, the cameras are configured with no preference to latitude or longitude coordinates. According to this “mosaic like” configuration, the field of vision of each lens overlaps to a great extent the fields of view of all adjacent lenses wherein the lenses are optionally equally distributed on the spherical surface. This configuration provides that any point in the collective field of vision is captured by lenses located around said point. The collection of images captured by this configuration facilitates stereoscopic perception from all viewing points, also when the viewer horizon is inclined with respect to ground.
  • The present invention is not limited to the type of the cameras used in the system. However, according to a preferred embodiment of the present invention, the cameras are digital video cameras having an electro-optical sensor such as a CCD sensor. The cameras can be selected from any commercially available digital video cameras
  • According to this embodiment, the encoded digital image information, captured by each of the lenses, is transmitted via communication means to a remote storage device wherein each image is stored separately. The use of digital cameras allows for enhanced manipulation of the captured images and facilitates controlling the image quality prior to storage. However, it should be realized that when non-digital cameras are used, it is possible to digitize the stored images information for further manipulation and/or in order to convert it to a format compatible with the displaying and viewing technology known in the art.
  • According to the preferred embodiment of the present invention, the camera system further includes an internal computerized system comprising a suitable software for controlling the internal robotics responsible for the sphere motion and stabilization. The computerized system may further include a central processing unit, a storage device, a memory device and algorithms for processing the captured images prior to storage. Alternatively, the image processing can be done at a remote computer.
  • The digital image information captured by each camera is processed separately by the internal computer to remove distortion from each image separately. The images are cropped into pre-defined sections which are required for keeping an adequate overlap between adjacent images, wherein the image edges are eliminated for sparing the storage space of unnecessary information. The image enhancement may further include other known methods such as stretching and/or compressing certain parts of the image in order to correct distortions and to enhance the image quality. The image processing may also include processing methods as those employed by standard camera control units (CCU) in order to balance the image and achieve uniformity between adjacent images.
  • According to a preferred embodiment of the present invention, the camera system further includes a marking means for marking the borders of the lenses or a set of predetermined points in the rims and/or area of each of the lenses. The marking means can comprise one or more laser sources, such as a small diode laser, or other small radiation sources for generating beams of visible or invisible light. Said beams scan the borders of the lenses or a predetermined set of points in the area of the lenses to form a set of imprinted markers in the captured images.
  • Said set of imprinted markers is identified and used by the computerized system for accurate positioning of the images relative to each other. Although the relative positions of the cameras are fixed, small shifts might occur during operation. The markers system eliminates possible misalignment of the images due to such small shifts. Within each image, the marked borders or points are used to facilitate image processing in accordance with known image processing methods in order to remove distortion and to enhance the image quality.
  • It will be realized by those skilled in the art that the present invention is not limited to an array of commercially available cameras but can be implemented by a multi-lens camera apparatus designed in accordance with the spirit of the present invention. Such a multi-lens camera apparatus, comprising a set of lenses distributed on a spherical surface and corresponding sensor means for recording the image information captured by each of said lenses, can be built from basic camera elements. The sensor means can comprise a set of electro-optical sensors such as CCD chips, each corresponding to one lens, for receiving and converting the images captured by the lenses into encoded digital information In another configuration, the number of sensors can be smaller than the number of lenses, wherein along with appropriate optical elements, each sensor receives image information from more than one lens. For example, the sensor means can comprise only one electro-optical sensor wherein the camera apparatus includes suitable optical elements for projecting the images formed by each lens onto said one sensor. According to one embodiment of such a configuration, the sensor can be designed to be large enough to capture all images simultaneously. Alternatively, all images can be sequentially projected onto the same sensor area wherein a special mechanism controls a system of shutters for switching between lenses. (Similar mechanism can project the image on celluloid film mechanism).
  • Reference is now made to FIG. 4 which, in conjunction with FIG. 5, describes the main steps of the image recording and displaying. The image (or sequence of images) information captured by each of lenses 1 in step 510, is optionally processed in step 520 by image processor 30 in order to enhance the image quality. In step 530 the images are recorded by recording system 40 and are optionally stored in a suitable storage medium. The image information (whether analog or digital, whether one image or sequence of images) from each of cameras 5 is recorded and stored separately with a lens indicator and a time indicator. The lens indicator includes information about the lens characteristics and position in terms of the sphere coordinates. Thus, a collection of recorded images is created wherein each image is stored separately and includes information about its relative position with respect to adjacent images.
  • The image processing in step 520 is optional. Depending on the optical characteristics of the lenses employed and on the desired quality of the image, various image processing techniques if any, may be employed. These techniques may vary from a simple cropping out of the image borders in order to completely remove distorted areas, to more sophisticated image processing methods such as stretching and/or compressing certain parts of the image in order to correct distortions and to enhance the image quality. The image processing may also include processing methods as those employed by standard camera control units (CCU) in order to balance the image and achieve uniformity between adjacent images.
  • The recorded images are displayed to a viewer via viewing system 50 employing known virtual reality equipment 60. Alternatively, viewing system 50 receives image information in real time, directly from the camera system.
  • It should be emphasized that according to the present invention the image processing is an optional step. As explained above, the image processing is performed on each of the images separately in order to maintain higher image quality and with no relation to the stereoscopic effect. The stereoscopic effect, according to the present invention, is based solely on optical principles and not requiring any image processing for that purpose, as explained down bellow.
  • Reference is now made to FIG. 6 which is a flow chart describing the steps for generating a pair of stereoscopic images from the collection of recorded images in order to be displayed to the viewer according to his viewing parameters. The viewer's viewing parameters are detected by employing known means in the art of virtual reality, designated 60. Such means can be a headset having sensors to detect the viewer's head position or a virtual glove having a sensor to detect the hand position. For displaying on a flat screen, such as a computer screen, these parameters are taken from the pointing device (for example: a mouse or a joystick), programmed for this purpose.
  • In step 610, when a viewer selects a specific view, either by actually turning his head while wearing a virtual reality headset, or by a pointing device coupled to a computer device, the viewing parameters are detected and received by the displaying system. The viewer's viewing parameters include the viewer's viewing direction and viewer's horizon.
  • In step 620, in accordance with these parameters, the viewer's field of vision is determined in terms of the coordinates of the sphere surrounding the viewer and is projected into the collection of stored images. In step 630 all the images that overlap at least partially the viewer's field of vision are selected from the collection of recorded images.
  • Steps 640-680 illustrates the images organization algorithms based on the optical principles of the present invention for creating a pair of merged images wherein said pair of merged images creates the perception of three dimensional panoramic images. This algorithm can be alternatively applied for images captured by a single (or more) cameras wherein the images are taken from different angles and/or location of the same cameras. Such alternative is preferably applied for creating spherical stereoscopic still image.
  • First, a stereoscopic image pair is created from the group of selected images in the following way: Each of the selected images is divided into a left part and a right part according to the viewer's horizon by a line which is perpendicular the viewer horizon and is passing through the center of the image (step 640). In step 650 all the left parts generated in step 640 which are included in the viewer's field of vision are merged into a one uniform two dimensional image that matches the viewer's field of vision. The formed image, is the right image of the stereoscopic pair to be displayed to the viewer's right eye (step 660). Following the same lines, a left image is formed by merging together the right parts generated in step 640 ( steps 670, 680 and 690).
  • FIG. 7 pictorially describes one example of generating a stereoscopic pair according to the present invention. Frame 100 represents the viewer's field of vision as received by viewing system 60. The illustrated frame has rectangular shape, however the same stereoscopic principles can be applied to circular frames depending on the cameras technological capabilities. Frames 31, 32 and 33 represent three adjacent images, taken by the lenses (as viewed from underneath) of the camera system 10, which overlap viewer's field of vision 100. The middle frame 32 covers frame 100 completely while frames 31 and 33 partly overlap the left and the right sides of frame 100, respectively. In accordance with steps 640-660 of FIG. 6, a new image 40, to be displayed to the viewer's right eye, is formed from images 32 and 33 by merging a left portion of image 32 with a left portion of image 33. Left image 41 is formed similarly by merging a right portion of image 32 with a right portion of image 31 to form a uniform image. New images 40 and 41 both cover the viewer's field of vision but obviously they are not identical, and are perceived as were taken from two viewpoints, thus giving the viewer a stereoscopic perception. Roughly, this new image pair is equivalent to a pair of images as if were taken by two virtual lenses having their optical axes directed forward in the viewer's viewing direction and having horizontal disparity.
  • The process illustrated in FIG. 7 is for a case where the viewer's field of vision overlaps three recorded images. However, it will be easily perceived that when the viewer's field of vision includes more recorded images (e.g., when it includes upper and lower parts of images from different latitudes or when each image covers only a small part of the viewer's field of vision), the stereoscopic pair is formed along the same lines, by merging together the appropriate parts from each of the recorded images and by removing redundant information. The same holds (when using the second configuration of cameras arrangement wherein the field of vision of each lens overlaps to a great extent the fields of view of all adjacent lenses and the lenses are optionally equally distributed on the spherical surface) for cases where the viewer's horizon is inclined with respect to ground (i.e., when the viewer eyes are not at the same height with respect to ground). For such cases, the selection of the recorded images is done by projecting the viewer's field of vision on the collection of recorded images and selecting those images that form with each other the closest angle to the inclining angle. Upon selecting the appropriate image, the stereoscopic pair is generated by following steps 640 to 680 of FIG. 6.
  • The stereoscopic image pair thus formed can be displayed on a flat screen such as a TV or a computer screen or by using a display device for virtual reality such as a virtual reality headset. When displayed on a flat screen, the images are displayed as a stereoscopic pair to be viewed by suitable spectacles in accordance with the standard color filtering method for stereoscopic viewing or by any other known in the art techniques for stereoscopic displaying: The part of the image being displayed changes according to the viewer viewpoint as explained above.
  • It will be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described herein above. Rather the scope of the present invention is defined only by the claims that follow.

Claims (5)

1. A process for capturing and for displaying a full stereoscopic curved image of a visual environment, to be displayed to a viewer by a system having means for stereoscopic displaying and viewing, said process comprising:
arranging a three dimensional array of lenses for capturing a collection of images, wherein a collective field of vision captured by all of said lenses covers a whole visual environment surrounding said array of lenses, and wherein any point in said collective field of vision is captured by at least two of said lenses;
receiving from said system for stereoscopic displaying and viewing, the coordinates of said viewer's field of vision;
creating a stereoscopic image pair from said collection of images, said stereoscopic image pair to cover said viewer's field of vision, and comprising divided left and right parts to form divided right and left images, respectively; and
displaying said stereoscopic image pair to said viewer by means of said system for stereoscopic displaying and viewing.
2. The process of claim 1 wherein creating a stereoscopic image pair from said collection of images comprises:
projecting said viewer's field of vision onto said collection of images;
selecting from said collection of images a selected group of images overlapping at least partially said viewer's field of vision;
dividing each of the selected group of images into a left part and a right part by a line perpendicular to a horizon of said viewer and passing through an image center to obtain a group of left image parts and a group of right image parts;
selecting from said group of left image parts those parts overlapping said viewer's field of vision to obtain a group of selected left parts;
merging said group of selected left parts into a uniform two-dimensional first image as the divided left part, said first image comprising a right image of said stereoscopic pair as the divided right image;
selecting from said group of right image parts those parts overlapping said viewer's field of vision to obtain a group of selected right parts; and
merging said group of selected right parts into a uniform two-dimensional second image as the divided right part, said second image comprising a left image of said stereoscopic pair as the divided left image.
3. The process of claim 1 further comprising processing each image in said collection of images to enhance image quality, to enhance image balance between adjacent images, and to crop out redundant data.
4. A process for capturing and for displaying a full stereoscopic curved image of a visual environment, to be displayed to a viewer by a system having means for stereoscopic displaying and viewing, said process comprising:
capturing a collection of images taken by at least one camera from different angles and/or locations, wherein a collective field of said images covers a surrounding visual environment and wherein any point in said collective field of vision is captured by at least two of said images;
receiving from said system for stereoscopic displaying and viewing, the coordinates of said viewer's field of vision;
creating a stereoscopic image pair from said collection of images, said stereoscopic image pair to cover said viewer's field of vision, and comprising divided left and right parts to form divided right and left images, respectively; and
displaying said stereoscopic image pair to said viewer by means of said system for stereoscopic displaying and viewing.
5. The process of claim 4 wherein creating a stereoscopic image pair from said collection of images comprises:
projecting said viewer's field of vision onto said collection of images;
selecting from said collection of images a selected group of images overlapping at least partially said viewer's field of vision;
dividing each image in said selected group of images into a left part and a right part by a line perpendicular to viewer's horizon and passing through an image center to obtain a group of left image parts and a group of right image parts;
selecting from said group of left image parts those parts overlapping said viewer's field of vision to obtain a group of selected left parts;
merging said selected left parts into a uniform two-dimensional first image as the divided left part, said first image comprising a right image of said stereoscopic image pair as the divided right image;
selecting from said group of right image parts those parts overlapping said viewer's field of vision to obtain a group of selected right parts; and
merging said group of selected right parts into a uniform two-dimensional second image as the divided right part, said second image comprising a left image of said stereoscopic image pair as the divided left image.
US12/107,048 2000-11-29 2008-04-21 System and method for spherical stereoscopic photographing Abandoned US20080316301A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/107,048 US20080316301A1 (en) 2000-11-29 2008-04-21 System and method for spherical stereoscopic photographing

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
IL139995A IL139995A (en) 2000-11-29 2000-11-29 System and method for spherical stereoscopic photographing
ILIL139995 2000-11-29
PCT/IL2001/001093 WO2002044808A2 (en) 2000-11-29 2001-11-28 System and method for spherical stereoscopic photographing
US10/416,533 US7429997B2 (en) 2000-11-29 2001-11-28 System and method for spherical stereoscopic photographing
US12/107,048 US20080316301A1 (en) 2000-11-29 2008-04-21 System and method for spherical stereoscopic photographing

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US10/416,533 Division US7429997B2 (en) 2000-11-29 2001-11-28 System and method for spherical stereoscopic photographing
PCT/IL2001/001093 Division WO2002044808A2 (en) 2000-11-29 2001-11-28 System and method for spherical stereoscopic photographing

Publications (1)

Publication Number Publication Date
US20080316301A1 true US20080316301A1 (en) 2008-12-25

Family

ID=11074860

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/416,533 Expired - Lifetime US7429997B2 (en) 2000-11-29 2001-11-28 System and method for spherical stereoscopic photographing
US12/107,048 Abandoned US20080316301A1 (en) 2000-11-29 2008-04-21 System and method for spherical stereoscopic photographing

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10/416,533 Expired - Lifetime US7429997B2 (en) 2000-11-29 2001-11-28 System and method for spherical stereoscopic photographing

Country Status (6)

Country Link
US (2) US7429997B2 (en)
EP (1) EP1346259A4 (en)
JP (1) JP2004514951A (en)
AU (1) AU2002218465A1 (en)
IL (1) IL139995A (en)
WO (1) WO2002044808A2 (en)

Cited By (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110001798A1 (en) * 2009-07-02 2011-01-06 Sony Corporation 3-d auto-convergence camera
US20110001797A1 (en) * 2009-07-02 2011-01-06 Sony Corporation 3-d auto-convergence camera
CN102472959A (en) * 2009-07-02 2012-05-23 索尼公司 3-d auto-convergence camera
WO2013036569A1 (en) * 2011-09-06 2013-03-14 King Abdullah University Of Science And Technology Stereo-panoramic data
US20140098193A1 (en) * 2012-10-05 2014-04-10 Research In Motion Limited Methods and devices for generating a stereoscopic image
WO2015054326A1 (en) * 2013-10-11 2015-04-16 Microsoft Corporation Eye tracking
US20150138336A1 (en) * 2013-04-18 2015-05-21 Okamoto Glass Co., Ltd. Housing for an underwater video camera
US9094540B2 (en) 2012-12-13 2015-07-28 Microsoft Technology Licensing, Llc Displacing image on imager in multi-lens cameras
US20150302642A1 (en) * 2014-04-18 2015-10-22 Magic Leap, Inc. Room based sensors in an augmented reality system
US20150321103A1 (en) * 2014-05-08 2015-11-12 Sony Computer Entertainment Europe Limited Image capture method and apparatus
US9294672B2 (en) * 2014-06-20 2016-03-22 Qualcomm Incorporated Multi-camera system using folded optics free from parallax and tilt artifacts
WO2016061640A1 (en) * 2014-10-22 2016-04-28 Parallaxter Method for collecting image data for producing immersive video and method for viewing a space on the basis of the image data
US9374516B2 (en) 2014-04-04 2016-06-21 Qualcomm Incorporated Auto-focus in low-profile folded optics multi-camera system
US9386222B2 (en) 2014-06-20 2016-07-05 Qualcomm Incorporated Multi-camera system using folded optics free from parallax artifacts
US9383550B2 (en) 2014-04-04 2016-07-05 Qualcomm Incorporated Auto-focus in low-profile folded optics multi-camera system
US9398264B2 (en) 2012-10-19 2016-07-19 Qualcomm Incorporated Multi-camera system using folded optics
US9438889B2 (en) 2011-09-21 2016-09-06 Qualcomm Incorporated System and method for improving methods of manufacturing stereoscopic image sensors
US20160307372A1 (en) * 2015-04-15 2016-10-20 Lytro, Inc. Capturing light-field volume image and video data using tiled light-field cameras
US9485495B2 (en) 2010-08-09 2016-11-01 Qualcomm Incorporated Autofocus for stereo images
US9541740B2 (en) 2014-06-20 2017-01-10 Qualcomm Incorporated Folded optic array camera using refractive prisms
US9549107B2 (en) 2014-06-20 2017-01-17 Qualcomm Incorporated Autofocus for folded optic array cameras
WO2017109274A1 (en) * 2015-12-23 2017-06-29 Nokia Technologies Oy Methods and apparatuses relating to the handling of a plurality of content streams
US9720559B2 (en) 2013-10-14 2017-08-01 Microsoft Technology Licensing, Llc Command authentication
US9819863B2 (en) 2014-06-20 2017-11-14 Qualcomm Incorporated Wide field of view array camera for hemispheric and spherical imaging
US9832381B2 (en) 2014-10-31 2017-11-28 Qualcomm Incorporated Optical image stabilization for thin cameras
US20170363949A1 (en) * 2015-05-27 2017-12-21 Google Inc Multi-tier camera rig for stereoscopic image capture
WO2018035347A1 (en) * 2016-08-17 2018-02-22 Google Llc Multi-tier camera rig for stereoscopic image capture
US10013764B2 (en) 2014-06-19 2018-07-03 Qualcomm Incorporated Local adaptive histogram equalization
US10038887B2 (en) 2015-05-27 2018-07-31 Google Llc Capture and render of panoramic virtual reality content
US10178373B2 (en) 2013-08-16 2019-01-08 Qualcomm Incorporated Stereo yaw correction using autofocus feedback
US10205896B2 (en) 2015-07-24 2019-02-12 Google Llc Automatic lens flare detection and correction for light-field images
US10275898B1 (en) 2015-04-15 2019-04-30 Google Llc Wedge-based light-field video capture
US10275892B2 (en) 2016-06-09 2019-04-30 Google Llc Multi-view scene segmentation and propagation
US10298834B2 (en) 2006-12-01 2019-05-21 Google Llc Video refocusing
US10334151B2 (en) 2013-04-22 2019-06-25 Google Llc Phase detection autofocus using subaperture images
US10341632B2 (en) 2015-04-15 2019-07-02 Google Llc. Spatial random access enabled video system with a three-dimensional viewing volume
US10354399B2 (en) 2017-05-25 2019-07-16 Google Llc Multi-view back-projection to a light-field
US10375381B2 (en) 2015-05-27 2019-08-06 Google Llc Omnistereo capture and render of panoramic virtual reality content
US10412373B2 (en) 2015-04-15 2019-09-10 Google Llc Image capture for virtual reality displays
US10419737B2 (en) 2015-04-15 2019-09-17 Google Llc Data structures and delivery methods for expediting virtual reality playback
US10440407B2 (en) 2017-05-09 2019-10-08 Google Llc Adaptive control for immersive experience delivery
US10444931B2 (en) 2017-05-09 2019-10-15 Google Llc Vantage generation and interactive playback
US10469873B2 (en) 2015-04-15 2019-11-05 Google Llc Encoding and decoding virtual reality video
US10474227B2 (en) 2017-05-09 2019-11-12 Google Llc Generation of virtual reality with 6 degrees of freedom from limited viewer data
US10540818B2 (en) 2015-04-15 2020-01-21 Google Llc Stereo image generation and interactive playback
US10545215B2 (en) 2017-09-13 2020-01-28 Google Llc 4D camera tracking and optical stabilization
US10546424B2 (en) 2015-04-15 2020-01-28 Google Llc Layered content delivery for virtual and augmented reality experiences
US10552947B2 (en) 2012-06-26 2020-02-04 Google Llc Depth-based image blurring
US10565734B2 (en) 2015-04-15 2020-02-18 Google Llc Video capture, processing, calibration, computational fiber artifact removal, and light-field pipeline
US10567464B2 (en) 2015-04-15 2020-02-18 Google Llc Video compression with adaptive view-dependent lighting removal
US10594945B2 (en) 2017-04-03 2020-03-17 Google Llc Generating dolly zoom effect using light field image data
US10679361B2 (en) 2016-12-05 2020-06-09 Google Llc Multi-view rotoscope contour propagation
US10965862B2 (en) 2018-01-18 2021-03-30 Google Llc Multi-camera navigation interface
TWI730298B (en) * 2018-02-22 2021-06-11 弗勞恩霍夫爾協會 Generating panoramic images
US11328446B2 (en) 2015-04-15 2022-05-10 Google Llc Combining light-field data with active depth data for depth map generation
EP3057066B1 (en) * 2015-02-10 2022-12-14 DreamWorks Animation LLC Generation of three-dimensional imagery from a two-dimensional image using a depth map

Families Citing this family (121)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7084904B2 (en) * 2002-09-30 2006-08-01 Microsoft Corporation Foveated wide-angle imaging system and method for capturing and viewing wide-angle images in real time
US7463280B2 (en) * 2003-06-03 2008-12-09 Steuart Iii Leonard P Digital 3D/360 degree camera system
KR20070007059A (en) 2003-12-26 2007-01-12 미코이 코포레이션 Multi-dimensional imaging apparatus, systems, and methods
US8681100B2 (en) 2004-07-30 2014-03-25 Extreme Realty Ltd. Apparatus system and method for human-machine-interface
US8928654B2 (en) 2004-07-30 2015-01-06 Extreme Reality Ltd. Methods, systems, devices and associated processing logic for generating stereoscopic images and video
US8872899B2 (en) * 2004-07-30 2014-10-28 Extreme Reality Ltd. Method circuit and system for human to machine interfacing by hand gestures
KR101238608B1 (en) 2004-07-30 2013-02-28 익스트림 리얼리티 엘티디. A system and method for 3D space-dimension based image processing
US7872665B2 (en) 2005-05-13 2011-01-18 Micoy Corporation Image capture and processing
TW200715830A (en) * 2005-10-07 2007-04-16 Sony Taiwan Ltd Image pick-up device of multiple lens camera system to create panoramic image
US20070285554A1 (en) * 2005-10-31 2007-12-13 Dor Givon Apparatus method and system for imaging
US9046962B2 (en) 2005-10-31 2015-06-02 Extreme Reality Ltd. Methods, systems, apparatuses, circuits and associated computer executable code for detecting motion, position and/or orientation of objects within a defined spatial region
US9182228B2 (en) * 2006-02-13 2015-11-10 Sony Corporation Multi-lens array system and method
US8498497B2 (en) * 2006-11-17 2013-07-30 Microsoft Corporation Swarm imaging
WO2008141270A2 (en) * 2007-05-10 2008-11-20 Mora Assad F Stereoscopic three dimensional visualization system and method of use
US8300086B2 (en) 2007-12-20 2012-10-30 Nokia Corporation Image processing for supporting a stereoscopic presentation
EP2342642A1 (en) * 2008-09-04 2011-07-13 Extreme Reality Ltd. Method system and software for providing image sensor based human machine interfacing
US20120159996A1 (en) * 2010-12-28 2012-06-28 Gary Edwin Sutton Curved sensor formed from silicon fibers
US9144714B2 (en) * 2009-05-02 2015-09-29 Steven J. Hollinger Ball with camera for reconnaissance or recreation and network for operating the same
CA2774867A1 (en) 2009-09-21 2011-03-24 Extreme Reality Ltd. Methods circuits apparatus and systems for human machine interfacing with an electronic appliance
US8878779B2 (en) 2009-09-21 2014-11-04 Extreme Reality Ltd. Methods circuits device systems and associated computer executable code for facilitating interfacing with a computing platform display screen
JP5337658B2 (en) * 2009-10-02 2013-11-06 株式会社トプコン Wide-angle imaging device and measurement system
CN102118554B (en) * 2010-01-04 2014-07-02 鸿富锦精密工业(深圳)有限公司 Monitoring system
US8836848B2 (en) * 2010-01-26 2014-09-16 Southwest Research Institute Vision system
US8942964B2 (en) 2010-06-08 2015-01-27 Southwest Research Institute Optical state estimation and simulation environment for unmanned aerial vehicles
US8730396B2 (en) * 2010-06-23 2014-05-20 MindTree Limited Capturing events of interest by spatio-temporal video analysis
WO2012056437A1 (en) 2010-10-29 2012-05-03 École Polytechnique Fédérale De Lausanne (Epfl) Omnidirectional sensor array system
US9473700B2 (en) 2010-11-03 2016-10-18 The Trustees Of Columbia University In The City Of New York Camera systems and methods for gigapixel computational imaging
WO2012136388A1 (en) * 2011-04-08 2012-10-11 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Capturing panoramic or semi-panoramic 3d scenes
US8466406B2 (en) 2011-05-12 2013-06-18 Southwest Research Institute Wide-angle laser signal sensor having a 360 degree field of view in a horizontal plane and a positive 90 degree field of view in a vertical plane
US9762794B2 (en) 2011-05-17 2017-09-12 Apple Inc. Positional sensor-assisted perspective correction for panoramic photography
US20120293607A1 (en) * 2011-05-17 2012-11-22 Apple Inc. Panorama Processing
JP2013051522A (en) * 2011-08-30 2013-03-14 Honda Elesys Co Ltd Image compression device, image compression method, and image compression program
WO2014007873A2 (en) 2012-03-20 2014-01-09 Wagreich David Image monitoring and display from unmanned vehicle
US20140327733A1 (en) 2012-03-20 2014-11-06 David Wagreich Image monitoring and display from unmanned vehicle
US9426430B2 (en) * 2012-03-22 2016-08-23 Bounce Imaging, Inc. Remote surveillance sensor apparatus
US9098922B2 (en) 2012-06-06 2015-08-04 Apple Inc. Adaptive image blending operations
US10306140B2 (en) 2012-06-06 2019-05-28 Apple Inc. Motion adaptive image slice selection
ITRM20120329A1 (en) * 2012-07-12 2012-10-11 Virtualmind Di Davide Angelelli 360 ° IMMERSIVE / SPHERICAL VIDEO CAMERA WITH 6-11 OPTICS 5-10 MEGAPIXEL WITH GPS GEOLOCALIZATION
US9479697B2 (en) * 2012-10-23 2016-10-25 Bounce Imaging, Inc. Systems, methods and media for generating a panoramic view
WO2014071400A1 (en) 2012-11-05 2014-05-08 360 Heros, Inc. 360 degree camera mount and related photographic and video system
US8902322B2 (en) 2012-11-09 2014-12-02 Bubl Technology Inc. Systems and methods for generating spherical images
WO2014105385A1 (en) * 2012-12-27 2014-07-03 The Regents Of The University Of California Anamorphic stretch image compression
US9413930B2 (en) * 2013-03-14 2016-08-09 Joergen Geerds Camera system
ITRM20130063U1 (en) * 2013-04-04 2014-10-05 Virtualmind Di Davide Angelelli PROBE FOR ENDOSCOPIC SHOOTS AND VIDEOINSPECTS, NAME REALWORLD360
US8982472B2 (en) * 2013-05-21 2015-03-17 Matvey Lvovskiy Method of widening of angular field of view of collimating optical systems
US9832378B2 (en) 2013-06-06 2017-11-28 Apple Inc. Exposure mapping and dynamic thresholding for blending of multiple images using floating exposure
KR102082300B1 (en) * 2013-07-01 2020-02-27 삼성전자주식회사 Apparatus and method for generating or reproducing three-dimensional image
US11019258B2 (en) 2013-08-21 2021-05-25 Verizon Patent And Licensing Inc. Aggregating images and audio data to generate content
US9451162B2 (en) 2013-08-21 2016-09-20 Jaunt Inc. Camera array including camera modules
KR101471927B1 (en) * 2013-09-13 2014-12-11 우석 계 Multi-camera, display apparatus and method using the multi-camera
US20150138311A1 (en) * 2013-11-21 2015-05-21 Panavision International, L.P. 360-degree panoramic camera systems
GB2525170A (en) * 2014-04-07 2015-10-21 Nokia Technologies Oy Stereo viewing
JP6599603B2 (en) * 2014-04-18 2019-10-30 東芝ライフスタイル株式会社 Autonomous vehicle
US9911454B2 (en) * 2014-05-29 2018-03-06 Jaunt Inc. Camera array including camera modules
US10368011B2 (en) 2014-07-25 2019-07-30 Jaunt Inc. Camera array removing lens distortion
US11108971B2 (en) 2014-07-25 2021-08-31 Verzon Patent and Licensing Ine. Camera array removing lens distortion
US9774887B1 (en) 2016-09-19 2017-09-26 Jaunt Inc. Behavioral directional encoding of three-dimensional video
US10186301B1 (en) 2014-07-28 2019-01-22 Jaunt Inc. Camera array including camera modules
US10701426B1 (en) 2014-07-28 2020-06-30 Verizon Patent And Licensing Inc. Virtual reality system including social graph
US9363569B1 (en) * 2014-07-28 2016-06-07 Jaunt Inc. Virtual reality system including social graph
US10440398B2 (en) 2014-07-28 2019-10-08 Jaunt, Inc. Probabilistic model to compress images for three-dimensional video
EP3183687B1 (en) 2014-08-21 2020-07-08 IdentiFlight International, LLC Avian detection system and method
US9856856B2 (en) * 2014-08-21 2018-01-02 Identiflight International, Llc Imaging array for bird or bat detection and identification
KR20160028165A (en) * 2014-09-03 2016-03-11 (주)에프엑스기어 Apparatus and method for recording panorama stereo image
EP3204824A4 (en) * 2014-10-07 2018-06-20 Nokia Technologies Oy Camera devices with a large field of view for stereo imaging
US10358214B2 (en) 2015-01-04 2019-07-23 Hangzhou Zero Zro Technology Co., Ltd. Aerial vehicle and method of operation
US10220954B2 (en) 2015-01-04 2019-03-05 Zero Zero Robotics Inc Aerial system thermal control system and method
US9836053B2 (en) 2015-01-04 2017-12-05 Zero Zero Robotics Inc. System and method for automated aerial system operation
US10126745B2 (en) 2015-01-04 2018-11-13 Hangzhou Zero Zero Technology Co., Ltd. System and method for automated aerial system operation
JP6283329B2 (en) * 2015-05-26 2018-02-21 株式会社オプティム Augmented Reality Object Recognition Device
US9277122B1 (en) * 2015-08-13 2016-03-01 Legend3D, Inc. System and method for removing camera rotation from a panoramic video
US10217189B2 (en) * 2015-09-16 2019-02-26 Google Llc General spherical capture methods
EP3374263A4 (en) 2015-11-10 2019-05-08 Matternet, Inc. Methods and systems for transportation using unmanned aerial vehicles
JP2019035584A (en) * 2015-12-01 2019-03-07 株式会社ソニー・インタラクティブエンタテインメント Distance measurement device
US20170195560A1 (en) * 2015-12-31 2017-07-06 Nokia Technologies Oy Method and apparatus for generating a panoramic view with regions of different dimensionality
EP3190566A1 (en) * 2016-01-05 2017-07-12 Giroptic Spherical virtual reality camera
WO2017136776A1 (en) * 2016-02-05 2017-08-10 Vantage Robotics, Llc Durable modular unmanned aerial vehicle
JP6673584B2 (en) * 2016-02-25 2020-03-25 株式会社ザクティ Imaging device
KR101784095B1 (en) * 2016-03-14 2017-10-11 윤승훈 Head-mounted display apparatus using a plurality of data and system for transmitting and receiving the plurality of data
US10057487B1 (en) * 2016-03-25 2018-08-21 Scott Zhihao Chen Panoramic imaging systems based on normal-lens cameras
EP3229073B1 (en) * 2016-04-06 2018-10-31 Facebook, Inc. Three-dimensional, 360-degree virtual reality camera system
US10230904B2 (en) * 2016-04-06 2019-03-12 Facebook, Inc. Three-dimensional, 360-degree virtual reality camera system
USD821479S1 (en) * 2016-04-08 2018-06-26 Facebook, Inc. Panoramic virtual reality camera
US10435144B2 (en) 2016-04-24 2019-10-08 Hangzhou Zero Zero Technology Co., Ltd. Aerial system propulsion assembly and method of use
US20170322396A1 (en) * 2016-05-06 2017-11-09 VideoStitch Inc. Compact camera design to adjust parallax effects
CN105739231B (en) * 2016-05-06 2019-04-26 中国科学技术大学 A kind of multi-cam full-view stereo imaging device of plane distribution
USD818024S1 (en) * 2016-06-30 2018-05-15 Zhihao Chen Panoramic camera
USD830444S1 (en) * 2016-06-30 2018-10-09 Facebook, Inc. Panoramic virtual reality camera
USD837275S1 (en) 2016-06-30 2019-01-01 Facebook, Inc. Panoramic virtual reality camera assembly
USD830445S1 (en) 2016-06-30 2018-10-09 Facebook, Inc. Panoramic virtual reality camera
CN106027919A (en) * 2016-06-30 2016-10-12 北京和兴宏图科技有限公司 Video camera
AT518256B1 (en) * 2016-08-02 2017-09-15 Innaq Gmbh GENERATING A PANORAMIC IMPOSITION FOR STEREOSCOPIC REPRODUCTION AND SUCH A PLAYBACK
JP1574354S (en) * 2016-08-25 2017-04-17
JP1572074S (en) * 2016-08-25 2017-03-21
US11032536B2 (en) 2016-09-19 2021-06-08 Verizon Patent And Licensing Inc. Generating a three-dimensional preview from a two-dimensional selectable icon of a three-dimensional reality video
US11032535B2 (en) 2016-09-19 2021-06-08 Verizon Patent And Licensing Inc. Generating a three-dimensional preview of a three-dimensional video
US10681341B2 (en) 2016-09-19 2020-06-09 Verizon Patent And Licensing Inc. Using a sphere to reorient a location of a user in a three-dimensional virtual reality video
USD815677S1 (en) * 2016-09-19 2018-04-17 Coasia Microelectronics(Shanghai) Co. Ltd 360° panoramic camera
JP6732617B2 (en) * 2016-09-21 2020-07-29 株式会社ソニー・インタラクティブエンタテインメント Information processing apparatus and image generation method
JP6683307B2 (en) * 2016-10-06 2020-04-15 カイ インコーポレイテッド Optimal spherical image acquisition method using multiple cameras
USD819718S1 (en) * 2016-10-25 2018-06-05 Shenzhen Arashi Vision Company Limited Camera
GB2555585A (en) * 2016-10-31 2018-05-09 Nokia Technologies Oy Multiple view colour reconstruction
JP6952456B2 (en) * 2016-11-28 2021-10-20 キヤノン株式会社 Information processing equipment, control methods, and programs
EP3339714A1 (en) * 2016-12-20 2018-06-27 Nokia Technologies Oy An apparatus and associated methods for virtual reality image capture
EP3339951A1 (en) 2016-12-20 2018-06-27 Nokia Technologies Oy Fill lighting apparatus
US11636572B2 (en) 2016-12-29 2023-04-25 Nokia Technologies Oy Method and apparatus for determining and varying the panning speed of an image based on saliency
TWD184894S (en) * 2017-01-23 2017-08-11 廣達電腦股份有限公司 Camera device
US10067513B2 (en) 2017-01-23 2018-09-04 Hangzhou Zero Zero Technology Co., Ltd Multi-camera system and method of use
TWD185642S (en) * 2017-01-23 2017-09-21 威盛電子股份有限公司 Camera device
US10652460B2 (en) * 2017-03-30 2020-05-12 Canon Kabushiki Kaisha Image-capturing apparatus and image-capturing control method
JP1596758S (en) * 2017-05-23 2018-02-05
KR102465248B1 (en) 2017-09-11 2022-11-11 삼성전자주식회사 Apparatus and method for processing image
US10440349B2 (en) 2017-09-27 2019-10-08 Facebook Technologies, Llc 3-D 360 degrees depth projector
EP3462128B1 (en) * 2017-09-27 2021-11-10 Facebook Technologies, LLC 3d 360-degree camera system
JP2019080223A (en) 2017-10-26 2019-05-23 株式会社ソニー・インタラクティブエンタテインメント Camera system
JP7011728B2 (en) * 2018-09-28 2022-01-27 株式会社ソニー・インタラクティブエンタテインメント Image data output device, content creation device, content playback device, image data output method, content creation method, and content playback method
US10694167B1 (en) 2018-12-12 2020-06-23 Verizon Patent And Licensing Inc. Camera array including camera modules
USD897398S1 (en) * 2019-03-25 2020-09-29 Arashi Vision Inc. Panoramic camera
US11178374B2 (en) * 2019-05-31 2021-11-16 Adobe Inc. Dynamically rendering 360-degree videos using view-specific-filter parameters
WO2022072058A1 (en) * 2020-09-29 2022-04-07 James Logan Wearable virtual reality (vr) camera system
CN113079365B (en) * 2021-03-26 2022-09-09 中国人民解放军陆军装甲兵学院 Integrated imaging unit image array generation method and system based on tilt-shift camera

Citations (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3187339A (en) * 1961-10-16 1965-06-01 Wallace A Clay Three dimensional viewing apparatus
US3357770A (en) * 1961-10-02 1967-12-12 Intermountain Res And Engineer Stereoscopic viewing apparatus which includes a curved lenticular screen in front ofa curved picture supporting surface
US3953869A (en) * 1974-09-24 1976-04-27 Dimensional Development Corporation Stereoscopic photography apparatus
US4063265A (en) * 1974-09-24 1977-12-13 Dimensional Development Corporation Apparatus for taking stereoscopic pictures
US4214821A (en) * 1978-10-30 1980-07-29 Termes Richard A Total environment photographic mount and photograph
US4475798A (en) * 1977-12-27 1984-10-09 The Three Dimensional Photography Corporation Camera for stereoscopic photography
US4525038A (en) * 1981-05-20 1985-06-25 Mueller Rolf Fish eye lens system
US4868682A (en) * 1986-06-27 1989-09-19 Yamaha Corporation Method of recording and reproducing video and sound information using plural recording devices and plural reproducing devices
US5023725A (en) * 1989-10-23 1991-06-11 Mccutchen David Method and apparatus for dodecahedral imaging system
US5130794A (en) * 1990-03-29 1992-07-14 Ritchey Kurtis J Panoramic display system
US5562572A (en) * 1995-03-10 1996-10-08 Carmein; David E. E. Omni-directional treadmill
US5650813A (en) * 1992-11-20 1997-07-22 Picker International, Inc. Panoramic time delay and integration video camera system
US5657073A (en) * 1995-06-01 1997-08-12 Panoramic Viewing Systems, Inc. Seamless multi-camera panoramic imaging with distortion correction and selectable field of view
US5668595A (en) * 1992-03-23 1997-09-16 Canon Kabushiki Kaisha Multi-lens imaging apparatus having a mechanism for combining a plurality of images without displacement of registration
US5703604A (en) * 1995-05-22 1997-12-30 Dodeca Llc Immersive dodecaherdral video viewing system
US5721585A (en) * 1996-08-08 1998-02-24 Keast; Jeffrey D. Digital video panoramic image capture and display system
US5822117A (en) * 1996-01-22 1998-10-13 Kleinberger; Paul Systems for three-dimensional viewing including first and second light polarizing layers
US5943056A (en) * 1995-07-11 1999-08-24 Fujitsu Ltd. Interference checking method
US5946077A (en) * 1994-09-06 1999-08-31 Herman D. Mims Method and apparatus for improved three dimensional photography
US5973726A (en) * 1993-09-24 1999-10-26 Canon Kabushiki Kaisha Panoramic image processing apparatus
US5980256A (en) * 1993-10-29 1999-11-09 Carmein; David E. E. Virtual reality system with enhanced sensory apparatus
US6002430A (en) * 1994-01-31 1999-12-14 Interactive Pictures Corporation Method and apparatus for simultaneous capture of a spherical image
US6016151A (en) * 1997-09-12 2000-01-18 Neomagic Corp. 3D triangle rendering by texture hardware and color software using simultaneous triangle-walking and interpolation for parallel operation
US6023588A (en) * 1998-09-28 2000-02-08 Eastman Kodak Company Method and apparatus for capturing panoramic images with range data
US6031540A (en) * 1995-11-02 2000-02-29 Imove Inc. Method and apparatus for simulating movement in multidimensional space with polygonal projections from subhemispherical imagery
US6034716A (en) * 1997-12-18 2000-03-07 Whiting; Joshua B. Panoramic digital camera system
US6075905A (en) * 1996-07-17 2000-06-13 Sarnoff Corporation Method and apparatus for mosaic image construction
US6097854A (en) * 1997-08-01 2000-08-01 Microsoft Corporation Image mosaic construction system and apparatus with patch-based alignment, global block adjustment and pair-wise motion-based local warping
US6108005A (en) * 1996-08-30 2000-08-22 Space Corporation Method for producing a synthesized stereoscopic image
US6112033A (en) * 1997-10-29 2000-08-29 Canon Kabushiki Kaisha Multiple-lens image pickup apparatus for viewing in stereoscopic panoramic, and ordinary photographing modes
US6141034A (en) * 1995-12-15 2000-10-31 Immersive Media Co. Immersive imaging method and apparatus
US6169858B1 (en) * 1999-07-26 2001-01-02 Eastman Kodak Company Panoramic image capture aid
US6236748B1 (en) * 1994-08-02 2001-05-22 Canon Kabushiki Kaisha Compound eye image pickup device utilizing plural image sensors and plural lenses
US6243103B1 (en) * 1996-05-28 2001-06-05 Canon Kabushiki Kaisha Panoramic image generation in digital photography
US6252601B1 (en) * 1997-09-19 2001-06-26 Nec Corporation Tetrahedral mesh generation and recording medium storing program therefor
US6292191B1 (en) * 1996-12-30 2001-09-18 Cirrus Logic, Inc. Dynamically selectable MIP map blending for a software graphics engine
US6323858B1 (en) * 1998-05-13 2001-11-27 Imove Inc. System for digitally capturing and recording panoramic movies
US6337683B1 (en) * 1998-05-13 2002-01-08 Imove Inc. Panoramic movies which simulate movement through multidimensional space
US6396639B1 (en) * 1999-02-04 2002-05-28 Olympus Optical Co., Ltd. Viewing optical system and image display apparatus using the same
US6469710B1 (en) * 1998-09-25 2002-10-22 Microsoft Corporation Inverse texture mapping using weighted pyramid blending
US6518963B1 (en) * 1998-07-20 2003-02-11 Geometrix, Inc. Method and apparatus for generating patches from a 3D mesh model
US6522325B1 (en) * 1998-04-02 2003-02-18 Kewazinga Corp. Navigable telepresence method and system utilizing an array of cameras
US6559846B1 (en) * 2000-07-07 2003-05-06 Microsoft Corporation System and process for viewing panoramic video
US6639596B1 (en) * 1999-09-20 2003-10-28 Microsoft Corporation Stereo reconstruction from multiperspective panoramas
US6665003B1 (en) * 1998-09-17 2003-12-16 Issum Research Development Company Of The Hebrew University Of Jerusalem System and method for generating and displaying panoramic images and movies
US20040001138A1 (en) * 2002-06-27 2004-01-01 Weerashinghe W.A. Chaminda P. Stereoscopic panoramic video generation system
US6831643B2 (en) * 2001-04-16 2004-12-14 Lucent Technologies Inc. Method and system for reconstructing 3D interactive walkthroughs of real-world environments
US6947059B2 (en) * 2001-08-10 2005-09-20 Micoy Corporation Stereoscopic panoramic image capture device
US7015954B1 (en) * 1999-08-09 2006-03-21 Fuji Xerox Co., Ltd. Automatic video system using multiple cameras
US7095418B2 (en) * 2003-10-30 2006-08-22 Sensable Technologies, Inc. Apparatus and methods for texture mapping
US7280113B1 (en) * 2000-01-11 2007-10-09 Intel Corporation Multiple texture compositing

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4825235A (en) * 1986-05-09 1989-04-25 Nikon Corporation Camera having soft focus filter
JPH07140569A (en) 1993-11-16 1995-06-02 Shunichi Kiwada Stereoscopic image photographing method, prism used therefor and stereoscopic image photographing device
US5703961A (en) 1994-12-29 1997-12-30 Worldscape L.L.C. Image transformation and synthesis methods
JP2826477B2 (en) * 1995-02-14 1998-11-18 義仁 下向 Kaleidoscope, pattern making apparatus and pattern making method using the kaleidoscope
US6552744B2 (en) 1997-09-26 2003-04-22 Roxio, Inc. Virtual reality camera
DE10030196C2 (en) 2000-06-22 2002-05-08 4D Vision Gmbh Arrangement and method for recording multiple views of a scene or an object
KR20070007059A (en) 2003-12-26 2007-01-12 미코이 코포레이션 Multi-dimensional imaging apparatus, systems, and methods
US7796152B2 (en) 2005-04-04 2010-09-14 Micoy Corporation Multi-dimensional imaging
US7872665B2 (en) 2005-05-13 2011-01-18 Micoy Corporation Image capture and processing
US7656403B2 (en) * 2005-05-13 2010-02-02 Micoy Corporation Image processing and display

Patent Citations (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3357770A (en) * 1961-10-02 1967-12-12 Intermountain Res And Engineer Stereoscopic viewing apparatus which includes a curved lenticular screen in front ofa curved picture supporting surface
US3187339A (en) * 1961-10-16 1965-06-01 Wallace A Clay Three dimensional viewing apparatus
US3953869A (en) * 1974-09-24 1976-04-27 Dimensional Development Corporation Stereoscopic photography apparatus
US4063265A (en) * 1974-09-24 1977-12-13 Dimensional Development Corporation Apparatus for taking stereoscopic pictures
US4475798A (en) * 1977-12-27 1984-10-09 The Three Dimensional Photography Corporation Camera for stereoscopic photography
US4214821A (en) * 1978-10-30 1980-07-29 Termes Richard A Total environment photographic mount and photograph
US4525038A (en) * 1981-05-20 1985-06-25 Mueller Rolf Fish eye lens system
US4868682A (en) * 1986-06-27 1989-09-19 Yamaha Corporation Method of recording and reproducing video and sound information using plural recording devices and plural reproducing devices
US5023725A (en) * 1989-10-23 1991-06-11 Mccutchen David Method and apparatus for dodecahedral imaging system
US5130794A (en) * 1990-03-29 1992-07-14 Ritchey Kurtis J Panoramic display system
US5668595A (en) * 1992-03-23 1997-09-16 Canon Kabushiki Kaisha Multi-lens imaging apparatus having a mechanism for combining a plurality of images without displacement of registration
US5650813A (en) * 1992-11-20 1997-07-22 Picker International, Inc. Panoramic time delay and integration video camera system
US5973726A (en) * 1993-09-24 1999-10-26 Canon Kabushiki Kaisha Panoramic image processing apparatus
US5980256A (en) * 1993-10-29 1999-11-09 Carmein; David E. E. Virtual reality system with enhanced sensory apparatus
US6002430A (en) * 1994-01-31 1999-12-14 Interactive Pictures Corporation Method and apparatus for simultaneous capture of a spherical image
US6236748B1 (en) * 1994-08-02 2001-05-22 Canon Kabushiki Kaisha Compound eye image pickup device utilizing plural image sensors and plural lenses
US5946077A (en) * 1994-09-06 1999-08-31 Herman D. Mims Method and apparatus for improved three dimensional photography
US5562572A (en) * 1995-03-10 1996-10-08 Carmein; David E. E. Omni-directional treadmill
US5703604A (en) * 1995-05-22 1997-12-30 Dodeca Llc Immersive dodecaherdral video viewing system
US5657073A (en) * 1995-06-01 1997-08-12 Panoramic Viewing Systems, Inc. Seamless multi-camera panoramic imaging with distortion correction and selectable field of view
US5943056A (en) * 1995-07-11 1999-08-24 Fujitsu Ltd. Interference checking method
US6683608B2 (en) * 1995-11-02 2004-01-27 Imove Inc. Seaming polygonal projections from subhemispherical imagery
US6320584B1 (en) * 1995-11-02 2001-11-20 Imove Inc. Method and apparatus for simulating movement in multidimensional space with polygonal projections from subhemispherical imagery
US6031540A (en) * 1995-11-02 2000-02-29 Imove Inc. Method and apparatus for simulating movement in multidimensional space with polygonal projections from subhemispherical imagery
US6141034A (en) * 1995-12-15 2000-10-31 Immersive Media Co. Immersive imaging method and apparatus
US5822117A (en) * 1996-01-22 1998-10-13 Kleinberger; Paul Systems for three-dimensional viewing including first and second light polarizing layers
US6243103B1 (en) * 1996-05-28 2001-06-05 Canon Kabushiki Kaisha Panoramic image generation in digital photography
US6075905A (en) * 1996-07-17 2000-06-13 Sarnoff Corporation Method and apparatus for mosaic image construction
US5721585A (en) * 1996-08-08 1998-02-24 Keast; Jeffrey D. Digital video panoramic image capture and display system
US6108005A (en) * 1996-08-30 2000-08-22 Space Corporation Method for producing a synthesized stereoscopic image
US6292191B1 (en) * 1996-12-30 2001-09-18 Cirrus Logic, Inc. Dynamically selectable MIP map blending for a software graphics engine
US6097854A (en) * 1997-08-01 2000-08-01 Microsoft Corporation Image mosaic construction system and apparatus with patch-based alignment, global block adjustment and pair-wise motion-based local warping
US6016151A (en) * 1997-09-12 2000-01-18 Neomagic Corp. 3D triangle rendering by texture hardware and color software using simultaneous triangle-walking and interpolation for parallel operation
US6252601B1 (en) * 1997-09-19 2001-06-26 Nec Corporation Tetrahedral mesh generation and recording medium storing program therefor
US6112033A (en) * 1997-10-29 2000-08-29 Canon Kabushiki Kaisha Multiple-lens image pickup apparatus for viewing in stereoscopic panoramic, and ordinary photographing modes
US6034716A (en) * 1997-12-18 2000-03-07 Whiting; Joshua B. Panoramic digital camera system
US6522325B1 (en) * 1998-04-02 2003-02-18 Kewazinga Corp. Navigable telepresence method and system utilizing an array of cameras
US6323858B1 (en) * 1998-05-13 2001-11-27 Imove Inc. System for digitally capturing and recording panoramic movies
US6654019B2 (en) * 1998-05-13 2003-11-25 Imove, Inc. Panoramic movie which utilizes a series of captured panoramic images to display movement as observed by a viewer looking in a selected direction
US6337683B1 (en) * 1998-05-13 2002-01-08 Imove Inc. Panoramic movies which simulate movement through multidimensional space
US6518963B1 (en) * 1998-07-20 2003-02-11 Geometrix, Inc. Method and apparatus for generating patches from a 3D mesh model
US6665003B1 (en) * 1998-09-17 2003-12-16 Issum Research Development Company Of The Hebrew University Of Jerusalem System and method for generating and displaying panoramic images and movies
US6469710B1 (en) * 1998-09-25 2002-10-22 Microsoft Corporation Inverse texture mapping using weighted pyramid blending
US6023588A (en) * 1998-09-28 2000-02-08 Eastman Kodak Company Method and apparatus for capturing panoramic images with range data
US6396639B1 (en) * 1999-02-04 2002-05-28 Olympus Optical Co., Ltd. Viewing optical system and image display apparatus using the same
US6169858B1 (en) * 1999-07-26 2001-01-02 Eastman Kodak Company Panoramic image capture aid
US7015954B1 (en) * 1999-08-09 2006-03-21 Fuji Xerox Co., Ltd. Automatic video system using multiple cameras
US6639596B1 (en) * 1999-09-20 2003-10-28 Microsoft Corporation Stereo reconstruction from multiperspective panoramas
US7280113B1 (en) * 2000-01-11 2007-10-09 Intel Corporation Multiple texture compositing
US6559846B1 (en) * 2000-07-07 2003-05-06 Microsoft Corporation System and process for viewing panoramic video
US6831643B2 (en) * 2001-04-16 2004-12-14 Lucent Technologies Inc. Method and system for reconstructing 3D interactive walkthroughs of real-world environments
US6947059B2 (en) * 2001-08-10 2005-09-20 Micoy Corporation Stereoscopic panoramic image capture device
US20040001138A1 (en) * 2002-06-27 2004-01-01 Weerashinghe W.A. Chaminda P. Stereoscopic panoramic video generation system
US7095418B2 (en) * 2003-10-30 2006-08-22 Sensable Technologies, Inc. Apparatus and methods for texture mapping

Cited By (109)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10298834B2 (en) 2006-12-01 2019-05-21 Google Llc Video refocusing
US20110001797A1 (en) * 2009-07-02 2011-01-06 Sony Corporation 3-d auto-convergence camera
CN102472959A (en) * 2009-07-02 2012-05-23 索尼公司 3-d auto-convergence camera
US8698878B2 (en) * 2009-07-02 2014-04-15 Sony Corporation 3-D auto-convergence camera
US8878908B2 (en) 2009-07-02 2014-11-04 Sony Corporation 3-D auto-convergence camera
US20110001798A1 (en) * 2009-07-02 2011-01-06 Sony Corporation 3-d auto-convergence camera
US9485495B2 (en) 2010-08-09 2016-11-01 Qualcomm Incorporated Autofocus for stereo images
US10061189B2 (en) 2011-09-06 2018-08-28 King Abdullah University of Science and Technology (KAUST) Stereo-panoramic data
WO2013036569A1 (en) * 2011-09-06 2013-03-14 King Abdullah University Of Science And Technology Stereo-panoramic data
US9438889B2 (en) 2011-09-21 2016-09-06 Qualcomm Incorporated System and method for improving methods of manufacturing stereoscopic image sensors
US10552947B2 (en) 2012-06-26 2020-02-04 Google Llc Depth-based image blurring
US9137517B2 (en) * 2012-10-05 2015-09-15 Blackberry Limited Methods and devices for generating a stereoscopic image
US20140098193A1 (en) * 2012-10-05 2014-04-10 Research In Motion Limited Methods and devices for generating a stereoscopic image
US10165183B2 (en) 2012-10-19 2018-12-25 Qualcomm Incorporated Multi-camera system using folded optics
US9398264B2 (en) 2012-10-19 2016-07-19 Qualcomm Incorporated Multi-camera system using folded optics
US9838601B2 (en) 2012-10-19 2017-12-05 Qualcomm Incorporated Multi-camera system using folded optics
US9094540B2 (en) 2012-12-13 2015-07-28 Microsoft Technology Licensing, Llc Displacing image on imager in multi-lens cameras
US20150138336A1 (en) * 2013-04-18 2015-05-21 Okamoto Glass Co., Ltd. Housing for an underwater video camera
US9829771B2 (en) * 2013-04-18 2017-11-28 Okamoto Glass Co., Ltd. Housing for an underwater video camera
US10334151B2 (en) 2013-04-22 2019-06-25 Google Llc Phase detection autofocus using subaperture images
US10178373B2 (en) 2013-08-16 2019-01-08 Qualcomm Incorporated Stereo yaw correction using autofocus feedback
WO2015054326A1 (en) * 2013-10-11 2015-04-16 Microsoft Corporation Eye tracking
US10754490B2 (en) 2013-10-14 2020-08-25 Microsoft Technology Licensing, Llc User interface for collaborative efforts
US9740361B2 (en) 2013-10-14 2017-08-22 Microsoft Technology Licensing, Llc Group experience user interface
US9720559B2 (en) 2013-10-14 2017-08-01 Microsoft Technology Licensing, Llc Command authentication
US9383550B2 (en) 2014-04-04 2016-07-05 Qualcomm Incorporated Auto-focus in low-profile folded optics multi-camera system
US9374516B2 (en) 2014-04-04 2016-06-21 Qualcomm Incorporated Auto-focus in low-profile folded optics multi-camera system
US9860434B2 (en) 2014-04-04 2018-01-02 Qualcomm Incorporated Auto-focus in low-profile folded optics multi-camera system
US9973680B2 (en) 2014-04-04 2018-05-15 Qualcomm Incorporated Auto-focus in low-profile folded optics multi-camera system
US10909760B2 (en) 2014-04-18 2021-02-02 Magic Leap, Inc. Creating a topological map for localization in augmented or virtual reality systems
US11205304B2 (en) 2014-04-18 2021-12-21 Magic Leap, Inc. Systems and methods for rendering user interfaces for augmented or virtual reality
US9761055B2 (en) 2014-04-18 2017-09-12 Magic Leap, Inc. Using object recognizers in an augmented or virtual reality system
US9766703B2 (en) 2014-04-18 2017-09-19 Magic Leap, Inc. Triangulation of points using known points in augmented or virtual reality systems
US9767616B2 (en) 2014-04-18 2017-09-19 Magic Leap, Inc. Recognizing objects in a passable world model in an augmented or virtual reality system
US10262462B2 (en) 2014-04-18 2019-04-16 Magic Leap, Inc. Systems and methods for augmented and virtual reality
US10115232B2 (en) 2014-04-18 2018-10-30 Magic Leap, Inc. Using a map of the world for augmented or virtual reality systems
US10665018B2 (en) 2014-04-18 2020-05-26 Magic Leap, Inc. Reducing stresses in the passable world model in augmented or virtual reality systems
US10825248B2 (en) * 2014-04-18 2020-11-03 Magic Leap, Inc. Eye tracking systems and method for augmented or virtual reality
US10198864B2 (en) 2014-04-18 2019-02-05 Magic Leap, Inc. Running object recognizers in a passable world model for augmented or virtual reality
US10186085B2 (en) 2014-04-18 2019-01-22 Magic Leap, Inc. Generating a sound wavefront in augmented or virtual reality systems
US10846930B2 (en) 2014-04-18 2020-11-24 Magic Leap, Inc. Using passable world model for augmented or virtual reality
US9852548B2 (en) 2014-04-18 2017-12-26 Magic Leap, Inc. Systems and methods for generating sound wavefronts in augmented or virtual reality systems
US10109108B2 (en) 2014-04-18 2018-10-23 Magic Leap, Inc. Finding new points by render rather than search in augmented or virtual reality systems
US9881420B2 (en) 2014-04-18 2018-01-30 Magic Leap, Inc. Inferential avatar rendering techniques in augmented or virtual reality systems
US10115233B2 (en) 2014-04-18 2018-10-30 Magic Leap, Inc. Methods and systems for mapping virtual objects in an augmented or virtual reality system
US9911234B2 (en) 2014-04-18 2018-03-06 Magic Leap, Inc. User interface rendering in augmented or virtual reality systems
US9911233B2 (en) 2014-04-18 2018-03-06 Magic Leap, Inc. Systems and methods for using image based light solutions for augmented or virtual reality
US9922462B2 (en) 2014-04-18 2018-03-20 Magic Leap, Inc. Interacting with totems in augmented or virtual reality systems
US9928654B2 (en) 2014-04-18 2018-03-27 Magic Leap, Inc. Utilizing pseudo-random patterns for eye tracking in augmented or virtual reality systems
US20150302642A1 (en) * 2014-04-18 2015-10-22 Magic Leap, Inc. Room based sensors in an augmented reality system
US9972132B2 (en) 2014-04-18 2018-05-15 Magic Leap, Inc. Utilizing image based light solutions for augmented or virtual reality
US9984506B2 (en) 2014-04-18 2018-05-29 Magic Leap, Inc. Stress reduction in geometric maps of passable world model in augmented or virtual reality systems
US9996977B2 (en) 2014-04-18 2018-06-12 Magic Leap, Inc. Compensating for ambient light in augmented or virtual reality systems
US10008038B2 (en) 2014-04-18 2018-06-26 Magic Leap, Inc. Utilizing totems for augmented or virtual reality systems
US10043312B2 (en) 2014-04-18 2018-08-07 Magic Leap, Inc. Rendering techniques to find new map points in augmented or virtual reality systems
US10013806B2 (en) 2014-04-18 2018-07-03 Magic Leap, Inc. Ambient light compensation for augmented or virtual reality
US10127723B2 (en) * 2014-04-18 2018-11-13 Magic Leap, Inc. Room based sensors in an augmented reality system
US20150321103A1 (en) * 2014-05-08 2015-11-12 Sony Computer Entertainment Europe Limited Image capture method and apparatus
US9579574B2 (en) * 2014-05-08 2017-02-28 Sony Computer Entertainment Europe Limited Image capture method and apparatus
GB2526263B (en) * 2014-05-08 2019-02-06 Sony Interactive Entertainment Europe Ltd Image capture method and apparatus
US10013764B2 (en) 2014-06-19 2018-07-03 Qualcomm Incorporated Local adaptive histogram equalization
US9733458B2 (en) 2014-06-20 2017-08-15 Qualcomm Incorporated Multi-camera system using folded optics free from parallax artifacts
US9549107B2 (en) 2014-06-20 2017-01-17 Qualcomm Incorporated Autofocus for folded optic array cameras
US9294672B2 (en) * 2014-06-20 2016-03-22 Qualcomm Incorporated Multi-camera system using folded optics free from parallax and tilt artifacts
US9386222B2 (en) 2014-06-20 2016-07-05 Qualcomm Incorporated Multi-camera system using folded optics free from parallax artifacts
US9541740B2 (en) 2014-06-20 2017-01-10 Qualcomm Incorporated Folded optic array camera using refractive prisms
US9854182B2 (en) 2014-06-20 2017-12-26 Qualcomm Incorporated Folded optic array camera using refractive prisms
US10084958B2 (en) 2014-06-20 2018-09-25 Qualcomm Incorporated Multi-camera system using folded optics free from parallax and tilt artifacts
US9843723B2 (en) 2014-06-20 2017-12-12 Qualcomm Incorporated Parallax free multi-camera system capable of capturing full spherical images
US9819863B2 (en) 2014-06-20 2017-11-14 Qualcomm Incorporated Wide field of view array camera for hemispheric and spherical imaging
KR20170074902A (en) * 2014-10-22 2017-06-30 패럴랙스터 Method for collecting image data for producing immersive video and method for viewing a space on the basis of the image data
US10218966B2 (en) 2014-10-22 2019-02-26 Parallaxter Method for collecting image data for producing immersive video and method for viewing a space on the basis of the image data
KR102343678B1 (en) * 2014-10-22 2021-12-27 패럴랙스터 Method for collecting image data for producing immersive video and method for viewing a space on the basis of the image data
WO2016061640A1 (en) * 2014-10-22 2016-04-28 Parallaxter Method for collecting image data for producing immersive video and method for viewing a space on the basis of the image data
CN107079145A (en) * 2014-10-22 2017-08-18 帕拉拉克斯特公司 For collecting the method for the view data for producing immersion video and for the method based on view data viewing space
AU2015336869B2 (en) * 2014-10-22 2019-06-06 Parallaxter Method for collecting image data for producing immersive video and method for viewing a space on the basis of the image data
US9832381B2 (en) 2014-10-31 2017-11-28 Qualcomm Incorporated Optical image stabilization for thin cameras
EP3057066B1 (en) * 2015-02-10 2022-12-14 DreamWorks Animation LLC Generation of three-dimensional imagery from a two-dimensional image using a depth map
US10540818B2 (en) 2015-04-15 2020-01-21 Google Llc Stereo image generation and interactive playback
US10567464B2 (en) 2015-04-15 2020-02-18 Google Llc Video compression with adaptive view-dependent lighting removal
US11328446B2 (en) 2015-04-15 2022-05-10 Google Llc Combining light-field data with active depth data for depth map generation
US10275898B1 (en) 2015-04-15 2019-04-30 Google Llc Wedge-based light-field video capture
US10412373B2 (en) 2015-04-15 2019-09-10 Google Llc Image capture for virtual reality displays
US10419737B2 (en) 2015-04-15 2019-09-17 Google Llc Data structures and delivery methods for expediting virtual reality playback
US10085005B2 (en) * 2015-04-15 2018-09-25 Lytro, Inc. Capturing light-field volume image and video data using tiled light-field cameras
US20160307372A1 (en) * 2015-04-15 2016-10-20 Lytro, Inc. Capturing light-field volume image and video data using tiled light-field cameras
US10469873B2 (en) 2015-04-15 2019-11-05 Google Llc Encoding and decoding virtual reality video
US10341632B2 (en) 2015-04-15 2019-07-02 Google Llc. Spatial random access enabled video system with a three-dimensional viewing volume
US10565734B2 (en) 2015-04-15 2020-02-18 Google Llc Video capture, processing, calibration, computational fiber artifact removal, and light-field pipeline
US10546424B2 (en) 2015-04-15 2020-01-28 Google Llc Layered content delivery for virtual and augmented reality experiences
US10375381B2 (en) 2015-05-27 2019-08-06 Google Llc Omnistereo capture and render of panoramic virtual reality content
US20170363949A1 (en) * 2015-05-27 2017-12-21 Google Inc Multi-tier camera rig for stereoscopic image capture
US10038887B2 (en) 2015-05-27 2018-07-31 Google Llc Capture and render of panoramic virtual reality content
US10205896B2 (en) 2015-07-24 2019-02-12 Google Llc Automatic lens flare detection and correction for light-field images
WO2017109274A1 (en) * 2015-12-23 2017-06-29 Nokia Technologies Oy Methods and apparatuses relating to the handling of a plurality of content streams
US11252451B2 (en) * 2015-12-23 2022-02-15 Nokia Technologies Oy Methods and apparatuses relating to the handling of a plurality of content streams
US10275892B2 (en) 2016-06-09 2019-04-30 Google Llc Multi-view scene segmentation and propagation
CN109361912A (en) * 2016-08-17 2019-02-19 谷歌有限责任公司 Multilayer camera apparatus for 3 D visual image capture
WO2018035347A1 (en) * 2016-08-17 2018-02-22 Google Llc Multi-tier camera rig for stereoscopic image capture
US10679361B2 (en) 2016-12-05 2020-06-09 Google Llc Multi-view rotoscope contour propagation
US10594945B2 (en) 2017-04-03 2020-03-17 Google Llc Generating dolly zoom effect using light field image data
US10444931B2 (en) 2017-05-09 2019-10-15 Google Llc Vantage generation and interactive playback
US10440407B2 (en) 2017-05-09 2019-10-08 Google Llc Adaptive control for immersive experience delivery
US10474227B2 (en) 2017-05-09 2019-11-12 Google Llc Generation of virtual reality with 6 degrees of freedom from limited viewer data
US10354399B2 (en) 2017-05-25 2019-07-16 Google Llc Multi-view back-projection to a light-field
US10545215B2 (en) 2017-09-13 2020-01-28 Google Llc 4D camera tracking and optical stabilization
US10965862B2 (en) 2018-01-18 2021-03-30 Google Llc Multi-camera navigation interface
TWI730298B (en) * 2018-02-22 2021-06-11 弗勞恩霍夫爾協會 Generating panoramic images
US11310420B2 (en) 2018-02-22 2022-04-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Generating panoramic images

Also Published As

Publication number Publication date
US7429997B2 (en) 2008-09-30
IL139995A0 (en) 2002-02-10
JP2004514951A (en) 2004-05-20
EP1346259A2 (en) 2003-09-24
IL139995A (en) 2007-07-24
US20040066449A1 (en) 2004-04-08
WO2002044808A3 (en) 2003-04-10
EP1346259A4 (en) 2007-01-17
AU2002218465A1 (en) 2002-06-11
WO2002044808A2 (en) 2002-06-06

Similar Documents

Publication Publication Date Title
US7429997B2 (en) System and method for spherical stereoscopic photographing
US7224382B2 (en) Immersive imaging system
EP1048167B1 (en) System and method for generating and displaying panoramic images and movies
US8004558B2 (en) Stereoscopic wide field of view imaging system
US6947059B2 (en) Stereoscopic panoramic image capture device
EP3206398B1 (en) Stereoscopic camera device
USRE44087E1 (en) Presenting panoramic images with geometric transformation
US20170227841A1 (en) Camera devices with a large field of view for stereo imaging
EP1367819A1 (en) Image input device
EP1178352A1 (en) Method of and apparatus for presenting panoramic images at a local receiver, and a corresponding computer program
JP5483027B2 (en) 3D image measurement method and 3D image measurement apparatus
WO2014162324A1 (en) Spherical omnidirectional video-shooting system
CN1247994A (en) Stereoscopic full views watching system
US9883159B2 (en) Camera to minimize parallax effects
US6836286B1 (en) Method and apparatus for producing images in a virtual space, and image pickup system for use therein
US20210185299A1 (en) A multi-camera device and a calibration method
JP2006515128A (en) Stereo panoramic image capturing device
CN109769111A (en) Image display method, device, system, storage medium and processor
CN108646776B (en) Imaging system and method based on unmanned aerial vehicle
US7154528B2 (en) Apparatus for placing primary image in registration with lenticular lens in system for using binocular fusing to produce secondary 3D image from primary image
Prechtl et al. Stereoscopic wide field of view imaging system
JP2000347133A (en) Stereo camera
JPH1026738A (en) Stereoscope covering major portion of visual field with pictures

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION