Systems and methods for encoding light field image files
원문보기
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G06K-009/36
G06T-009/00
G06T-009/20
H04N-019/597
H04N-013/00
H04N-013/02
H04N-019/625
H04N-019/136
G06T-003/40
H04N-019/85
G06T-015/08
G06T-007/50
출원번호
US-0667495
(2015-03-24)
등록번호
US-9811753
(2017-11-07)
발명자
/ 주소
Venkataraman, Kartik
Nisenzon, Semyon
Lelescu, Dan
출원인 / 주소
FotoNation Cayman Limited
대리인 / 주소
KPPB LLP
인용정보
피인용 횟수 :
16인용 특허 :
256
초록▼
Systems and methods configured to store images synthesized from light field image data and metadata describing the images in electronic files and render images using the stored image and the metadata in accordance with embodiments of the invention are disclosed. One embodiment includes a processor a
Systems and methods configured to store images synthesized from light field image data and metadata describing the images in electronic files and render images using the stored image and the metadata in accordance with embodiments of the invention are disclosed. One embodiment includes a processor and memory containing an encoding application and light field image data, where the light field image data comprises a plurality of low resolution images of a scene captured from different viewpoints. In addition, the encoding application configures the processor to synthesize a higher resolution image of the scene from a reference viewpoint using the low resolution images, where synthesizing the higher resolution image involves creating a depth map that specifies depths from the reference viewpoint for pixels in the higher resolution image; encode the higher resolution image; and create a light field image file including the encoded image and metadata including the depth map.
대표청구항▼
1. An image processing system, comprising: a processor; andmemory containing an encoding application;wherein the encoding application configures the processor to: control capture of a plurality of images of a scene from different viewpoints by an array of cameras;generate a depth map that specifies
1. An image processing system, comprising: a processor; andmemory containing an encoding application;wherein the encoding application configures the processor to: control capture of a plurality of images of a scene from different viewpoints by an array of cameras;generate a depth map that specifies depths for pixels from a reference viewpoint using the plurality of captured images;synthesize an image of the scene from the reference viewpoint using the plurality of captured images, where synthesizing the image involves using the depth map that specifies depths from the reference viewpoint for pixels in the synthesized image;encode the synthesized image; andstore the synthesized image in an image file, and store the depth map and at least one of the plurality of captured images in the image file. 2. The system of claim 1, wherein the encoding application configures the processor to encode the depth map and store the depth map as metadata within the image file. 3. The system of claim 2, wherein the metadata also includes a description of a focal plane. 4. The system of claim 1, wherein: synthesizing the image involves identifying pixels in the plurality of images of the scene that are occluded in the reference viewpoint; andwherein the encoding application configure the processor to store metadata describing the encoded image in the image file, wherein the stored metadata includes descriptions of the occluded pixels. 5. The system of claim 4, wherein the descriptions of the occluded pixels include colors, locations, and depths of the occluded pixels. 6. The system of claim 1, wherein: generating a depth map involves creating a confidence map for the depth map, where the confidence map indicates the reliability of the depth value for a pixel in the depth map; andwherein the encoding application configure the processor to store metadata describing the encoded image in the image file, wherein the metadata includes the depth map and the confidence map. 7. The system of claim 6, wherein the encoding application configures the processor to encode the confidence map. 8. The system of claim 1, wherein: the encoding application configures the processor to generate an edge map that indicates pixels in the synthesized image that lie on a discontinuity; andwherein the encoding application configure the processor to store metadata describing the encoded image in the image file, wherein the metadata includes the edge map. 9. The system of claim 8, wherein the edge map identifies whether a pixel lies on an intensity discontinuity. 10. The system of claim 8, wherein the edge map identifies whether a pixel lies on an intensity and depth discontinuity. 11. The system of claim 8, wherein the encoding application configures the processor to encode the edge map. 12. The system of claim 1, wherein: the encoding application configures the processor to generate a missing pixel map that indicates pixels in the synthesized image that do not correspond to a pixel from the plurality of images of the scene and that are generated by interpolating pixel values from adjacent pixels in the synthesized image; andwherein the encoding application configure the processor to store metadata describing the encoded image in the image file, wherein the metadata includes the missing pixels map. 13. The system of claim 12, wherein the encoding application configures the processor to encode the missing pixels map. 14. The system of claim 1, wherein the image file conforms to the JPEG File Interchange Format (JFIF) standard. 15. The system of claim 14, wherein the synthesized image is encoded in accordance with the Joint Photographic Experts Group (JPEG) standard. 16. The system of claim 15, wherein the encoding application configure the processor to store metadata describing the encoded image within an application marker segment within the image file. 17. The system of claim 16, wherein the application marker segment is identified using an APP9 marker. 18. The system of claim 16, wherein the encoding application configures the processor to encode the depth map using lossless compression and the encoded depth map is stored within the application marker segment containing the metadata describing the encoded image. 19. The system of claim 1, wherein the depth map and at least one of the plurality of captured images are stored as metadata in the image file. 20. A method for encoding light field image data as a image file, comprising: capture a plurality of images of a scene from different viewpoints by an array of cameras;generate a depth map that specifies depths for pixels from a reference viewpoint using the plurality of captured images using a processor;synthesize an image of a scene from the reference viewpoint based upon the plurality of captured images using a processor, wherein synthesizing the image involves using the depth map that specifies depths from the reference viewpoint for pixels in the synthesized image;encoding the synthesized image using a processor; andstoring the synthesized image in an image file using a processor, and storing the depth map and at least one of the plurality of captured images in the image file. 21. A non-transitory machine readable medium containing processor instructions, where execution of the instructions by a processor causes the processor to perform a process comprising: controlling capturing of a plurality of images of a scene from different viewpoints by an array of cameras;generating a depth map that specifies depths for pixels from a reference viewpoint using the plurality of captured images;synthesizing an image of a scene from the reference viewpoint using the plurality of captured images, wherein synthesizing the image involves using the depth map that specifies depths from the reference viewpoint for pixels in the synthesized image;encoding the synthesized image; andstoring the synthesized image in an image file, and storing the depth map and at least one of the plurality of captured images in the image file.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (256)
Hines, Stephen P, 3-D motion-parallax portable display software application.
Wilburn, Bennett; Joshi, Neel; Levoy, Marc C.; Horowitz, Mark, Apparatus and method for capturing a scene using staggered triggering of dense camera arrays.
Iwase Toshihiro (Nara JPX) Kanekura Hiroshi (Yamatokouriyama JPX), Apparatus for and method of converting a sampling frequency according to a data driven type processing.
Venkataraman, Kartik; Gallagher, Paul; Jain, Ankit; Nisenzon, Semyon; Lelescu, Dan; Ciurea, Florian; Molina, Gabriel, Array cameras including an array camera module augmented with a separate camera.
Duparre, Jacques; Lelescu, Dan; Venkataraman, Kartik, Array cameras incorporating monolithic array camera modules with high MTF lens stacks for capture of images used in super-resolution processing.
Duparre, Jacques; Lelescu, Dan; Venkataraman, Kartik, Array cameras incorporating optics with modulation transfer functions greater than sensor Nyquist frequency for capture of images used in super-resolution processing.
Boisvert, David Michael; McMahon, Andrew Kenneth John, CCD output processing stage that amplifies signals from colored pixels based on the conversion efficiency of the colored pixels.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images including occlusions captured by arrays of luma and chroma cameras.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images including occlusions captured by camera arrays.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images including occlusions focused on an image sensor by a lens stack array.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images using camera array incorperating Bayer cameras having different fields of view.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Capturing and processing of images using monolithic camera array with heterogeneous imagers.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images using monolithic camera array with heterogeneous imagers.
Yamashita,Syugo; Murata,Haruhiko; Iinuma,Toshiya; Nakashima,Mitsuo; Mori,Takayuki, Device and method for converting two-dimensional video to three-dimensional video.
Pertsel, Shimon; Meitav, Ohad; Pozniansky, Eli; Galil, Erez, Digital camera with selectively increased dynamic range by control of parameters during image acquisition.
Ward, Gregory John; Seetzen, Helge; Heidrich, Wolfgang, Electronic camera having multiple sensors for capturing high dynamic range images and related methods.
Hornback,Bert; Harwood,Doug; Boyd,W. Eric; Carlson,Randy, Imaging device with multiple fields of view incorporating memory-based temperature compensation of an uncooled focal plane array.
Abell Gurdon R. (West Woodstock CT) Cook Francis J. (Topsfield MA) Howes Peter D. (Sudbury MA), Method and apparatus for arraying image sensor modules.
Kim, Yong-tae; Park, Ha-joong; Lee, Gun-ill; Min, Houng-sog; Hong, Sung-bin; Choi, Kwang-cheol, Method and apparatus for providing a multi-view still image service, and method and apparatus for receiving a multi-view still image service.
Sawhney,Harpreet Singh; Tao,Hai; Kumar,Rakesh; Hanna,Keith, Method and apparatus for synthesizing new video and/or still imagery from a collection of real video and/or still imagery.
Burt Peter J. (Mercer County NJ) van der Wal Gooitzen S. (Mercer NJ) Kolczynski Raymond J. (Mercer NJ) Hingorani Rajesh (Mercer NJ), Method for fusing images and apparatus therefor.
Burt Peter J. (Princeton NJ) van der Wal Gooitzen S. (Hopewell Borough ; Mercer County NJ) Kolczynski Raymond J. (Hamilton Township ; Mercer County NJ) Hingorani Rajesh (West Windsor Township ; Merce, Method for fusing images and apparatus therefor.
Han, Hee-chul; Choi, Yang-lim; Cho, Seung-ki, Method of generating image data by an image device including a plurality of lenses and apparatus for generating image data.
Alexander David H. (Santa Monica CA) Hershman George H. (Carlsbad CA) Jack Michael D. (Carlsbad CA) Koda N. John (Vista CA) Lloyd Randahl B. (San Marcos CA), Monolithic imager for near-IR.
Doering, Hans-Joachim; Heinitz, Joachim, Multi-beam modulator for a particle beam and use of the multi-beam modulator for the maskless structuring of a substrate.
Hornbaker ; III Cecil V. (New Carrolton MD) Driggers Thomas C. (Falls Church VA) Bindon Edward W. (Fairfax VA), Scanning apparatus using multiple CCD arrays and related method.
Neal Margulis ; Chad Fogg, System and method for using temporal gamma and reverse super-resolution to process images for use in digital display systems.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, System and methods for measuring depth using an array camera employing a bayer filter.
Lelescu, Dan; Molina, Gabriel; Venkataraman, Kartik, Systems and methods for dynamic refocusing of high resolution images generated using images captured by a plurality of imagers.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints.
Venkataraman, Kartik; Lelescu, Dan; Molina, Gabriel, Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for generating depth maps and corresponding confidence maps indicating depth estimation reliability.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Systems and methods for generating depth maps using a set of images containing a baseline image.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using images captured by camera arrays incorporating cameras having different fields of view.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using light focused on an image sensor by a lens element array.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for measuring depth in the presence of occlusions using a subset of images.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for measuring depth using an array of independently controllable cameras.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for measuring depth using images captured by a camera array including cameras surrounding a central camera.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for parallax detection and correction in images captured using array cameras that contain occlusions using subsets of images to perform depth estimation.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for parallax measurement using camera arrays incorporating 3 x 3 camera configurations.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for performing depth estimation using image data from multiple spectral channels.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for performing post capture refocus using images captured by camera arrays.
Lelescu, Dan; Duong, Thang, Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information.
Lelescu, Dan; Molina, Gabriel; Venkataraman, Kartik, Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Systems and methods for synthesizing higher resolution images using a set of images containing a baseline image.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for synthesizing higher resolution images using images captured by camera arrays.
Venkataraman, Kartik; Nisenzon, Semyon; Chatterjee, Priyam; Molina, Gabriel, Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies.
Ludwig, Lester F., Vignetted optoelectronic array for use in synthetic image formation via signal processing, lensless cameras, and integrated camera-displays.
Rieger Albert,DEX ; Barclay David ; Chapman Steven ; Kellner Heinz-Andreas,DEX ; Reibl Michael,DEX ; Rydelek James G. ; Schweizer Andreas,DEX, Watertight body for accommodating a photographic camera.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images including occlusions focused on an image sensor by a lens stack array.
Srikanth, Manohar; Ramamoorthi, Ravi; Venkataraman, Kartik; Chatterjee, Priyam, System and methods for depth regularization and semiautomatic interactive matting using RGB-D images.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints.
Venkataraman, Kartik; Lelescu, Dan; Molina, Gabriel, Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras.
Duparre, Jacques; McMahon, Andrew Kenneth John; Lelescu, Dan, Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.