Capturing and processing of images using non-monolithic camera arrays
원문보기
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
H04N-005/225
H04N-005/232
H01L-027/146
H04N-003/14
H04N-005/369
H04N-009/04
H04N-013/02
G06T-003/40
G06T-007/00
H04N-005/228
H04N-005/33
H04N-005/262
H04N-005/265
H04N-005/357
H04N-009/083
H04N-001/195
H04N-013/00
출원번호
US-0475449
(2014-09-02)
등록번호
US-9060124
(2015-06-16)
발명자
/ 주소
Venkataraman, Kartik
Jabbi, Amandeep S.
Mullis, Robert H.
출원인 / 주소
Pelican Imaging Corporation
대리인 / 주소
KPPB LLP
인용정보
피인용 횟수 :
57인용 특허 :
115
초록▼
A camera array, an imaging device and/or a method for capturing image that employ a plurality of imagers fabricated on a substrate is provided. Each imager includes a plurality of pixels. The plurality of imagers include a first imager having a first imaging characteristics and a second imager havin
A camera array, an imaging device and/or a method for capturing image that employ a plurality of imagers fabricated on a substrate is provided. Each imager includes a plurality of pixels. The plurality of imagers include a first imager having a first imaging characteristics and a second imager having a second imaging characteristics. The images generated by the plurality of imagers are processed to obtain an enhanced image compared to images captured by the imagers. Each imager may be associated with an optical element fabricated using a wafer level optics (WLO) technology.
대표청구항▼
1. A camera array, comprising: a plurality of cameras configured to capture images of a scene, where each camera is formed on a separate semiconductor substrate and comprises: optics comprising at least one lens element and at least one aperture where the optics are configured so that each camera ha
1. A camera array, comprising: a plurality of cameras configured to capture images of a scene, where each camera is formed on a separate semiconductor substrate and comprises: optics comprising at least one lens element and at least one aperture where the optics are configured so that each camera has a field of view that captures a sub-pixel shifted view of the same scene; anda sensor comprising a two dimensional array of pixels and control circuitry for controlling imaging parameters; andat least one spectral filter located within the camera, where each spectral filter is configured to pass a specific spectral band of light;a controller configured to control operation parameters of the plurality of cameras; andan image processing pipeline module;wherein the plurality of cameras comprises a plurality of different types of camera, where cameras of a given type share the same imaging characteristics and different types of cameras have different imaging characteristics including using different types of filters; andwherein the plurality of cameras includes a central camera and the plurality of cameras forms: at least one row of cameras above the central camera;at least one row of cameras below the central camera;at least one column of cameras to the left of the central camera; andat least one column of cameras to the right of the central camera;wherein cameras having the same imaging characteristics and filter are located in locations selected from the group consisting of: locations including at least one row in the camera array above the central camera and at least one row in the camera array below the central camera; and locations including at least one column in the camera array to the left of the central camera and at least one column in the camera array to the right of the central camera. 2. The camera array of claim 1, wherein the plurality of cameras are mounted to a motherboard. 3. The camera array of claim 2, wherein the motherboard is flexible. 4. The camera array of claim 1, wherein the camera array comprises an array of between 2×2 and 6×6 cameras. 5. The camera array of claim 1, wherein the control circuitry of the cameras including different spectral filters configures the cameras to operate with at least one difference in operating parameters. 6. The camera array of claim 5, wherein the at least one difference in operating parameters includes at least one imaging parameter selected from the group consisting of exposure time, gain, and black level offset. 7. The camera array of claim 1, further comprising: a controller configured to control operation parameters of the plurality of cameras; andan image processing pipeline module comprising a parallax confirmation and measurement module;wherein images captured by the plurality of cameras include different occlusions sets, where the occlusion set of a first camera is the portion of a scene visible to a second camera in the plurality of cameras that is occluded from the view of the first camera; andwherein the parallax confirmation and measurement module is configured to measure parallax using images captured by the plurality of cameras by: detecting parallax-induced changes that are consistent across all of the cameras in the two dimensional array of cameras taking into account the position of the cameras; andignoring pixels in the images captured by the plurality of cameras that are in the occlusion set of the central camera. 8. The camera array of claim 7, wherein the parallax confirmation and measurement module is further configured to estimate distance to an object in images captured by the plurality of cameras. 9. The camera array of claim 7, wherein the parallax confirmation and measurement module is further configured to generate a depth map using images captured by the plurality of cameras. 10. The camera array of claim 9, wherein the image processing pipeline module further comprises a super-resolution processing module configured to generate at least one higher resolution image using images captured by the plurality of cameras and parallax measurements from the parallax confirmation and measurement module to compensate for parallax in the captured images. 11. The camera array of claim 10, wherein generating at least one higher resolution image using images captured by the plurality of cameras and parallax measurements comprises: aligning portions of different images to compensate for parallax based upon the measured parallax; andperforming at least one process selected from the group consisting of: fusing the aligned image portions to obtain a higher resolution image; andperforming super-resolution processing on the aligned image portions to synthesize a super-resolution image. 12. The camera array of claim 7, wherein the super-resolution processing module is configured to select at least one distance as a focal plane and to apply blurring to pixels in an image produced by the camera array with depths in the depth map that are not proximate a focal plane. 13. A camera array, comprising: a plurality of cameras configured to capture images of a scene, where each camera is formed on a separate semiconductor substrates and comprises: optics comprising at least one lens element and at least one aperture where the optics are configured so that each camera has a field of view that captures a sub-pixel shifted view of the same scene; anda sensor comprising a two dimensional array of pixels and control circuitry for controlling imaging parameters; andat least one spectral filter located within the camera, where each spectral filter is configured to pass a specific spectral band of light;a controller configured to control operation parameters of the plurality of cameras;an image processing pipeline module comprising a parallax confirmation and measurement module and a super-resolution processing module;wherein the plurality of cameras comprises a plurality of different types of camera, where cameras of a given type share the same imaging characteristics and different types of cameras have different imaging characteristics including using different types of filters; andwherein the plurality of cameras includes a central camera and the plurality of cameras forms: at least one row of cameras above the central camera;at least one row of cameras below the central camera;at least one column of cameras to the left of the central camera; andat least one column of cameras to the right of the central camera;wherein cameras having the same imaging characteristics and filter are located in locations selected from the group consisting of: locations including at least one row in the camera array above the central camera and at least one row in the camera array below the central camera; andlocations including at least one column in the camera array to the left of the central camera and at least one column in the camera array to the right of the central camera; andwherein the parallax confirmation and measurement module is configured to measure parallax using images captured by the plurality of cameras by: detecting parallax-induced changes that are consistent across all of the cameras in the two dimensional array of cameras taking into account the position of the cameras; andignoring pixels in the images captured by the plurality of cameras that are in the occlusion set of the central camera;wherein the parallax confirmation and measurement module is further configured to generate a depth map using images captured by the plurality of cameras; andwherein the super-resolution processing module is configured to generate at least one higher resolution super-resolved image using images captured by the plurality of cameras and parallax measurements from the parallax confirmation and measurement module to compensate for parallax in the captured images. 14. A linear camera array, comprising: a plurality of cameras that form a linear array of cameras configured to capture images of a scene, where each camera is formed on a separate semiconductor substrate and comprises:optics comprising at least one lens element and at least one aperture where the optics are configured so that each camera has a field of view that captures a sub-pixel shifted view of the same scene; anda sensor comprising a two dimensional array of pixels and control circuitry for controlling imaging parameters; andat least one spectral filter located within the camera, where each spectral filler is configured to pass a specific spectral band of light; andan image processing pipeline module;wherein the plurality of cameras comprises a plurality of different types of camera, where cameras of a given type share the same imaging characteristics and different types of cameras have different imaging characteristics including using different types of filters;wherein the linear array of cameras includes a central camera; andwherein cameras having the same imaging characteristics and filter are located on either side of the central camera;a controller configured to control operation parameters of the linear camera array; andan image processing pipeline module comprising a parallax confirmation and measurement module;wherein images captured by the plurality of cameras include different occlusions sets, where the occlusion set of a first camera is the portion of a scene visible to a second camera in the plurality of cameras that is occluded from the view of the first camera; andwherein the parallax confirmation and measurement module is configured to measure parallax using images captured by the plurality of cameras by:detecting parallax-induced changes that are consistent across all of the cameras in the linear array of cameras taking into account the position of the cameras; andignoring pixels in the images captured by the plurality of cameras that are in the occlusion set of the central camera. 15. The camera array of claim 14, wherein the plurality of cameras are mounted to a motherboard. 16. The camera array of claim 14, wherein the motherboard is flexible. 17. The camera array of claim 14, wherein the camera array comprises at least one selected from the group consisting of: a 1×4 array of imagers; and a 1×10 array of imagers. 18. The camera array of claim 14, wherein the parallax confirmation and measurement module is further configured to estimate distance to an object in a scene using images captured by the plurality of cameras. 19. The camera array of claim 14, wherein the parallax confirmation and measurement module is further configured to generate a depth map using images captured by the plurality of cameras. 20. The camera array of claim 19, wherein the super-resolution processing module is configured to select at least one distance as a focal plane and to apply blurring to pixels in at least one image produced by the camera array with depths in the depth map that are not proximate a focal plane. 21. The camera array of claim 14, wherein the image processing pipeline module further comprises a super-resolution processing module configured to generate at least one higher resolution image using images captured by the plurality of cameras and parallax measurements from the parallax confirmation and measurement module to compensate for parallax in the captured images. 22. The camera array of claim 21, wherein generating at least one higher resolution image using images captured by the plurality of cameras and parallax measurements comprises: aligning portions of different images to compensate for parallax based upon the measured parallax; andperforming at least one process selected from the group consisting of: fusing the aligned image portions to obtain a higher resolution image; andperforming super-resolution processing on the aligned image portions to synthesize a super-resolution image.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (115)
Wilburn, Bennett; Joshi, Neel; Levoy, Marc C.; Horowitz, Mark, Apparatus and method for capturing a scene using staggered triggering of dense camera arrays.
Boisvert, David Michael; McMahon, Andrew Kenneth John, CCD output processing stage that amplifies signals from colored pixels based on the conversion efficiency of the colored pixels.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Capturing and processing of images using monolithic camera array with heterogeneous imagers.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images using monolithic camera array with heterogeneous imagers.
Yamashita,Syugo; Murata,Haruhiko; Iinuma,Toshiya; Nakashima,Mitsuo; Mori,Takayuki, Device and method for converting two-dimensional video to three-dimensional video.
Ward, Gregory John; Seetzen, Helge; Heidrich, Wolfgang, Electronic camera having multiple sensors for capturing high dynamic range images and related methods.
Abell Gurdon R. (West Woodstock CT) Cook Francis J. (Topsfield MA) Howes Peter D. (Sudbury MA), Method and apparatus for arraying image sensor modules.
Sawhney,Harpreet Singh; Tao,Hai; Kumar,Rakesh; Hanna,Keith, Method and apparatus for synthesizing new video and/or still imagery from a collection of real video and/or still imagery.
Alexander David H. (Santa Monica CA) Hershman George H. (Carlsbad CA) Jack Michael D. (Carlsbad CA) Koda N. John (Vista CA) Lloyd Randahl B. (San Marcos CA), Monolithic imager for near-IR.
Hornbaker ; III Cecil V. (New Carrolton MD) Driggers Thomas C. (Falls Church VA) Bindon Edward W. (Fairfax VA), Scanning apparatus using multiple CCD arrays and related method.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for parallax detection and correction in images captured using array cameras that contain occlusions using subsets of images to perform depth estimation.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for performing depth estimation using image data from multiple spectral channels.
Ludwig, Lester F., Vignetted optoelectronic array for use in synthetic image formation via signal processing, lensless cameras, and integrated camera-displays.
Rieger Albert,DEX ; Barclay David ; Chapman Steven ; Kellner Heinz-Andreas,DEX ; Reibl Michael,DEX ; Rydelek James G. ; Schweizer Andreas,DEX, Watertight body for accommodating a photographic camera.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images including occlusions focused on an image sensor by a lens stack array.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H.; Duparre, Jacques; Hu, Shane Ching-Feng, Capturing and processing of images using camera array incorperating Bayer cameras having different fields of view.
Srikanth, Manohar; Ramamoorthi, Ravi; Venkataraman, Kartik; Chatterjee, Priyam, System and methods for depth regularization and semiautomatic interactive matting using RGB-D images.
Nayar, Shree; Venkataraman, Kartik; Pain, Bedabrata; Lelescu, Dan, Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures.
Lelescu, Dan; Venkataraman, Kartik, Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing.
Duparre, Jacques; McMahon, Andrew Kenneth John; Lelescu, Dan; Venkataraman, Kartik; Molina, Gabriel, Systems and methods for detecting defective camera arrays and optic arrays.
Ciurea, Florian; Venkataraman, Kartik; Molina, Gabriel; Lelescu, Dan, Systems and methods for estimating depth and visibility from a reference viewpoint for pixels in a set of images captured from different viewpoints.
Venkataraman, Kartik; Lelescu, Dan; Molina, Gabriel, Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information.
Venkataraman, Kartik; Lelescu, Dan; Molina, Gabriel, Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using a camera arrays incorporating monochrome and color cameras.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for generating depth maps using images captured by camera arrays incorporating cameras having different fields of view.
Duparre, Jacques; McMahon, Andrew Kenneth John; Lelescu, Dan, Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors.
Duparre, Jacques; McMahon, Andrew Kenneth John; Lelescu, Dan, Systems and methods for manufacturing camera modules using active alignment of lens stack arrays and sensors.
Venkataraman, Kartik; Jabbi, Amandeep S.; Mullis, Robert H., Systems and methods for measuring depth using images captured by a camera array including cameras surrounding a central camera.
Venkataraman, Kartik; Huang, Yusong; Jain, Ankit K.; Chatterjee, Priyam, Systems and methods for performing high speed video capture and depth estimation using array cameras.
Lelescu, Dan; Duong, Thang, Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information.
Lelescu, Dan; Molina, Gabriel; Venkataraman, Kartik, Systems and methods for synthesizing high resolution images using images captured by an array of independently controllable imagers.
Venkataraman, Kartik; Nisenzon, Semyon; Chatterjee, Priyam; Molina, Gabriel, Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies.
Venkataraman, Kartik; Nisenzon, Semyon; Chatterjee, Priyam; Molina, Gabriel, Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.