[미국특허]
Device and algorithm for capturing high dynamic range (HDR) video
원문보기
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
H04N-005/355
H04N-005/232
H04N-005/235
G06T-001/20
G06T-005/50
G06T-005/00
G06T-007/00
출원번호
US-0540476
(2012-07-02)
등록번호
US-9489706
(2016-11-08)
발명자
/ 주소
Levy, Noam
Rapaport, Guy
출원인 / 주소
QUALCOMM Technologies, Inc.
대리인 / 주소
Knobbe Martens Olson & Bear LLP
인용정보
피인용 횟수 :
0인용 특허 :
5
초록▼
Embodiments are directed towards enabling digital cameras to digitally process a captured a Low Dynamic Range image sequence at a real time video rate, and to convert the image sequence into an High Dynamic Range (HDR) image sequence using a pipelined architecture. Two or more image frames are captu
Embodiments are directed towards enabling digital cameras to digitally process a captured a Low Dynamic Range image sequence at a real time video rate, and to convert the image sequence into an High Dynamic Range (HDR) image sequence using a pipelined architecture. Two or more image frames are captured using different exposure settings and then combined to form a single HDR output frame in a video sequence. The pipelined architecture operate on adjacent image frames by performing an image alignment, an image mixing, and a tone mapping on the adjacent image frames to generate the HDR image sequence.
대표청구항▼
1. A processor based method for pipelining a plurality of actions at a real-time video rate to generate a high dynamic range (HDR) video output stream, comprising: capturing a plurality of images as low dynamic range images, at least one image having a different exposure setting than at least one ot
1. A processor based method for pipelining a plurality of actions at a real-time video rate to generate a high dynamic range (HDR) video output stream, comprising: capturing a plurality of images as low dynamic range images, at least one image having a different exposure setting than at least one other image in the plurality of images;performing image alignment between the images in the plurality of images;determining misaligned pixels between the images via comparing the difference between corresponding pixels of the images to a threshold;creating an HDR image by combining each of the aligned images with each other aligned image based on at least a pixel-wise blending, the blending comprising pushing a weighting value closer to one of first and second whole numbers closest to the weighting value when the difference between the corresponding pixels of the images is greater than the threshold, the first whole number being less than the weighting value and the second whole number being greater than the weighting value, the pushing comprising: i) pushing the weighting value closer to the first whole number when the weighting value is closer to the first whole number than the second whole number, and ii) pushing the weighting value closer to the second whole number when the weighting value is closer to the second whole number than the first whole number, the pushed weighting value being applied to the misaligned pixels in a selected one of the images being blended; andencoding the HDR image into a HDR video output stream. 2. The processor based method of claim 1, wherein performing image alignment further comprises: performing a motion vector extraction between corresponding pixels between the images; andemploying a result of the motion vector extraction transforming the images onto image coordinates of at least one of the images in the plurality of images. 3. The processor based method of claim 2, wherein the motion vector extraction further comprises: employing a motion search using a motion of a block of pixels in at least one of the images to identify another block of pixels in at least one other image in the plurality based on when a mean reduced sum of the absolute differences (MRSAD) is determined to be minimal;using a result of the motion search to identify a set of correspondences between pixels in at least one of the images with pixels in at least one other image; andperforming a global motion estimation on the set of correspondences to generate an estimated global transformation. 4. The processor based method of claim 3, wherein the motion search further includes performing a corner detection on at least one of the images and at least one other image in the plurality. 5. The processor based method of claim 1, wherein combining the aligned images further comprises: for each image in the plurality that is represented in a gamma non-linear domain, performing a de-gamma transformation on the image;normalizing a respective image having a brighter exposure setting than another darker image in the plurality, to an exposure value of the other darker image;determining weighting values based in part on luminosity values of the image having the brighter exposure setting; andusing the weighting values to blend dark pixels from the brighter image with bright pixels from the darker image to generate the HDR image. 6. The processor based method of claim 1, wherein pipelining a plurality of actions, further comprises: performing a tone mapping to reduce a number of bits used to represent a color component in the HDR image, the tone mapping including: performing a gamma global mapping on the HDR image; andperforming a non-linear local mapping that maps a pixel based on a neighbor of the pixels' average luminosity value. 7. The processor based method of claim 1, wherein the different exposure setting is different in at least one of an exposure time, an f-number, or an International for Standardization (ISO) level. 8. An image system, comprising: an image sensing device configured to receive a plurality of images as low dynamic range images; andone or more circuits having a plurality of components thereon operating in a pipeline architecture to perform a plurality of actions at a real-time video rate for a sequence of captured images, including: receiving the plurality of images, at least one image captured with a different exposure setting than at least one other image in the plurality of images;performing image alignment between the images;determining misaligned pixels between the images via comparing the difference between corresponding pixels of the images to a threshold;creating a high dynamic range (HDR) image by combining each of the aligned images with each other aligned image using a pixel-wise blending, the blending comprising pushing a weighting value closer to one of first and second whole numbers closest to the weighting value when the difference between the corresponding pixels of the images is greater than the threshold, the first whole number being less than the weighting value and the second whole number being greater than the weighting value, the pushing comprising: i) pushing the weighting value closer to the first whole number when the weighting value is closer to the first whole number than the second whole number and ii) pushing the weighting value closer to the second whole number when the weighting value is closer to the second whole number than the first whole number, the pushed weighting value being applied to the misaligned pixels in a selected one of the images being blended; andencoding the HDR image into an HDR video output stream. 9. The image system of claim 8, wherein performing image alignment further comprises: performing a motion vector extraction between corresponding pixels between the images; andemploying a result of the motion vector extraction transforming the images onto image coordinates of at least one image in the plurality. 10. The image system of claim 9, wherein the motion vector extraction further comprises: employing a motion search using a motion of a block of pixels in at least one of the images to identify another block of pixels in at least one other image based on when a mean reduced sum of the absolute differences (MRSAD) is determined to be minimal;using a result of the motion search to identify a set of correspondences between pixels in the images; andperforming a global motion estimation on the set of correspondences to generate an estimated global transformation. 11. The image system of claim 10, wherein the motion search further includes performing a corner detection on the images. 12. The image system of claim 8, wherein blending further comprises for saturated areas within one of the images, using pixels from a darker image in the plurality to recover burned out details in a scene when creating the HDR image. 13. The image system of claim 8, wherein the different exposure setting is different in at least one of an exposure time, an f-number, or an International for Standardization (ISO) level. 14. The image system of claim 8, wherein the pipelined plurality of actions, further comprises: performing a tone mapping to reduce a number of bits used to represent a color component in the HDR image, the tone mapping including: performing a gamma global mapping on the HDR image; andperforming a non-linear local mapping that maps a pixel based on a neighbor of the pixels' average luminosity value. 15. A storage device having stored thereon a plurality of computer-executable instructions that when installed on a digital camera performed actions, comprising: iteratively performing within a pipelined architecture using the computer-executable instructions to perform actions for images within a sequence of images, including: capturing a first image with a first exposure setting;capturing a second image with a second exposure setting that is different from the first exposure setting, the first image and the second image being captured as low dynamic range images;performing image alignment between the first image and the second image;determining misaligned pixels between the first image and the second image via comparing the difference between corresponding pixels of the first and second images to a threshold;creating a high dynamic range (HDR) image by blending the aligned first image and the second image using a pixel-wise blending, the blending comprising pushing a weighting value closer to one of first and second whole numbers closest to the weighting value when the difference between the corresponding pixels of the first and second images is greater than the threshold, the first whole number being less than the weighting value and the second whole number being greater than the weighting value, the pushing comprising: i) pushing the weighting value closer to the first whole number when the weighting value is closer to the first whole number than the second whole number and ii) pushing the weighting value closer to the second whole number when the weighting value is closer to the second whole number than the first whole number, the pushed weighting value being applied to the misaligned pixels in a selected one of the images being blended; andencoding the HDR image into an HDR video output stream. 16. The storage device of claim 15, wherein performing image alignment further comprises: performing a motion vector extraction between corresponding pixels between the images; andemploying a result of the motion vector extraction transforming one of the first image or the second image onto image coordinates of the other of the second image or the first image. 17. The storage device of claim 16, wherein the motion vector extraction further comprises: employing a motion search using a motion of a block of pixels in the first image to identify another block of pixels in the second image based on when a mean reduced sum of the absolute differences (MRSAD) is determined to be minimal;using a result of the motion search to identify a set of correspondences between pixels in the first image with pixels in the second image; andperforming a global motion estimation on the set of correspondences to generate an estimated global transformation. 18. The storage device of claim 15, wherein blending further comprises for saturated areas within one of the first image or second image, using pixels from a darker image among the first image and the second image to recover burned out details in a scene when creating the HDR image. 19. The storage device of claim 15, wherein the first exposure setting and the second exposure setting are different in at least one of an exposure time, an f-number or an International for Standardization (ISO) level. 20. The storage device of claim 15, wherein the pipelined plurality of actions, further comprises: performing a tone mapping to reduce a number of bits used to represent a color component in the HDR image, the tone mapping including:performing a gamma global mapping on the HDR image; andperforming a non-linear local mapping that maps a pixel based on a neighbor of pixels' average luminosity value.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (5)
Ward, Gregory John; Jia, Wenhui; Ninan, Ajit; Ten, Arkady; Wang, Gaven, Encoding, decoding, and representing high dynamic range images.
Tzur, Meir; Pinto, Victor; Pinhasov, Eran, Techniques of motion estimation when acquiring an image of a scene that may be illuminated with a time varying luminance.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.