Methods and systems for assigning pixels distance-cost values using a flood fill technique
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G06K-009/34
G06K-009/00
G06T-007/50
G06T-007/73
G06K-009/46
G06T-007/00
G06K-009/52
G06T-007/194
출원번호
US-0425331
(2017-02-06)
등록번호
US-9953223
(2018-04-24)
발명자
/ 주소
Lin, Dennis
Nguyen, Quang
Dang, Gia
Zhang, Yi
Venshtain, Simion
Nguyen, Cong
출원인 / 주소
Personify, Inc.
대리인 / 주소
Invention Mine LLC
인용정보
피인용 횟수 :
0인용 특허 :
116
초록▼
Disclosed herein are methods and systems for assigning pixels distance-cost values using a flood fill technique. One embodiment takes the form of a process that includes obtaining video data depicting a head of a user, obtaining depth data associated with the video data, and selecting seed pixels fo
Disclosed herein are methods and systems for assigning pixels distance-cost values using a flood fill technique. One embodiment takes the form of a process that includes obtaining video data depicting a head of a user, obtaining depth data associated with the video data, and selecting seed pixels for a flood fill at least in part by using the depth information. The process also includes performing the flood fill from the selected seed pixels. The flood fill assigns respective distance-cost values to pixels of the video data based on position-space cost values and color-space cost values. In some embodiments, the process also includes classifying pixels of the video data as foreground based at least in part on the assigned distance-cost values. In some other embodiments, the process also includes assigning pixels of the video data foreground-likelihood values based at least in part on the assigned distance-cost values.
대표청구항▼
1. A method comprising: obtaining video data depicting a head of a user;obtaining depth data associated with the video data;selecting seed pixels for a hair-identification flood fill for identifying pixels depicting hair of the head of the user, the seed pixels selected at least in part by using the
1. A method comprising: obtaining video data depicting a head of a user;obtaining depth data associated with the video data;selecting seed pixels for a hair-identification flood fill for identifying pixels depicting hair of the head of the user, the seed pixels selected at least in part by using the obtained depth data;performing the hair-identification flood fill from the selected seed pixels, the hair-identification flood fill assigning respective distance-cost values to pixels of the video data based on respective position-space-cost values and respective color-space-cost values, wherein performing the hair-identification flood fill comprises: identifying a plurality of neighbor pixels of a current pixel;determining respective step-cost values from the current pixel to each pixel in the plurality of neighbor pixels; andassigning each pixel in the plurality of neighbor pixels a respective distance-cost value based on a distance-cost value of the current pixel and the respective step-cost values; andidentifying, using a foreground identification process, a persona of the user from the video data, wherein the respective distance-cost values assigned to pixels by the hair-identification flood fill is one of a plurality of inputs of the foreground identification process. 2. The method of claim 1, wherein selecting seed pixels for the hair-identification flood fill further comprises selecting seed pixels for the hair-identification flood fill at least in part by using the video data. 3. The method of claim 1, further comprising: obtaining a head contour that estimates an outline of the depicted head of the user, the head contour being based at least in part on the depth data associated with the video data,wherein the selected seed pixels are on an upper contour, wherein the upper contour is an upper portion of the head contour. 4. The method of claim 3, wherein the selected seed pixels are equally distributed along the upper contour. 5. The method of claim 3, wherein the selected seed pixels are of colors that are found in a user-hair-color model. 6. The method of claim 1, wherein selecting seed pixels for the hair-identification flood fill comprises: identifying pixels having noisy depth values over a series of frames; andselecting the identified noisy depth-pixels as seed pixels for the hair-identification flood fill. 7. The method of claim 6, wherein the selected seed pixels are located within an extended head box. 8. The method of claim 6, wherein the selected seed pixels have intermittent depth values that are within a threshold tolerance of a depth value corresponding to the head of the user. 9. The method of claim 6, wherein the selected seed pixels are of colors that are found in a user-hair-color model. 10. The method of claim 1, wherein a first set of the selected seed pixels are on an upper contour and a second set of the selected seed pixels have noisy depth values over a series of frames, wherein the upper contour is an upper portion of a head contour that estimates an outline of the depicted head of the user, the method further comprising: initializing the distance-cost values of the seed pixels in the first set to be zero; andinitializing the distance-cost values of the seed pixels in the second set to be non-zero. 11. The method of claim 1, wherein performing the hair-identification flood fill comprises: determining a minimum distance-cost value from at least one of the selected seed pixels to a current pixel; andassigning the current pixel a distance-cost value that is the determined minimum distance-cost value. 12. The method of claim 11, wherein determining a minimum distance-cost value comprises: comparing a current distance-cost value corresponding with a current flood-fill path to a prior distance-cost value corresponding with a prior flood-fill path. 13. The method of claim 12, wherein the current flood-fill path and the prior flood-fill path originate from a common seed pixel. 14. The method of claim 12, wherein the current flood-fill path and the prior flood-fill path originate from different seed pixels. 15. The method of claim 1, wherein performing the hair-identification flood fill comprises performing the hair-identification flood fill along a plurality of flood-fill paths, the method further comprising: terminating the hair-identification flood fill along a current flood-fill path in response to at least one termination criteria, the termination criteria comprising: a current pixel not being a user-hair color according to a user-hair-color model;the current pixel being a background color according to a background-color model;a distance-cost value to the current pixel being greater than a distance-cost threshold; anda step-cost value to the current pixel being greater than a step-cost threshold. 16. A system comprising: a communication interface;a processor; anddata storage containing instructions executable by the processor for causing the system to carry out a set of functions, the set of functions including: obtaining video data depicting a head of a user;obtaining depth data associated with the video data;selecting seed pixels for a hair-identification flood fill for identifying pixels depicting hair of the head of the user, the seed pixels selected at least in part by using the depth data;performing the hair-identification flood fill from the selected seed pixels, the hair-identification flood fill assigning respective distance-cost values to pixels of the video data based on respective position-space-cost values and respective color-space-cost values, wherein performing the hair-identification flood fill comprises: identifying a plurality of neighbor pixels of a current pixel;determining respective step-cost values from the current pixel to each pixel in the plurality of neighbor pixels; andassigning each pixel in the plurality of neighbor pixels a respective distance-cost value based on a distance-cost value of the current pixel and the respective step-cost values; andidentifying, using a foreground identification process, a persona of the user from the video data, wherein the respective distance-cost values assigned to pixels by the hair-identification flood fill is one of a plurality of inputs of the foreground identification process.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (116)
Cipolla Roberto (Cambridge GBX) Okamoto Yasukazu (Chiba-ken JPX) Kuno Yoshinori (Osaka-fu JPX), 3D human interface apparatus using motion recognition based on dynamic image processing.
Panahpour Tehrani, Mehrdad; Ishikawa, Akio; Sakazawa, Shigeyuki, Apparatus, method and computer program for classifying pixels in a motion picture as foreground or background.
Clanton,Charles H.; Ventrella,Jeffrey J.; Paiz,Fernando J., Cinematic techniques in avatar-centric communication during a multi-user online simulation.
DeMenthon Daniel F. (Columbia MD), Computer vision system for position monitoring in three dimensions using non-coplanar light sources attached to a monito.
Tian, Dihong; Mauchly, J. William; Friel, Joseph T., Generating and rendering synthesized views with multiple video streams in telepresence video conference sessions.
Iwamoto, Masayuki; Fujimura, Koichi, Image processing apparatus, method for processing and image and computer-readable recording medium for causing a computer to process images.
Carter, James; Yaacob, Arik; Darrah, James F., Managing the layout of multiple video streams displayed on a destination display screen during a videoconference.
Bang, Gun; Um, Gi-Mun; Chang, Eun-Young; Kim, Taeone; Hur, Nam-Ho; Kim, Jin-Woong; Lee, Soo-In, Method and apparatus for improving quality of depth image.
Yeh Hwa-Young M ; Lure Yuan-Ming F ; Lin Jyh-Shyan, Method and system for re-screening nodules in radiological images using multi-resolution processing, neural network, and image processing.
Berman Arie ; Vlahos Paul ; Dadourian Arpag, Method for removing from an image the background surrounding a selected subject by generating candidate mattes.
Colmenarez, Antonio J.; Gutta, Srinivas, Person tagging in an image processing system utilizing a statistical model based on both appearance and geometric features.
Haskell, Barin Geoffry; Puri, Atul; Schmidt, Robert Lewis, Scene description nodes to support improved chroma-key shape representation of coded arbitrary images and video objects.
Mackie, David J.; Tian, Dihong; Weir, Andrew P.; Buttimer, Maurice; Friel, Joseph T.; Mauchly, J. William; Chen, Wen-Hsiung, System and method for providing enhanced video processing in a network environment.
Prahlad, Anand; Schwartz, Jeremy A.; Ngo, David; Brockway, Brian; Muller, Marcus S., Systems and methods for classifying and transferring information in a storage network.
Weiser, Reginald; McGravie, Richard; Diouskine, Roman; Teboul, Jeremy, Systems and methods for providing video conferencing services via an ethernet adapter.
Rudolph, Eric; Rui, Yong; Malvar, Henrique S; He, Li Wei; Cohen, Michael F; Tashev, Ivan, Systems and methods for real-time audio-visual communication and data collaboration in a network conference environment.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.