System and methods for persona identification using combined probability maps
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G06K-009/00
G06K-009/48
H04N-007/15
G06T-007/00
H04N-007/14
G06K-009/62
출원번호
US-0145874
(2013-12-31)
등록번호
US-9414016
(2016-08-09)
발명자
/ 주소
Lin, Dennis
Francisco, Glenn
Nguyen, Quang
Dang, Long
출원인 / 주소
PERSONIFY, INC.
대리인 / 주소
Invention Mine LLC
인용정보
피인용 횟수 :
2인용 특허 :
87
초록▼
Disclosed herein are systems and methods for extracting person image data comprising: obtaining at least one frame of pixel data and corresponding image depth data; processing the at least one frame of pixel data and the image depth data with a plurality of persona identification modules to generate
Disclosed herein are systems and methods for extracting person image data comprising: obtaining at least one frame of pixel data and corresponding image depth data; processing the at least one frame of pixel data and the image depth data with a plurality of persona identification modules to generate a corresponding plurality of persona probability maps; combining the plurality of persona probability maps to obtain an aggregate persona probability map; and generating a persona image by extracting pixels from the at least one frame of pixel data based on the aggregate persona probability map.
대표청구항▼
1. A method comprising: obtaining at least one frame of pixel data and corresponding image depth data;processing the at least one frame of pixel data and the corresponding image depth data with a plurality of persona identification modules to generate a corresponding plurality of persona probability
1. A method comprising: obtaining at least one frame of pixel data and corresponding image depth data;processing the at least one frame of pixel data and the corresponding image depth data with a plurality of persona identification modules to generate a corresponding plurality of persona probability maps, wherein the plurality of persona probability maps includes a hair-identification-module persona probability map, and wherein the plurality of persona identification modules comprises a hair-identification module for generating the hair-identification-module probability map at least in part by: identifying a plurality of pixel columns that cross an identified head contour; andfor each pixel column in the plurality of pixel columns: performing a color-based segmentation of the pixels in the pixel column into a foreground segment, a hair segment, and a background segment; andassigning the pixels in the hair segment an increased foreground-probability value in the hair-identification-module persona probability map;combining the plurality of persona probability maps to obtain an aggregate persona probability map; andgenerating a persona image by extracting pixels from the at least one frame of pixel data based on the aggregate persona probability map. 2. The method of claim 1, wherein: the at least one frame of pixel data comprises two frames of stereo pixel data; andthe corresponding image depth data is obtained from disparity data generated by a stereo disparity module. 3. The method of claim 2, wherein processing the at least one frame of pixel data and the corresponding image depth data comprises: generating a foreground-background map from the disparity data by designating pixels having a disparity value above a threshold as foreground pixels. 4. The method of claim 2, wherein: the disparity data comprises a plurality of disparity values for each pixel, each of the plurality of disparity values having an associated confidence value; andprocessing the at least one frame of pixel data and the corresponding image depth data comprises generating a foreground-background map from the disparity data by identifying pixels having a cumulative confidence measure above a threshold as foreground pixels. 5. The method of claim 1, wherein processing the at least one frame of pixel data and the corresponding image depth data comprises converting the corresponding image depth data to a foreground-background map using a thresholding operation. 6. The method of claim 5, wherein processing the at least one frame of pixel data and the corresponding image depth data further comprises performing a distance transform on the foreground-background map to obtain a persona probability map. 7. The method of claim 1, wherein extracting pixels is performed using a graph cut module. 8. The method of claim 1, wherein the aggregate persona probability map is formed by combining the plurality of persona probability maps using predetermined weights. 9. A method comprising: obtaining at least one frame of pixel data and corresponding image depth data;processing the corresponding image depth data to generate a foreground-background map;processing the at least one frame of pixel data and the foreground-background map to generate a plurality of persona probability maps at least in part through use of a hair-identification module, wherein the plurality of persona probability maps includes a hair-identification-module persona probability map that is generated by the hair-identification module at least in part by: identifying a plurality of pixel columns that cross an identified head contour; andfor each pixel column in the plurality of pixel columns: performing a color-based segmentation of the pixels in the pixel column into a foreground segment, a hair segment, and a background; andassigning the pixels in the hair segment an increased foreground-probability value in the hair-identification-module persona probability map;combining the plurality of persona probability maps to obtain an aggregate persona probability map; andgenerating a persona image by extracting pixels from the at least one frame of pixel data based on the aggregate persona probability map. 10. The method of claim 9, wherein the foreground-background map is generated from a disparity data volume. 11. The method of claim 9, wherein processing the at least one frame of pixel data and the foreground-background map to generate a plurality of persona probability maps comprises: performing a distance transform on the foreground-background map to generate a persona probability map. 12. An apparatus comprising: a foreground-background module configured to generate a foreground-background map based on image depth data;a plurality of persona identification modules configured to generate a corresponding plurality of persona probability maps, wherein the plurality of persona identification modules comprises a hair-identification module, wherein the plurality of persona probability maps includes a hair-identification-module persona probability map that is generated by the hair-identification module at least in part by: identifying a plurality of pixel columns that cross an identified head contour; andfor each pixel column in the plurality of pixel columns: performing a color-based segmentation of the pixels in the pixel column into a foreground segment, a hair segment, and a background segment; andassigning the pixels in the hair segment an increased foreground-probability value in the hair-identification-module persona probability map;a combiner module configured to generate an aggregate persona probability map based on the plurality of persona probability maps; anda persona extraction module configured to generate a persona image by extracting pixels from at least one frame of pixel data based on the aggregate persona probability map. 13. The method of claim 1, further comprising converting the head contour into a multi-segment polygon that approximates the head contour, the multi-segment polygon being formed of multiple head-contour segments, wherein identifying the plurality of pixel columns that cross the identified head contour comprises identifying pixel columns that cross one of the head-contour segments. 14. The method of claim 1, wherein performing a color-based segmentation comprises performing a color-based segmentation using a clustering algorithm. 15. The method of claim 14, wherein the clustering algorithm is a k-means algorithm with k=3. 16. The method of claim 1, wherein performing the color-based segmentation of the pixels in a given pixel column into the foreground segment, the hair segment, and the background segment of the given pixel column comprises: identifying an average foreground-pixel color, an average hair-pixel color, and an average background-pixel color for the given pixel column; andidentifying the foreground segment, the hair segment, and the background segment of the given pixel column using a clustering algorithm to cluster the pixels in the given pixel column around the identified average foreground-pixel color, the identified average hair-pixel color, and the identified average background-pixel color for the given pixel column, respectively. 17. The method of claim 16, wherein: identifying the average foreground-pixel color for the given pixel column comprises identifying the average foreground-pixel color for the given pixel column based on a first set of pixels at an innermost end of the given pixel column;identifying the average hair-pixel color for the given pixel column comprises identifying the average hair-pixel color for the given pixel column based on a second set of pixels that includes a point where the given pixel column crosses the identified head contour; andidentifying the average background-pixel color for the given pixel column comprises identifying the average background-pixel color for the given pixel column based on a third set of pixels at an outermost end of the given pixel column. 18. The method of claim 1, further comprising, for each pixel column in the plurality of pixel columns: assigning the pixels in the foreground and background segments an equal probability of being in the foreground and being in the background in the hair-identification-module persona probability map. 19. The method of claim 1, wherein assigning the pixels in the hair segment an increased foreground-probability value in the hair-identification-module persona probability map comprises: assigning a first value to the pixels in the hair segment in the hair-identification-module persona probability map; andassigning a second value to the pixels in the foreground and background segments in the hair-identification-module persona probability map, wherein the first value corresponds to a higher probability of being a foreground pixel than does the second value.
연구과제 타임라인
LOADING...
LOADING...
LOADING...
LOADING...
LOADING...
이 특허에 인용된 특허 (87)
Cipolla Roberto (Cambridge GBX) Okamoto Yasukazu (Chiba-ken JPX) Kuno Yoshinori (Osaka-fu JPX), 3D human interface apparatus using motion recognition based on dynamic image processing.
Panahpour Tehrani, Mehrdad; Ishikawa, Akio; Sakazawa, Shigeyuki, Apparatus, method and computer program for classifying pixels in a motion picture as foreground or background.
Clanton,Charles H.; Ventrella,Jeffrey J.; Paiz,Fernando J., Cinematic techniques in avatar-centric communication during a multi-user online simulation.
DeMenthon Daniel F. (Columbia MD), Computer vision system for position monitoring in three dimensions using non-coplanar light sources attached to a monito.
Tian, Dihong; Mauchly, J. William; Friel, Joseph T., Generating and rendering synthesized views with multiple video streams in telepresence video conference sessions.
Iwamoto, Masayuki; Fujimura, Koichi, Image processing apparatus, method for processing and image and computer-readable recording medium for causing a computer to process images.
Carter, James; Yaacob, Arik; Darrah, James F., Managing the layout of multiple video streams displayed on a destination display screen during a videoconference.
Bang, Gun; Um, Gi-Mun; Chang, Eun-Young; Kim, Taeone; Hur, Nam-Ho; Kim, Jin-Woong; Lee, Soo-In, Method and apparatus for improving quality of depth image.
Haskell, Barin Geoffry; Puri, Atul; Schmidt, Robert Lewis, Scene description nodes to support improved chroma-key shape representation of coded arbitrary images and video objects.
Mackie, David J.; Tian, Dihong; Weir, Andrew P.; Buttimer, Maurice; Friel, Joseph T.; Mauchly, J. William; Chen, Wen-Hsiung, System and method for providing enhanced video processing in a network environment.
Prahlad, Anand; Schwartz, Jeremy A.; Ngo, David; Brockway, Brian; Muller, Marcus S., Systems and methods for classifying and transferring information in a storage network.
Weiser, Reginald; McGravie, Richard; Diouskine, Roman; Teboul, Jeremy, Systems and methods for providing video conferencing services via an ethernet adapter.
Rudolph, Eric; Rui, Yong; Malvar, Henrique S; He, Li Wei; Cohen, Michael F; Tashev, Ivan, Systems and methods for real-time audio-visual communication and data collaboration in a network conference environment.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.