[미국특허]
Video scene classification by activity
원문보기
IPC분류정보
국가/구분
United States(US) Patent
등록
국제특허분류(IPC7판)
G06K-009/00
H04N-005/232
H04N-005/77
G11B-027/031
H04N-021/8549
G11B-027/00
G11B-027/22
G11B-027/34
H04N-005/91
H04N-009/82
출원번호
US-0513151
(2014-10-13)
등록번호
US-9984293
(2018-05-29)
발명자
/ 주소
Hodulik, Nick
Taylor, Jonathan
출원인 / 주소
GoPro, Inc.
대리인 / 주소
Sheppard Mullin Richter & Hampton LLP
인용정보
피인용 횟수 :
0인용 특허 :
31
초록▼
Video and corresponding metadata is accessed. Events of interest within the video are identified based on the corresponding metadata, and best scenes are identified based on the identified events of interest. A video summary can be generated including one or more of the identified best scenes. The v
Video and corresponding metadata is accessed. Events of interest within the video are identified based on the corresponding metadata, and best scenes are identified based on the identified events of interest. A video summary can be generated including one or more of the identified best scenes. The video summary can be generated using a video summary template with slots corresponding to video clips selected from among sets of candidate video clips. Best scenes can also be identified by receiving an indication of an event of interest within video from a user during the capture of the video. Metadata patterns representing activities identified within video clips can be identified within other videos, which can subsequently be associated with the identified activities.
대표청구항▼
1. A method for identifying scenes in videos, the method comprising: obtaining one or more electronic files defining a video captured with a camera, the one or more electronic files including event of interest information indicating reception of verbal input during the capture of the video, the verb
1. A method for identifying scenes in videos, the method comprising: obtaining one or more electronic files defining a video captured with a camera, the one or more electronic files including event of interest information indicating reception of verbal input during the capture of the video, the verbal input identifying an occurrence of an event of interest within the video, the event of interest occurring at an event moment within the video, the event of interest information identifying (i) a given input type of the verbal input, input types of the verbal input including a first input type, a second input type, and a third input type, and (ii) an input moment during the capture of the video at which the verbal input was received, wherein the first input type indicates the input moment occurring before the event moment, the second input type indicates the input moment occurring during the event moment, and the third input type indicates the input moment occurring after the event moment;identifying the input moment based on the event of interest information;identifying the given input type of the verbal input based on the event of interest information;identifying the event moment based on the input moment and the given input type, wherein the event moment is identified to occur before the input moment based on the given input type being the first type, the event moment is identified to occur during the input moment based on the given input type being the second type, and the event moment is identified to occur after the input moment based on the given input type being the third type;identifying a portion of the video as a video clip associated with the event of interest based on the event of interest information, the video clip comprising a first time amount of the video occurring before the event moment and a second time amount of the video occurring after the event moment, the first time amount and the second time amount being determined based on a type of an activity captured within the video; andstoring clip information indicating the association of the video clip with the event of interest and the portion of the video included in the video clip. 2. The method of claim 1, wherein the occurrence of the event of interest is identified further based on metadata associated with the video, the metadata captured during the capture of the video, the metadata characterizing velocity or acceleration of the activity captured within the video. 3. The method of claim 2, wherein a metadata criteria for identifying the occurrence of the event of interest based on the metadata associated with the video is based on the type of the activity captured within the video such that a first metadata criteria is used to identify the occurrence of the event of interest based on the activity being of a first type and a second metadata criteria is used to identify the occurrence of the event of interest based on the activity being of a second type, the first metadata criteria being different from the second metadata criteria. 4. The method of claim 1, wherein the verbal input comprises a spoken command associated with tagging events of interest. 5. The method of claim 1, wherein the event moment includes a point in time within the video or a duration of time within the video. 6. The method of claim 1, further comprising: identifying one or more non-event moments within the video, the one or more non-event moments not associated with any event of interest within the video; andstoring non-event information indicating the one or more non-event moments. 7. The method of claim 6, wherein the one or more non-event moments are identified based on matching metadata associated with the video with a metadata pattern determined to not be of interest to a user, the metadata characterizing velocity of the activity captured within the video, acceleration of the activity captured within the video, visuals captured within the video, or audio captured within the video. 8. A system for identifying scenes in videos, the system comprising: one or more physical processors configured by computer-readable instructions to: obtain one or more electronic files defining a video captured with a camera, the one or more electronic files including event of interest information indicating reception of verbal input during the capture of the video, the verbal input identifying an occurrence of an event of interest within the video, the event of interest occurring at an event moment within the video, the event of interest information identifying (i) a given input type of the verbal input, input types of the verbal input including a first input type, a second input type, and a third input type, and (ii) an input moment during the capture of the video at which the verbal input was received, wherein the first input type indicates the input moment occurring before the event moment, the second input type indicates the input moment occurring during the event moment, and the third input type indicates the input moment occurring after the event moment;identify the input moment based on the event of interest information;identify the given input type of the verbal input based on the event of interest information;identify the event moment based on the input moment and the given input type, wherein the event moment is identified to occur before the input moment based on the given input type being the first type, the event moment is identified to occur during the input moment based on the given input type being the second type, and the event moment is identified to occur after the input moment based on the given input type being the third type;identify a portion of the video as a video clip associated with the event of interest based on the event of interest information, the video clip comprising a first time amount of the video occurring before the event moment and a second time amount of the video occurring after the event moment, the first time amount and the second time amount being determined based on a type of an activity captured within the video; andstore clip information indicating the association of the video clip with the event of interest and the portion of the video included in the video clip. 9. The system of claim 8, wherein the occurrence of the event of interest is identified further based on metadata associated with the video, the metadata captured during the capture of the video, the metadata characterizing velocity or acceleration of the activity captured within the video. 10. The system of claim 9, wherein a metadata criteria for identifying the occurrence of the event of interest based on the metadata associated with the video is based on the type of the activity captured within the video such that a first metadata criteria is used to identify the occurrence of the event of interest based on the activity being of a first type and a second metadata criteria is used to identify the occurrence of the event of interest based on the activity being of a second type, the first metadata criteria being different from the second metadata criteria. 11. The system of claim 8, wherein the verbal input comprises a spoken command associated with tagging events of interest. 12. The system of claim 8, wherein the event moment includes a point in time within the video or a duration of time within the video. 13. The system of claim 8, wherein the one or more physical processors are further configured by the computer-readable instructions to: identify one or more non-event moments within the video, the one or more non-event moments not associated with any event of interest within the video; andstore non-event information indicating the one or more non-event moments. 14. The system of claim 13, wherein the one or more non-event moments are identified based on matching metadata associated with the video with a metadata pattern determined to not be of interest to a user, the metadata characterizing velocity of the activity captured within the video, acceleration of the activity captured within the video, visuals captured within the video, or audio captured within the video. 15. A non-transitory computer-readable storage medium storing instructions for identifying scenes in videos, the instructions, when executed by one or more physical processors, configured to cause the one or more physical processors to: obtain one or more electronic files defining a video captured with a camera, the one or more electronic files including event of interest information indicating reception of verbal input during the capture of the video, the verbal input identifying an occurrence of an event of interest within the video, the event of interest occurring at an event moment within the video, the event of interest information identifying (i) a given type of the verbal input, input types of the verbal input including a first input type, a second input type, and a third input type, and (ii) an input moment during the capture of the video at which the verbal input was received, wherein the first input type indicates the input moment occurring before the event moment, the second input type indicates the input moment occurring during the event moment, and the third input type indicates the input moment occurring after the event moment;identify the input moment based on the event of interest information;identify the given input type of the verbal input based on the event of interest information;identify the event moment based on the input moment and the given input type, wherein the event moment is identified to occur before the input moment based on the given input type being the first type, the event moment is identified to occur during the input moment based on the given input type being the second type, and the event moment is identified to occur after the input moment based on the given input type being the third type;identify a portion of the video as a video clip associated with the event of interest based on the event of interest information, the video clip comprising a first time amount of the video occurring before the event moment and a second time amount of the video occurring after the event moment, the first time amount and the second time amount being determined based on a type of an activity captured within the video; andstore clip information indicating the association of the video clip with the event of interest and the portion of the video included in the video clip. 16. The computer-readable storage medium of claim 15, wherein the occurrence of the event of interest is identified further based on metadata associated with the video, the metadata captured during the capture of the video, the metadata characterizing velocity or acceleration of the activity captured within the video. 17. The computer-readable storage medium of claim 16, wherein a metadata criteria for identifying the occurrence of the event of interest based on the metadata associated with the video is based on the type of the activity captured within the video such that a first metadata criteria is used to identify the occurrence of the event of interest based on the activity being of a first type and a metadata second criteria is used to identify the occurrence of the event of interest based on the activity being of a second type, the first metadata criteria being different from the second metadata criteria. 18. The computer-readable storage medium of claim 15, wherein the verbal input comprises a spoken command associated with tagging events of interest. 19. The computer-readable storage medium of claim 15, wherein the event moment includes a point in time within the video or a duration of time within the video. 20. The computer-readable storage medium of claim 15, wherein the instructions, when executed by the one or more physical processors, are further configured to cause the one or more physical processors to, identify one or more non-event moments within the video, the one or more non-event moments not associated with any event of interest within the video; andstore non-event information indicating the one or more non-event moments.
Edwards,Jeffrey L.; Ahmad,Subutai, Automatic editing of a visual recording to eliminate content of unacceptably low quality and/or very little or no interest.
Chuang, Daniel B.; Candell, Lawrence M.; Ross, William D.; Beattie, Mark E.; Fang, Cindy Y.; Ren, Bobby; Blanchard, Jonathan P., Imaging system for immersive surveillance.
Swenson, Anne; Agnoli, Giovanni; Rodriguez, Enrique; Lyons, Charles; Meaney, Brian; Cerf, Dave; Stern, Mike, Media editing application for auditioning different types of media clips.
Meaney, Brian; Pendergast, Colleen; Matsuda, Ken; Agnoli, Giovanni; Khan, Itrat U.; Minjack, Zachury Bryant; Stern, Mike; De Marco, Vincenzo; Diephouse, Matthew D.; LaSalle, Louis; Fleischhauer, Michael; McCommons, Jordan P., Media editing with multi-camera media clips.
Matsuda, Ken; Cerf, Dave; Khan, Itrat U.; Diephouse, Matthew D.; Meaney, Brian; De Marco, Vincenzo; McCommons, Jordan P.; LaSalle, Louis, Media-editing application with anchored timeline.
Curcio, Igor Danilo Diego; Mate, Sujeet Shyamsundar; Cricri, Francesco; Roininen, Mikko Joonas; Sathish, Sailesh, Method and apparatus for semantic extraction and video remix creation.
Evans, Matt; Lagemann, Ole; Danty, John; Helms, Jan-Hinnerk; Lengeling, Gerhard; Soren, Alexander; Martin, Timothy Benjamin; Pillhofer, Stefan, Method and system to process digital audio data.
※ AI-Helper는 부적절한 답변을 할 수 있습니다.