Recherche dans les collections de brevets nationales et internationales
Certains contenus de cette application ne sont pas disponibles pour le moment.
Si cette situation persiste, veuillez nous contacter àObservations et contact
1. (WO2017003777) ZOOM ET PANORAMIQUE BASÉS SUR LE CONTENU POUR L'ÉDITION DE CONTENU VIDÉO
Note: Texte fondé sur des processus automatiques de reconnaissance optique de caractères. Seule la version PDF a une valeur juridique

CLAIMS

WHAT IS CLAIMED IS:

1. A computer-implemented method comprising:

receiving input video data comprising video frames having a first aspect ratio greater than 2: 1 ;

determining an event of interest represented in the video data;

determining a beginning of the event in a first video frame of the video data;

determining first pixel coordinates in the first video frame associated with the beginning of the event;

determining a first cropped window from the first video frame, the first cropped window comprising a portion of the first video frame including the first pixel coordinates, the first cropped window having a second aspect ratio less than 2: 1 and the first cropped window having a first size and a first position within the first video frame;

determining an end of the event in a second video frame of the video data;

determining second pixel coordinates in the second video frame associated with the end of the event, the second pixel coordinates different than the first pixel coordinates;

determining a second cropped window from the second video frame, the second cropped window comprising a portion of the second video frame including the second pixel coordinates, the second cropped window having the second aspect ratio and the second cropped window having a second size and a second position within the video frame; and determining output data corresponding to the first cropped window and the second cropped window.

2. The computer-implemented method of claim 1, further comprising:

identifying an object of interest in the video data; and

tracking the object of interest across multiple video frames,

wherein, prior to determining the beginning of the event and determining the end of the event, the determining the event further comprises:

determining a third video frame corresponding to the event of interest based on the object of interest.

3. The computer-implemented method of claim 1, wherein determining the output data further comprises:

determining output video data simulating at least one of panning and zooming from the first cropped window to the second cropped window.

4. The computer-implemented method of claim 5, wherein determining the output data further comprises:

generating a first video tag corresponding to the first cropped window, the first video tag including the first pixel coordinates, the first size, the first position and a first timestamp associated with the first video frame; and

generating a second video tag corresponding to the second cropped window, the second video tag including the second pixel coordinates, the second size, the second position and a second timestamp associated with the second video frame.

5. The computer-implemented method of claim 1, wherein determining the event of interest further comprises:

identifying a first person represented in the video data;

identifying a second person represented in the video data;

determining, at a first time, that a first number of pixels between the first person and the second person in the video data exceeds a threshold; and

determining, at a second time following the first time, that a second number of pixels between the first person and the second person in the video data is less than the threshold, wherein the second time is associated with the event of interest.

6. The computer-implemented method of claim 1, further comprising:

determining a first direction between the first pixel coordinates and the second pixel coordinates,

wherein the determining the first cropped window further comprises:

determining the first cropped window, the first cropped window comprising a portion of the first image including the first pixel coordinates and an area of pixels in the first direction from the first pixel coordinates.

7. The computer-implemented method of claim 1, wherein:

determining the second cropped window comprises determining the second position relative to the second video frame is different from the first position relative to the first video frame; and

determining output video data comprises determining output video data simulating panning from the first cropped window at the first position to the second cropped window at the second position.

8. The computer-implemented method of claim 1, wherein:

determining the second cropped window comprises determining the second size is different from the first size; and

determining output video data comprises determining output video data simulating zooming from the first cropped window having the first size to the second cropped window having the second size.

9. A system, comprising:

at least one processor;

a memory including instructions operable to be executed by the at least one processor to cause the system to perform a set of actions comprising:

receiving input video data comprising video frames having a first aspect ratio greater than 2: 1 ;

determining an event of interest represented in the video data; determining a beginning of the event in a first video frame of the video data; determining first pixel coordinates in the first video frame associated with the beginning of the event;

determining a first cropped window from the first video frame, the first cropped window comprising a portion of the first video frame including the first pixel coordinates, the first cropped window having a second aspect ratio less than 2: 1 and the first cropped window having a first size and a first position within the first video frame;

determining an end of the event in a second video frame of the video data; determining second pixel coordinates in the second video frame associated

with the end of the event, the second pixel coordinates different than the first pixel coordinates;

determining a second cropped window from the second video frame, the second cropped window comprising a portion of the second video frame including the second pixel coordinates, the second cropped window having the second aspect ratio and the second cropped window having a second size and a second position within the video frame; and

determining output data corresponding to the first cropped window and the second cropped window.

10. The system of claim 9, the set of actions further comprising:

identifying an object of interest in the video data;

tracking the object of interest across multiple video frames;

determining, prior to determining the beginning of the event and determining the end of the event, a third video frame corresponding to the event of interest based on the object of interest;

determining a first color histogram corresponding to the object;

determining a second color histogram corresponding to third video frame; and comparing the first color histogram with the second color histogram.

11. The system of claim 9, the set of actions further comprising:

generating a first video tag corresponding to the first cropped window, the first video tag including the first pixel coordinates, the first size, the first position and a first timestamp associated with the first video frame; and

generating a second video tag corresponding to the second cropped window, the second video tag including the second pixel coordinates, the second size, the second position and a second timestamp associated with the second video frame.

The system of claim 9, the set of actions further comprising:

identifying a first person represented in the video data;

identifying a second person represented in the video data;

determining, at a first time, that a first number of pixels between the first person and the second person in the video data exceeds a threshold; and

determining, at a second time following the first time, that a second number of pixels between the first person and the second person in the video data is less than the threshold, wherein the second time is associated with the event of interest.

13. The system of claim 9, the set of actions further comprising:

determining a first direction between the first pixel coordinates and the second pixel coordinates; and

determining the first cropped window, the first cropped window comprising a portion of the first image including the first pixel coordinates and an area of pixels in the first direction from the first pixel coordinates.

14. The system of claim 9, the set of actions further comprising:

determining the second cropped window comprises determining the second position relative to the second video frame is different from the first position relative to the first video frame; and

determining output video data comprises determining output video data simulating panning from the first cropped window at the first position to the second cropped window at the second position.

15. The system of claim 9, the set of actions further comprising:

determining the second cropped window comprises determining the second size is different from the first size; and

determining output video data comprises determining output video data simulating zooming from the first cropped window having the first size to the second cropped window having the second size.