البحث في مجموعات البراءات الوطنية والدولية
بعض محتويات هذا التطبيق غير متوفرة في الوقت الحالي.
إذا استمرت هذه الحالة ، يرجى الاتصال بنا علىتعليق وإتصال
1. (AU2016203579) VIDEO CAMERA SCENE TRANSLATION
ملاحظة: نص مبني على عمليات التَعرف الضوئي على الحروف. الرجاء إستخدام صيغ PDF لقيمتها القانونية
The claims defining the invention are as follows:
1.          A computer-implemented method, including:
accessing, by a computing system, a first video that shows a first two-dimensional (2D) scene of a real-world environment, the first video captured by a first camera located in the real-world environment having a first field of view;
defining a trip wire for performing video analytics at a first position of the first 2D scene shown in the first video;
obtaining a three-dimensional (3D) model of at least a portion of the real-world environment that is shown in the first 2D scene of the first video;
mapping, by the computing system, the trip wire from the first position of the first 2D scene shown in the first video to a first position of the 3D model that corresponds to the first position of the first 2D scene;
accessing, by the computing system, a second video that shows a second 2D scene of the real-world environment, the second video captured by the first camera having a second field of view that is different from the first field of view or by a second camera having the second field of view; and
projecting the trip wire from the first position of the 3D model to a first position of the second 2D scene that corresponds to a same location in the real-world environment as the first position of the first 2D scene.
2.          A method according to claim 1, wherein the defined trip wire is associated with one or more rules for performing video analytics.
3.          A method according to either claim 1 or claim 2, wherein the one or more rules for performing video analytics include counting a number of objects crossing the trip line.
4.          A method according to any one of the preceding claims, further including performing video analytics using the projected trip wire at the first position of the second 2D scene.
5.          A method according to any one of the preceding claims, further including determining that the first camera has moved and has a second field of view that is different from the first field of view.
6.          A method according to claim 5, wherein the first camera is at a fixed position in the environment and determining that the first camera has moved includes determining that the first camera has one or more of (i) panned, (ii) tilted, and/or (iii) zoomed.
7.          A method according to either claim 5 or claim 6, wherein the first camera is mobile and determining that the first camera has moved includes determining translations and rotations of the video camera.
8.          A method according to any one of claims 5 to 7, wherein determining that the first camera has moved includes using one or more sensors to identify movement of the first camera.
9.          A method according to any one of claims 5 to 8, wherein determining that the first camera has moved includes calculating a displacement between the first field of view and the second field of view.
10.        A method according to any one of claims 5 to 9, wherein determining that the first camera has moved includes using one or more sensors to identify movement of the first camera and calculating a displacement between the first field of view and the second field of view.
11.        A method according to any one of the preceding claims, wherein the 3D model includes (i) a ground plane of at least a portion of the real-world environment that is shown in the first 2D scene of the first video, and (ii) a first camera location relative to the ground plane.
12.        A method according to any one of the preceding claims, wherein the 3D model includes a detailed 3D model of at least a portion of the real-world environment that is shown in the first 2D scene of the first video, wherein the detailed 3D model includes objects and buildings in the portion of the real-world environment that is shown in the first 2D scene of the first video.
13.        A method according to any one of the preceding claims, wherein the 3D model is automatically generated from one or more 2D fields of view of the first camera.
14.        A method according to any one of the preceding claims, wherein the 3D model is generated using standard 3D modeling techniques.
15.        A method according to any one of the preceding claims, wherein the defined trip wire includes a line drawn in space relative to a ground plane of the portion of the real world environment that is shown in the first 2D scene of the first video.
16.        A method according to any one of the preceding claims, wherein the defined trip wire includes (i) a line, or (ii) one more lines defining a closed boundary, drawn in various directions in the 3D model.
17.        A method according to any one of the preceding claims, wherein defining a trip wire for performing video analytics includes identifying user input specifying the first position of the trip wire.
18.        A computer-implemented method, including:
receiving, at a computing system, information that defines a trip wire with respect to a first position of a two-dimensional (2D) map of a real-world environment;
obtaining a three-dimensional (3D) model of at least a portion of the real-world environment;
mapping, by the computing system, the trip wire from the first position of the 2D map of the real-world environment to a first position of the 3D model that corresponds to a same location in the real-world environment as the first position of the 2D map;
accessing, by the computing system, a video that shows a 2D scene of the real world environment; and
projecting the trip wire from the first position of the 3D model to a first position of the 2D scene that corresponds to the same location in the real-world environment as the first position of the 2D map.
19.        A method according to claim 18, wherein the 2D map of the real-world environment represents an overhead view of the real-world environment.
20.        One or more non-transitory computer-readable media having instructions stored thereon that, when executed by one or more processors, cause performance of operations including:
accessing, by a computing system, a first video that shows a first two-dimensional (2D) scene of a real-world environment, the first video captured by a first camera located in the real-world environment having a first field of view;
defining a trip wire for performing video analytics at a first position of the first 2D scene shown in the first video;
obtaining a three-dimensional (3D) model of at least a portion of the real-world environment that is shown in the first 2D scene of the first video;
mapping, by the computing system, the trip wire from the first position of the first 2D scene shown in the first video to a first position of the 3D model that corresponds to the first position of the first 2D scene;
accessing, by the computing system, a second video that shows a second 2D scene of the real-world environment, the second video captured by the first camera having a second field of view that is different from the first field of view or by a second camera having the second field of view; and
projecting the trip wire from the first position of the 3D model to a first position of the second 2D scene that corresponds to a same location in the real-world environment as the first position of the first 2D scene.