Intelligent Analysis for Georeferenced Video Using Context-Based Random Graphs

Video sensor networks are formed by the joining of heterogeneous sensor nodes, which is frequently reported as video of communication functionally bound to geographical locations. Decomposition of georeferenced video stream presents the expression of video from spatial feature set. Although it has been studied extensively, spatial relations underlying the scenario are not well understood, which are important to understand the semantics of georeferenced video and behavior of elements. Here we propose a method of mapping georeferenced video sequences for geographical scenes and use contextual random graphs to investigate semantic knowledge of georeferenced video, leading to correlation analysis of the target motion elements in the georeferenced video stream. We have used the connections of motion elements, both the correlation and continuity, to present a dynamic structure in time series that reveals clues to the event development of the video stream. Furthermore, we have provided a method for the effective integration of semantic and campaign information. Ultimately, the experimental results show that the provided method offers a better description of georeferenced video elements that cannot be achieved with existing schemes. In addition, it offers a new way of thinking for the semantic description of the georeferenced video scenarios.


Introduction
The notion of wireless multimedia sensor networks (WMSNs) is frequently reported as the convergence between the concepts of wireless sensor networks and distributed smart cameras [1].As a result, an increasing number of video clips is being collected, which has created complex datahandling challenges [2].Further, some types of video data are naturally tied to geographical locations.For example, video data from traffic monitoring may not contain much meaning without its associated location information.Therefore, most potential applications of a WMSN require the sensor network paradigm to support location-based multimedia services as well as manipulate large scale data at the same time to provide a high quality of experience (QoE), which raises an important issue.How to investigate an intelligent processing method for georeferenced multimedia?Although the question has been extensively addressed theoretically, the method of mapping video sequences to geographical scenes remains to be described.On the other hand, with the growth of geographic information system (GIS) whose major growth area is the convergence between GIS and multimedia technology, a new paradigm named video-GIS emerged [3][4][5].The major researches facing video-GIS are the coding of georeferenced video and the content and types of services that should be provided by georeferenced video.Further improvement of these processes is contingent on deeper understanding of video, as well as improved understanding of the spatial relationship of geographic space.It is due to the necessity of using video-GIS to visualize the relationship between the video analysis methods and the real geographical scene, resulting in georeferenced multimedia intelligent processing method based on context-based random graphs.
Georeferenced video is fundamental process in video-GIS development.Prior research activities on georeferenced video technologies and applications have been conducted.Most of them make use of video and GPS sensors.In [6,7], Stefanakis and Peterson and Klamma et al. proposed a unified framework for hypermedia and GIS.Pissinou et al. [8] explored topology and direction under the proposed georeferenced video.The work of Hwang et al. and Joo et al. [9,10] defined the metadata of georeferenced video, which International Journal of Distributed Sensor Networks support interoperability between GIS and video images.In the field of georeferenced video search, Liu et al. [11] presented a sensor-enhanced video annotation system, which searches video clips for the appearance of particular objects.Ay et al. proposed the use of geographical properties of videos [12], while Wang gave a method of time-spatial images to extract the basic movement information [13].Although single media have been studied extensively, its semantics in geographic space are poorly understood.How to determine the spatial relationship of video elements is one of the most important operations on georeferenced video.For instance, a moving video element changes its position, shape, size, speed, and attribute values over time.Understanding the changing process and rules of these attributes is of important significance to the geographical description of the video.
Many techniques for video event recognition have been proposed.As the work on model-driven methodology which has become well established and approached maturity, the most common and popular conceptualization of fusion systems is the SVM model [14,15].However, such methodology not only cannot solve the problems, such as multi-instance, diversity, and multimodal, but needs a large number of training samples.Most previous studies to date have used data-driven method [16] which has been carefully designed to signal clear and distinct semantic of the videos [17][18][19][20][21].In our event recognition application, we observe that some events may share common motion patterns.Though involved in pattern discovery, data-driven method also contributes to social network during pattern discovery [22][23][24][25].These works have showed a high accuracy in the differentiating of video and its semantic extraction frame.However, most multimedia applications are unknown and uncertain, which are extremely difficult to meet the requirements of real-time stream processing.
Previous studies have shown that multimedia intelligent processing method is important to the development of video-GIS and have achieved inspiring progress.However, these solution methods have suffered from the classical ensemble average limitation presented by the analysis of low-level characteristics.Therefore, the spatial data gathered are sometimes inconclusive and, in part, contradictory.These algorithms usually build or learn a model of the target object first and then use it for tracking, without adapting the model to account for the changes in the appearance of the object, for example, large variation of pose or facial expression, or the surroundings, for example, lighting variation.Furthermore, it is assumed that all images are acquired with a stationary camera.Such approach, in our view, is prone to performance instability, and thus it needs to be addressed when building a robust visual tracker.
To overcome these problems, we will begin by looking at some valid models, which are suitable for georeferenced video understanding and behavior analysis.In this paper, we propose a new event recognition framework for consumer videos by leveraging a large amount of videos.As we know, graph structure provides a complex, dynamic, and robust framework for assembling complex relationships involved in the objects [26], which is suitable for our goal.Thus, multiple random behaviors are presented in certain movement, making the graph structure unsuitable for describing the real video scenario.To circumvent this problem, random graph model has been taken into consideration, which can be seen as a rather simplified model of the evolution of certain communication net [27].In our research, it could simplify the analysis of the interaction between video objects substantially for revealing the new insight into the relationships between objects and its complex interaction.Our analysis focuses on describing spatial relationships bound to objects using random graph grammar in georeferenced video, developing a scientific analysis of behavior and structured methods of georeferenced video understanding.

Preliminary
Surveillance video data is mostly non-ortho image data so that it does not match up with the geography scene vector data using the traditional method.To solve this problem, a mapping method of video scene imaging data to geography scene vector data is adopted in the paper, as showed in Figure 1.Firstly, the virtual viewpoint camera is constructed by the camera interior and exterior parameters.Secondly, geography scene virtual imaging can be gained from geography scene vector data using the process of model transformation, viewpoint transformation, and pruning according to the computer graphics rendering process, with the corresponding relationship between an object in virtual imaging and vector object.Thirdly, the image matching technology based on the features that have invariant character for translation, scale and rotation is used to match the geography scene virtual imaging and video image.Finally, the corresponding relationship between video image and vector data is established using that between an object in virtual imaging and vector object, with the purpose of accomplishing the mapping of video scene to objects in geography scene.
In the following part, we will introduce several preliminary key steps.

Selection Algorithm of Multicamera Based on Spatial Correlation and Target
Priority.Multicamera surveillance system should not only gain detecting and tracking information of motion element of the single camera, but also make the coherent dynamic scene description using all the observations to some extent.Meanwhile, every motion element could be tracked by cameras simultaneously.How to select cameras for tracking a specific target is particularly important in video sensor networks.Based on the spatial correlation [28] and target priority, the paper proposes a selection algorithm of multicamera with task allocation optimized to achieve the automatic selection according to the target priority at each moment.
The algorithm is based on the assumption that a camera with no task carries out the basic single camera tracking which has lower power consumption, and the high-priority task could be preempted when bending.The selection algorithm of multicamera is shown as in Algorithm 1.
The set of images  = { 1 ,  2 , . . .,   } is observed by these  cameras, and  denotes the set of cameras selected.  is correlation coefficient of the two images   and   .The larger the correlation coefficient, the more correlated the two images.In step 6,  denotes the task priority with a default value  0 , which can be marked manually by monitoring person.It assigns cameras to the motion element with high priority and coordinates cameras to track different targets based on spatial correlation and target priority.

Organization of Video and Location Data.
We have put forward a coding model of video-GIS that is comprised of video and camera's position in conjunction with its view direction and distance.Thus, the location data can be collected automatically by various small sensors to a camera, such as a GPS and a compass (see Figure 2).This eliminates manual work and allows the annotation process to be accurate and scalable.Therefore, we investigate the realtime collection, coding, and integration of video information and GPS information on the SEED-VPM642 platform, and finally we can obtain two different bit-rate location-based streaming media.The lower bit-rate one can be positioned to the wireless network broadcast live, and the higher one can be positioned to the hard disk storage.
In the coding of video-GIS, we need to calculate the three-dimensional coordinate of the video object [29].As video-GIS coding based on mobile sensor cannot calculate single video frame by three-dimensional control field, the most effective way is using digital map and spatial geometrical relations (see Figure 3).
Therefore, the geometric relationship among GPS, posture sensor, imaging space, and object space should be built.It is assumed that the axis of imaging space , ,  is parallel with that of object spatial , , , respectively.Consider In detail,   is the coordinate vector of point  in the three-dimensional space.The coordinate function of GPS antenna in the given mapping frame is expressed as  GPS (). Att () represents the rotation matrix function while   represents the proportional relationship of image frame and object GPS is the excursion of the geometric center of GPS antenna and the camera lens.
For acquiring a more precise spatial locating information, we need to get the GPS information and attitude information generated by a posture sensor at least.Therefore, the spatial locating information is described by the combination of GPS and angle direction elementary (Heading, Pitch, and Roll), which obtained by Micro Inertial Measurement Unit (MIMU), as shown in Table 1.
As shown in Table 1, there are two kinds of the spatial locating information: (1) GPS information: such as UTC time and longitude latitude; (2) angle direction elementary information: including Heading, Pitch, and Roll.

Digital Map-Based Image
Resolution.The features of digital maps are expressed by a two-dimensional plane on the vertical projection of the vector data.From the standpoint of this work, the video image is a raster data expression of the feature in the height direction of the information, and video image can also be expressed as the data format of the dotted line surface after the vector processing.Video images and digital map on the point, the line, and the corresponding expression of the surface can be shown at Table 2.
From the view of technology, we subject map-based image resolution to a three-dimensional measurement challenge and then use single-frame video images and digital map matching to define the changes in three steps.The first step is feature extraction of dense range image, which aims to extract the features of point and line.Under the premise of the full calibration to video frame, we can identify the particular characteristics of extracted target to meet the special requirement.For instance, the corners of building or telegraph pole as a fixed line characteristic for the expression of video image is perpendicular to the target.Once formulated, the second step is to combine the line characteristics into the characteristics of the surface using texture information.The third step is matching with digital map vector data.The contents include a variety of different  matching points, points and lines, a line and a line, and the line and the plane between form and technique, which is shown in Figure 4.  Measure indicates the measuring relation among motion element using some measure in measuring space, such as distance.In the same planar reference domain, Direction is the equity mutual relationship between source target and reference target.Definition 5. Visual Feature.In the georeferenced video stream, the visual characters of one motion element, including color, texture, and shape, will be dynamically changed with the time .Therefore, the changes of visual characters of a motion element within the scope of spatial constraint should be described accurately [30].And the visual characters mainly include Color, Texture, Shape, and Size.Texture can reflect the structure mode and gray space distribution formed by local pixels in motion element, while the low-level features can clearly define and describe the motion element.

Behavior and Interaction of Motion Element.
In the georeferenced video stream, Behavior of the motion element within the specific scope of spatial constraint represents the behavior state sequence, as shown in Figure 6.Let the state set of Behavior be a BehaviorState, and the typical element is  with the definition as follows: (2)

𝑃(𝑡)
Object Under the influence of temporal subspace  and spatial relation , Interaction is the description of mutual influence between motion element   and   .Behavior state of Object can be any state instance in the BehaviorState collection, and interaction production rule and interaction optimization update rule are involved in Rule.Therefore, the measuring of interaction has two influence factors, temporal and spatial factors.
Due to the close correlation of spatial relation at any time point  +1 and former   , the spatial relation at  +1 is always closely related to that at former time point   .Thus, the spatial relation evolution process among motion elements can be defined as a Markov chain in the temporal subspace , with its evolution having Markov quality Meanwhile, the measuring value  of interaction between the two motion element established Relation can be computed based on the planar spatial distance Distance, velocity magnitude, and direction angle, including the current topology at time point   , as shown in Figure 7.
In the georeferenced video stream, the dynamic update function of interaction relation within the scope of spatial constraint is shown as follows: (5) Among them, () represents the interaction relation measuring value between a certain motion element and others, with the range of  ∈ [0, 1].The higher value indicates the more hospitable relationship.When the interaction is established by behavior state instances, the initial value works as (0) =  1 × (, ) +  2 × (, ).() indicates the duration of interaction relation with the current state, and the dynamic change of  parameter is shown as follows:

Semantics and Formalization of Georeferenced Video
For the accurate description and behavior understanding of motion elements in the georeferenced video stream, the paper proposes an analysis method based on sparse random graphs with the purpose of observing the character evolution with time and presents an indicating and measuring method of video motion element with dynamic topology structure information based on context-sensitive sparse random graph grammar.

Formalization of Georeferenced Video.
Random graph  = (, , Ω) is defined on triple, while the edge set  of graph  with the vertex set  is defined in probabilistic spaces Ω.Consider Each edge of random graph  is mutually independent; namely, any two vertexes that established incidence relation connected independently with probability .As the spatial relation will be dynamically changed during the movement with the time factor, it is necessary to describe the motion state and interaction relationships within specific spatial area using random graph.Context-sensitive sparse random graph grammar can be defined as five-meshes  = (,   , , , ℎ) . (8) Among them,  is the root vertex that an initial vertex of semantic event in the georeferenced video stream.There is only one  vertex in the video event sequence.Vertex   = { 1 ,  2 ,  3 , . ..} involves all the motion elements emerged in the specific spatial area. in the formula means the evolution process and rule of random graph  while  the state transition functions.The cohesion of random subgraph ℎ indicates the inner coupling degree of motion element group.
It shows the motion status and interaction information of a motion element   labeled index at the time point Time.
Among them, Location and Speed represent the position coordinate and the velocity of motion element   in the planar area, respectively.Interaction is the description of interaction while  = (, , ) the spatial relation existed in the motion element.Virtual feature  shows low-level features information of a motion element including Color, Shape, and Size at the time point Time. = {, V, , } is the basic state of motion element.

Evolution Rule.
As a posterior method, dynamic process of motion elements in the video stream can be visually described and showed based on sparse random graph.The temporal and spatial evolution model of motion element is able to describe the basic character and dynamic process of spatial relation accurately.The essence of dynamic evolution process of sparse random graph is the continuous transition process of state space in random graphs.Therefore, the state transition function of sparse random graph can be defined as a mapping relation Among them, Θ is the state space of sparse random graph, Θ = ( 1 ,  2 , . . .,   )  , and  is a variable in state region.
The dynamic evolution process of sparse random graph includes its character update of motion element vertex   , emerging vertex with the Appear and Disappear behavior states, and the dynamic adjustment of edge set  and interaction relation  of random graphs.For the accurate description of event development process in georeferenced video stream, evolution rule algorithm of sparse random graph is shown in Algorithm 2.
We can get the corresponding dynamic evolution model of sparse random graph using the evolution rule algorithm.
Step (2) in the algorithm shows the creating and adding root vertex , and  0 = ( 0 ,  0 ) := ({}, 0).Adding a new motion element vertex  tmp in sparse random graph  Active is in step (5) while deleting the vanish vertex   and its association edge in step (11).Among them, function getRestriction(  ) in step (18) and getAttract (  ) in step (20) indicate whether it can delete or add the edge that vertex   associated, respectively.
Step (27)   generated by video event: notify and alarm, which is shown in Figure 8.
The paper is mainly to detect the unexpected crowd incident and conflict in the massive video events and proposes a novel two-layer discriminate method, which consists of individual attribute layer and group attribute layer.Once occurring video abnormal event, the corresponding real-time status of random graph must be described, which can be expressed as follows.
(1) Individual Attribute Layer.The owned velocity of multiple random graph nodes has modified radically in per unit of time , and the relevant movement direction has also changed significantly.
Specifically speaking, the detection and selection of variation range or interval of movement attributes in random graph can use sliding window.In the continuous movement attribute value  = { 1 ,  2 , . . .,   } in time series,  1 exists before the emergence of  2 , while  2 exists before  3 .The difference is obtained by the two continuous attribute values.In the paper, the data in the sliding interval Δ is viewed as the discriminative and forecasting sample, when the continuous difference (  ,   , ) is larger than the given threshold, and the sliding intervals Δ is within the max time threshold.Otherwise, recalibrate over the entire sliding intervals for new computation.
(2) Group Attribute Layer.The multiple interaction and distance values among random graph nodes in groups fluctuate greatly, or the multiple numerical variations of interaction relationship in random subgraph are changed significantly.The discriminant analysis of video abnormal event is achieved according to the check whether the change rate of parameter value ⃗  is greater than the given threshold ℎℎ, as Once either circumstance occurred, it must be entering the next notify phase.
When entering the notify discriminative phase, the random subgraph showing diffusion or flocking status makes numerical calculation.Using the computing method of structure entropy value, the corresponding random subgraph status is measured, and the entropy value ℎ(  ) is viewed as the warning degree of video abnormal behavior and emergency.With regard to different levels of urgency and security, the warning degree () is set to different threshold intervals as follows: The warning degree () is divided into three warning threshold intervals in the paper, which are Warning1, Warning2, and Warning3.Specifically, Warning1 indicates the early warning degree, which means that video abnormal event will be occurred in the next unit time and the discriminative module obtains alertness.Warning2 shows the probable warning degree and is the identifying processing transformed into the CBR phase.If the entropy value of random subgraph is greater than the max value of given threshold interval, the CBR discrimination phase works.Based on the video event features, the traditional CBR method is used to further identification.Warning3 expresses the confirmed warning degree, which can enter the Alarm phase of video abnormal event directly without the traditional CBR method.
The discriminate method based on the random graph is defined as graph-based reasoning (GBR) in the paper, while the improved GBR fused with traditional CBR method is GBR-C.The intelligent analysis for different video scenes plays an important role in the real-time detection of video abnormal behaviors and mass incidents.The instantaneous status information of video motion element is integrated with the random graph model and summarizes the random subgraph patterns and behavior rules with a statistical description.In violation of the behavior regularity of common video events, it is a latent exceptional event, and extracts the features of video motion elements involved which are recorded in object layer stream for the efficient retrieval of content-based video.

Experiment and Analysis
In order to verify the feasibility and availability of the proposed framework, space information of a motion element is extracted at real-time based on the detection and tracking [31,32].According to the dynamic change situation of space semantics, a timing description method using random graph grammar depicts the event development of video stream clearly.

Interaction Description.
Interaction is the mutual incidence relation among motion element.For the accurate description of the dynamic change process of interaction relation, interaction  should be calculated real-time based on the spatial information in experimental video including planar spatial distance, velocity magnitude, and direction angle.And the calculation results of real-time interaction update function () of the video clip trim from frame 550 to frame 685 is shown in Figure 9.
In Figure 9, function  1 shows a changing trend of increasing first and then decreasing gradually in the video clip.The minimum value of interaction  1 is at frame 685 with the value 0.11 while the maximum is at frame 586 with the value 0.38.And function  2 indicates the changing process of two close targets.The minimum value of  2 is at frame 592 with the value 0.23 while the maximum is at frame 685 with the value 0.79.The increasing planar spatial distance Distance and motion direction variation of two motion elements make the decreasing interaction value.On the contrary, as the planar spatial distance decreases and the duration of interaction continues to increase, interaction value  increases gradually.
The previous results show that it can accurately depict the dynamic varying changes of the interaction relation of video motion elements.However, the accurate depiction is an indispensable premise for the description of the georeferenced video stream.

Georeferenced Video Stream Description.
Based on the richer spatial semantic of motion elements in the georeferenced video stream, we can realize the intelligent parsing of georeferenced video content using context-sensitive sparse random graph grammar.The spatial relationship of motion elements in image space is transformed to that of object space, and the motion status and interaction relation can be depicted using random graph.The continuous transition process of inner state space in random graph is enforced with the dynamic evolution process of sparse random graph.
With the spatial reference data, the sparse random graph evolution processing based on the monitoring target is achieved.And the consecutive people emerged within the video surveillance range are labeled as A, B, C, and D which are shown in Figure 10.As soon as the moving object appears, a new random graph node will express it; when it leaves the surveillance confine, the corresponding node will disappear while the edge set constituted by the interaction that associated with the node is set to null.Using our video test data, the evolutionary process and timing evolving description diagram of the video clip trim from frame 1041 to frame 1712 is shown in Figure 10.
We can see that the timing evolving description diagram can be constructed by the automatic intelligent analysis and calculation of a video clip, and it verifies the correctness and effectiveness of the evolution rule algorithm of sparse random graph.Within the scope of the specific geographical space, the time-varying attributes of random graph nodes are visual displayed, such as behavior state, spatial location, and movement parameter.And the basis recorded information of each video motion element is shown in Algorithm 3.
Among them, the basic information consists of attribute information, spatial location information, and other movement parameter, which are shown in Algorithm 3. The attribute information State indicates the behavior status of the video motion element with succinct expressional number 0, 1, 2 and 3, which are described respectively with the four basic behavior  {, , V, }.And the interaction relationship attribute including the index of two elements, the numerical calculation value of interaction, and the relative spatial directions.The whole structural description of video motion element generated automatically is shown in Figure 11.

Conclusion
In summary, findings from the present study are all based on low-level visual features, which mean that there was a shortage of spatial constraints and coupling analysis with geography environment.It is necessary to establish the relationship between video analysis method and the real geographical scene.A georeferenced video analysis method is proposed based on the context-based random graph.The data are obtained using a wireless network of environmental sensors scattered at the supervising area and a vision sensor monitoring the same geographical area.Experimental results prove that the proposed description method of georeferenced video using random graph is feasible and efficient.Through the intelligent parsing of the georeferenced video data stream, we can get a novel visual description method using random graph which can clearly depict the development clue of video scenes and also offer the possibility to browse the video stream quickly.Meanwhile, random graph can be used as an effective nonlinear indexing for the content-based video indexing and browsing application.
As a future work, we will propose the enhancement of the implemented algorithms with alternative combination rules and the fusion of audio and video to deal with the uncertainty, imprecision, and incompleteness of the underlying information.In addition, large amounts of data should be conducted to set various parameters, such as thresholds, false alarm rates, and fusion weights.

Figure 2 :Figure 3 :
Figure 2: Experimental hardware and software to acquire georeferenced video.

Figure 4 :
Figure 4: Mapping from Image to Digital Map.

Definition 3 .
Relation.Relation is an incidence relation of mutual influence between two motion elements in the same time subspace . = (  ,   , ) shows the relationship between motion element   and   in time subspace  which means one-dimensional time coordinates.The measurement of interaction established between the two elements uses probability , which is dynamic adjustment with the influence of temporal-spatial factor, and  ∈ [0, 1].Definition 4. Spatial Relation.Spatial Relation includes measuring relation, direction relation, and topological relation.Spatial Relation  = (, , ).

Figure 7 :
Figure 7: A diagram of interaction relation.

Figure 11 :
Figure 11: Structural description of video motion feature.
(a).According to the warning degree of video abnormal behavior and emergency, three different warning threshold intervals are set in our test.And the Warning2 degree occurred between 252 and 270 seconds shown in Figure12(b).The Warning1 indicates the early warning degree in most of the time, which means that video abnormal event will be emerged.Figure12(c)shows the real-time nodes number of random subgraph in the video surveillance scope while Figure12(d)shows the total graph node number.

Table 1 :
Sample of GPS and MIMU.

Table 2 :
Correspondence between video images and digital map.
Behavior state sequence of motion element.Definition 2. Behavior Attribute.Behavior description of a single typical motion element mainly includes spatial location and speed.Spatial location can be defined as () = (  ,   ,   ), which means that the spatial location of the motion element Object at time point   is (  ,   ), and   and   represent the horizontal and vertical ordinate values in the two-dimensional plane, respectively.() = { Value ,  Vector ,   } indicates the motion element Object with velocity magnitude  Value and velocity direction  Vector at the time point   , among which  Vector is the unit vector in a general planar domain.
3.1.SyntacticDescription of Motion Element.Video motion element mainly refers to the entity objects that could be identified clearly in visual and are important in morphology, such as pedestrians in video surveillance.The description methods of motion element are mainly based on color and texture at present, which is difficult to support the definition of motion element, behavior analysis, and behavior understanding.For a better description of the dynamic characteristic of the video motion element, the paper first gives a definition to some related concepts of motion element.(a)Appear.The emerging motion element is newly appear and distinguished from the existing ones in the specific area of geographical boundary, and the state of which is called Appear.Then the motion element starts to be detected and tracked.Appear instance is regarded as the first instance of motion element.Target trajectory Stop Figure 5: The definition of Stop.(b) Disappear.In contrast with the Appear state definition, Disappear means the state of disappearance in the geographical boundary specific area or the untraceable state within a specific time, which is viewed as the last instance for the state description.Disappear state is the signal of canceling motion element detection and tracking.(c) Stop.Stop S is defined on triple  = ((),  min (),  max ()).Among them, () means the spatial plane area, and  min () and  max () represent the (d) Move.Within the scope of spatial constraint, Move  is a general designation of connecting the other three basic states in a continuous motion process of motion element.An instance of Move can be represented as  = ( |   ,  +1 | ).By connecting the other three basic state instances, Move can form a linear sequence formed through the combination of Appear, Stop, and Disappear.
accomplishes the dynamic update of interaction relation  in sparse random graph  Active .For   ∈   do //Update all Edges in  Active For   ∈   do //Update  of Graph in  Active

Table 3 :
Three test sample videos.
of time series, movement status information of each motion element including the behavior state sequence, real-time spatial location information, and the statistical information about interaction relation can be queried directly from the XML file.It also provides a novel simple nonlinear indexing for the understanding and description of video content.5.3.Performance of Video Event Warning.To validate the proposed early warning method of video abnormal behavior and emergency, we analyzed the performance of various attributes using the video test data which involves a crowd video scene.Experimental analysis mainly contains the real-time warning entropy value of random subgraph, warning degree, and realtime changes of corresponding subgraph node number and

Table 4 :
Comparison of crossing sample A with different methods.

Table 5 :
Comparison of flocking sample B with different methods.

Table 6 :
Comparison of conflict sample C with different methods.