1 Introduction

There are two basic types of network intrusion detection technology, that is, anomaly detection and misuse detection. At present, anomaly detection is the main direction of intrusion detection research. This detection technology establishes the normal behavior pattern of the system or the user and detects the intrusion by the comparison and matching between the monitoring system or the actual behavior pattern of the user and the normal mode [1, 2]. It is characterized by the lack of excessive knowledge of the system defects, strong adaptability, and the ability to detect the unknown intrusion modes. The high probability of false alarm is the main factor that restricts the application of abnormal detection at present. The key problem of anomaly detection is the establishment of normal behavior pattern and how to compare and judge the current behavior by using normal behavior pattern [3, 4].

The application of intelligent technology such as neural network and machine learning in abnormal detection has been carried out at home and abroad. The research goal is to improve the accuracy, real time, efficiency, and adaptability of the detection system. Some of the research results have been close to or reached the level of practicality in detection performance and operability [5]. This paper first introduces the user behavior anomaly detection model based on machine learning proposed by Lane et al. On the basis of which a new detection model is proposed, which is mainly used for the human intrusion detection system (IDS) with the shell command as the audit data [6, 7]. The model uses a variety of shell command sequences with different length to express the user’s behavior pattern and sets up a number of sample sequences to describe the normal behavior contour of the legitimate users in the network. When the detection is detected, the similarity degree is assigned by the sequence of variable length commands, and the user behavior is judged by the similarity degree after the noise filtering. Experiments using Unix user shell command data show that the new detection model has high detection performance and strong operability.

2 A fixed length command sequence detection model for machine learning

2.1 Basic principle of machine learning

Machine learning is a new branch of artificial intelligence. Through the study of human cognitive mechanism, it establishes various learning models with the aid of machine (computer system) and gives the ability of machine learning. On this basis, it builds a task-oriented learning system with specific application. A machine learning system consists mainly of a learning unit, a knowledge base, and an executive unit, in which a learning unit builds a knowledge base by using information provided by an external source and makes improvements to it (adding new knowledge or reorganizing the existing knowledge). The execution unit performs tasks by using knowledge in the knowledge base, and the information after task execution is fed back to the learning unit as input for further learning. The learning unit is the core of the learning function of the machine learning system. It involves the way of dealing with the outside information and the methods used in the process of acquiring new knowledge. The knowledge base is used to store knowledge, including the original domain knowledge of the system (this knowledge is long term, relatively stable), and the various new knowledge acquired through learning (this knowledge is short term and varied). The choice of knowledge representation plays a very important role in the design of learning system. Execution unit is a key part that makes learning system practical and at the same time able to evaluate learning methods.

A large part of machine learning research focuses on two areas: classification and problem solving. After more than 30 years of development, there have been a lot of learning methods, such as inductive learning, case learning, genetic learning, etc., but these methods have their limitations. It is the mainstream of current research to explore new learning methods and algorithms in combination with specific application fields.

2.2 Description of sequence detection model of fixed length command

Lane of Purdue University has proposed an anomaly detection model of user behavior based on machine learning and conducted in-depth research and experiments on the model. The model uses the fixed length shell command sequence to represent the user’s behavior pattern and sets up the sample sequence library to describe the normal behavior profile of the legitimate users. When working, the current command sequence of the monitored user is compared and matched to the sequence library of the command and the behavior of the monitored user is judged according to the similarity of the two. The main points of the model are as follows:

  1. 1.

    The shell command sequence with fixed length is used as the minimum data unit to describe the user’s behavior pattern, and an example learning method is used to establish the sample sequence library of the legitimate user.

  2. 2.

    A similarity degree between two sequences is defined, which is used to represent the similarity between the behavior patterns represented by the two sequences. On this basis, a similarity between a sequence and a sample sequence library is defined, which is used to represent the maximum similarity between the behavior patterns represented by the sequence and the various normal behavior patterns.

  3. 3.

    When the model works, it calculates the similarity between each sequence in the monitored user sequence stream and the sample sequence library. Then, the similarity degree is processed with window filtering, and the similarity judgment value arranged in time sequence is obtained. According to the size of the judgment, the behavior of the monitored user is judged in real time.

There are several key problems in the model:

  1. 1.

    The selection of optimal sequence length;

  2. 2.

    The extraction of sample sequence;

  3. 3.

    The definition of similarity function;

  4. 4.

    Selection of noise filtering algorithm.

Lane et al. performed a large number of experiments on UNIX user shell command data for the above problems. The following are the experimental results:

  1. 1.

    The optimal sequence length is related to the behavior characteristics of specific users. As the length of the sequence increases (from 1 to 15), the detection performance of the model changes with different users.

  2. 2.

    In a variety of similarity functions, the similarity function corresponding to the correlation between adjacent commands is superior to the similarity function without considering correlation. Mean noise filtering and median filtering algorithms have little difference in detection performance.

  3. 3.

    The clustering method is more adaptable to different users in the sampling method such as clustering, occurrence probability extraction, time sequence interception, random selection and so on.

3 Dense crowd state perception model

3.1 Technology roadmap

The population state perception process mainly consists of four processes, namely video acquisition and processing, optical flow calculation, vector light flow mapping to geographic space, population state detection and analysis, as shown in Fig. 1.

Fig. 1
figure 1

Population state perception technical route

The concrete steps are as follows:

  1. 1.

    First, through the mobile phone or camera in the shopping mall, hospital, station, bank and other people crowded and easy to happen public events in a place to gather experimental data. It mainly includes normal crowd behavior video and abnormal behavior video, such as fighting, running, gathering, abrupt, retrograde and so on, providing test data for the detection of abnormal behavior of the crowd. Based on high-definition remote sensing image or GPS, the location of the center point of camera is collected. Based on camera calibration or direct setting, the camera’s height, tilt angle, azimuth, and main distance are obtained [7].

  2. 2.

    Second, the optical flow field of each monitor video is calculated. Based on the two front and back images in the video, the real-time optical flow field of the current video is obtained by using the relevant optical flow calculation method.

  3. 3.

    Third, map the optical flow fields in each region to geospatial space. The optical flow field calculated by video image space is pixel coordinates. It is difficult to locate, measure or analyze in the geographic space. It needs to be projected into the map by the mapping method of image space and geospatial space.

  4. 4.

    Fourth, in geographic space, any vector light flow has a clear spatial reference. It is very easy to understand the position, direction, speed, acceleration, and other motion parameters of the crowd. It is an important basis for the detection and early warning of the crowd.

3.2 Key technology

The motion vector field based on video extraction is mainly dependent on the image space. The optical flow field is based on the image space in the direction of pixels, and the real measurement of the dynamic target in the geographic scene cannot be obtained, including the position, direction, and speed [8]. In order to make the state of the population be located, measurable, and analytical, and to have the advantage of reading, it is necessary to focus on the extraction of optical flow field and the mapping of the optical flow field to the geographic space.

3.2.1 Optical flow calculation method of Lucas–Kanade

Optical flow field computation is one of the most important key technologies of crowd state perception. The calculation of optical flow field is first proposed by American scholars HOM and SCHUNCK, and the improvement and optimization of optical flow field correlation algorithms are constantly promoted in the study. The optical flow field is widely applied to pedestrian detection, pedestrian gait analysis, and motion segmentation. Some scholars also evaluate the efficiency of optical flow method [9].

In this paper, the Lucas–Kanade optical flow method is used to obtain the motion state of the crowd. The method first converts video into sequence image and transforms it into sequence gray scale. In a short time, the gray value of the same point of moving target will not change and then the equation is constructed and the optical flow vector is calculated [10]. The change in crowd state can be expressed by field model. The change in sequence grayscale image is the expression of crowd movement. The faster the grayscale changes, the faster the crowd will move. A vector field can be formed by the constraint of the same gray value of the feature points in the near time. The magnitude and direction of each vector in vector field are the speed and direction of crowd movement. It should be noted that the vector field is not the real velocity of the crowd, but the perspective view of the crowd movement.

The Lucas–Kanade method searches the motion vectors of each pixel according to the change in the gray gradient of the adjacent frames. Based on the research of HORN and SCHUNCK, the Lucas–Kanade optical flow algorithm is effective. The method is a gradient-based locally parameterized optical flow estimation method. The method assumes that the optical flow vector in a small window in the image space remains unchanged, and the optical flow vector can be computed by the least square method.

It is assumed that the brightness of one pixel point \(\left( {x,y} \right)\) in the image plane is \(E\left( {x,y,t} \right)\) at the t moment, \(u\left( {x_{0} ,y_{0} } \right)\) is the horizontal variation component of the pixel point, and \(v\left( {x,y} \right)\) is the vertical component of the pixel point.

$$u = {\text{d}}x / {\text{d}}t$$
(1)
$$v = {\text{d}}y / {\text{d}}t$$
(2)

After a time interval of \(\Delta t\), the point corresponds to the brightness of \(E\left( {x + \Delta x,y + \Delta y,t + \Delta t} \right)\), and when the \(\Delta t\) is very small to 0, it is assumed that the brightness remains unchanged, such as formula (3).

$$E\left( {x,y,t} \right) = E\left( {x + \Delta x,y + \Delta y,t + \Delta t} \right)$$
(3)

When the brightness of the point changes, the brightness of the moving point is developed by the Taylor formula as shown in Eq. (4).

$$E\left( {x + \Delta x,y + \Delta y,t + \Delta t} \right) = E\left( {x,y,t} \right) + \frac{\partial E}{\partial x}\Delta x + \frac{\partial E}{\partial y}\Delta y + \frac{\partial E}{\partial t}\Delta t + \varepsilon$$
(4)

If the derivative of the two order is infinitesimal, the \(\Delta t\) delta T is 0, as shown in formula (5).

$$- \frac{\partial E}{\partial t} = \;\frac{\partial E}{\partial x}\frac{{\text{d}}x}{{\text{d}}t} + \frac{\partial E}{\partial y}\frac{{\text{d}}y}{{\text{d}}t} = \nabla Egw$$
(5)

The formula \(w = (u,v)\) is the basic optical flow constraint equation.

\(E_{x} = \partial E/\partial x,\;E_{y} = \partial E/\partial y,\;E_{t} = \partial E/\partial t\) shows the gradient of the pixel gray level in the \(x,y,t\) direction in the image and can be rewritten as: \(E_{x} u + E_{y} v + E_{t} = 0\), that is, gray consistency constraint conditions. The input video size is \(X \times Y\); all the Harris corners of the current image are detected as the feature points; then, the Pyramid LK optical flow method is used to take the characteristic window of \(w \times w\) and to solve the optical flow UV of the feature points based on the constraint conditions of the optical flow, such as formula (6).

$$\left[ {\begin{array}{*{20}l} {E_{x1} } \hfill & {} \hfill & {E_{y1} } \hfill \\ {E_{x2} } \hfill & {} \hfill & {E_{y2} } \hfill \\ {} \hfill & \cdots \hfill & {} \hfill \\ {E_{xn} } \hfill & {} \hfill & {E_{yn} } \hfill \\ \end{array} } \right]\left[ \begin{aligned} u \hfill \\ v \hfill \\ \end{aligned} \right] = \left[ {\begin{array}{*{20}l} { - E_{t1} } \hfill \\ { - E_{t2} } \hfill \\ \cdots \hfill \\ { - E_{tn} } \hfill \\ \end{array} } \right]$$
(6)

That is \(Av = - b\).

Then, the least squares method is used, such as formulas (7) and (8).

$$A^{T} Av = A^{T} ( - b)$$
(7)
$$\left[ \begin{aligned} u \hfill \\ v \hfill \\ \end{aligned} \right] = \left[ {\begin{array}{*{20}l} {\sum {E_{{xi^{2} }} } } \hfill & {\sum {E_{xi} \sum {E_{yi} } } } \hfill \\ {\sum {E_{xi} } \sum {E_{yi} } } \hfill & {\sum {E_{{yi^{2} }} } } \hfill \\ \end{array} } \right]\left[ \begin{aligned} - \sum {E_{xi} \quad \sum {E_{ti} } } \hfill \\ - \sum {E_{yi} \quad \sum {E_{ti} } } \hfill \\ \end{aligned} \right]$$
(8)

The experimental video contains a total of \(n_{\text{Frame}}\) frames, and the time is \(t_{\text{Sum}}\) seconds. The time required for each frame can be calculated by formula (9).

$$t_{\text{Each}} = t_{\text{Sum}} /n_{\text{Frame}}$$
(9)

The summation is calculated from 1 to \(n\) and calculated by image derivative and accumulated separately. At the same time, the weight function \(W\left( {i,j,k} \right)\), \(i,j,k \in \left[ {1,m} \right]\) is needed to highlight the coordinates of the central point of the window, and the Gauss function is used to make the calculation effect obvious, and the real-time optical flow field of the current video can be obtained. The algorithm has better monitoring effect for specific application scenarios, such as station crowd, and has broad application prospects.

3.2.2 Mapping the optical field to 2D geospatial

The optical flow field obtained by the video has its coordinates in pixel coordinates, which is difficult to locate, measure, or analyze in the geographic space. It needs to be projected onto the map through the mapping method of image space and geographic space. The clear spatial reference and macroscopic observation perspectives of the map will help to grasp the movement of the large-scale population as a whole. Mapping optical flow field to 2D geospatial is another key technique of this paper [11].

The traditional method needs to calculate the 2D homography matrix. Firstly, three or more control points are selected in the image to obtain the image coordinates of the control points; then, corresponding points are selected in the high-definition remote sensing image or map to obtain the spatial right-angle coordinates of the control points; finally, the 2D homography matrix H is calculated. This method not only requires interactive selection, but also requires high-definition remote sensing images or topographic maps. It is suitable for small-scale video surveillance scenarios and is difficult to use in large-scale video surveillance systems. In particular, the 2D homography approach is ineffective when acquiring 3D geometric information of the target. In this context, Xingguo et al. proposed a mutual mapping model for surveillance video and 2D geospatial space, which can better achieve a one-to-one correspondence between image space and two-dimensional geospatial space.

In order to obtain the real three-dimensional information of the crowd, taking into account the PTZ camera, this article has introduced the mutual mapping model of surveillance video mapping to 2D geospatial. This model requires the camera internal and external parameters to be known for ease of calculation. The camera model is represented by a homogeneous equation, such as formula (10).

$$\lambda \left[ \begin{aligned} x \hfill \\ y \hfill \\ 1 \hfill \\ \end{aligned} \right] = \left[ {\begin{array}{*{20}l} {f_{x} } \hfill & s \hfill & {u_{0} } \hfill & 0 \hfill \\ {} \hfill & {f_{y} } \hfill & {v_{0} } \hfill & 0 \hfill \\ {} \hfill & {} \hfill & 1 \hfill & 0 \hfill \\ \end{array} } \right]\left[ {\begin{array}{*{20}l} R \hfill & { - R\mathop c\limits^{ - } } \hfill \\ {0^{T} } \hfill & 1 \hfill \\ \end{array} } \right]\left[ \begin{aligned} X \hfill \\ Y \hfill \\ Z \hfill \\ \;1 \hfill \\ \end{aligned} \right]$$
(10)

In the formula, \((X,Y,Z)\) is the space right-angle coordinates in GIS, \(\mathop c\limits^{ - }\) is the position of camera center in the space right-angle coordinate system, \((x,y)\) is the image space coordinate, \((u_{0} ,v_{0} )\) is the main point coordinates, \(f_{x} = f/p_{x}\), \(f_{y} = f/p_{y}\) is the equivalent focal length, \(s\) is the non-orthogonal coordinate axis, and \(R\) is rotation transformation.

In the formula (10), the camera’s internal and external parameters are hidden in the matrix. In order to facilitate the use of the camera model, a camera model which is easy to use, such as formula (11), is obtained through comparative analysis of camera models of photogrammetry and computer vision.

$$\lambda \left[ \begin{aligned} x \hfill \\ y \hfill \\ 1 \hfill \\ \end{aligned} \right] = \left[ {\begin{array}{*{20}l} f \hfill & {u_{0} } \hfill \\ {} \hfill & {fv_{0} } \hfill \\ {} \hfill & 0 \hfill \\ \end{array} } \right]\left[ {\begin{array}{*{20}l} {a_{1} } \hfill & {b_{1} } \hfill & {c_{1} } \hfill \\ { - a_{2} } \hfill & { - b_{2} } \hfill & { - c_{2} } \hfill \\ { - a_{3} } \hfill & { - b_{3} } \hfill & { - c_{3} } \hfill \\ \end{array} } \right]\left[ {\begin{array}{*{20}l} 1 \hfill & 0 \hfill & 0 \hfill & { - X_{x} } \hfill \\ 0 \hfill & 1 \hfill & 0 \hfill & { - Y_{y} } \hfill \\ 0 \hfill & 0 \hfill & 1 \hfill & { - Z_{z} } \hfill \\ \end{array} } \right] \cdot \left[ \begin{aligned} X \hfill \\ Y \hfill \\ Z \hfill \\ 1 \hfill \\ \end{aligned} \right]$$
(11)

The parameter \(u_{0} = x_{{0{\text{ph}}}} ,v_{0} = H_{\text{pic}} - y_{{0{\text{ph}}}}\) is the coordinates of the principal points, and \(H_{\text{pic}}\) is the image height, which is consistent with the meaning of the internal and external parameters in photogrammetry.

3.3 Experiment and analysis

3.3.1 Experimental platform and data

Based on MATLAB, OpenCV, VSC#2012 and ArcEngine as an experimental platform, this paper integrates the algorithms implemented under various environments and develops a prototype system for the state perception of a group of individuals. The basic library of computer vision under MATLAB and OpenCV has many basic functions, such as video real-time access, video file playback, video frame acquisition, image size adjustment and so on. Based on this, the proposed algorithm is implemented and encapsulated as DLL. The related DLL is invoked under VSC#2012, and the prototype system is developed combined with the two development component ArcEngine of GIS.

In this paper, a number of video data containing people are collected from the restaurant and teaching building in a university, and QuickBird remote sensing images and high precision vector maps are collected in this area [12]. In order to obtain the internal and external reference of the camera, the method of extermination and extermination is used to calculate the main distance, and the external reference of the camera is obtained by combining high-definition remote sensing images, and the constraint rotation angle is 0. Thus, the mapping from video data to geospatial space is realized.

3.3.2 Experimental results and analysis

The experiment mainly includes three aspects: optical flow extraction, optical flow to geospatial mapping, and spatial analysis of optical flow. The video data of the restaurant and the teaching building are extracted from the restaurant and the teaching building, which reflects the trend of the crowd movement. As shown in Fig. 2, the arrow points to the direction of the optical flow, and its length represents the value of the optical flow. When someone moves normally in the crowd, the arrow tends to converge in a similar direction; otherwise, the arrow will be cluttered and the length will be longer.

Fig. 2
figure 2

Detection of abnormal behavior in people

Based on the mutual mapping model between video and image space and geospatial, the optical flow extracted from the video can be mapped to GIS in real time. In the form of arrows, GIS has clear spatial reference and measurement information, which can grasp the motion state of the regional population from the global perspective. As shown in Fig. 3, the prototype system displays the optical flow field at a certain time.

Fig. 3
figure 3

Optical flow field in a map

Based on GIS, spatial analysis of optical flow field can be carried out. Scattered point interpolation and contour generation experiments are carried out in this paper. The IDW method is selected for the scatter point interpolation. As shown in Fig. 4, we can see that the darker the color is, the faster we can understand the velocity field of the crowd movement according to the depth of the color. Interpolation-based DSM can automatically generate contour isoline distribution and also reflects the movement state of the crowd. The smaller the change in the value, the greater the distance between the lines. The faster the value goes up or down, the smaller the line spacing.

Fig. 4
figure 4

Optical flow interpolation

4 Variable length command sequence behavior detection model based on machine learning

The detection of abnormal events can be divided into two categories: one is the local anomaly detection, and an abnormal event is detected in a certain area of a given frame in a given video. The two is global anomaly detection. In a given video, it detects abnormal events at the beginning of a frame, aiming at the whole frame, rather than a certain area [13]. In general, the global anomaly detection means that each frame in the video is judged to be normal or abnormal. Since an early warning signal for one frame in the video can meet the requirements of video surveillance, this paper mainly studies the second methods, that is, global anomaly detection.

4.1 Description of variable length command sequence detection model

Lane et al. have two disadvantages: the fixed length command sequence detection model.

It lacks flexibility and adaptability in the representation of user behavior patterns. Behavior pattern refers to a certain regularity embodied in the process of user operation: in practice, the behavior patterns of different users are different, and the number of orders executed by the same user when completing different behavior patterns is not the same [14]. Therefore, the command sequence with fixed length is difficult to represent the overall behavior profile of users in a comprehensive and accurate way.

  1. 1.

    It is not easy to estimate the optimal sequence length for specific users. Lane et al. mainly use experimental methods to determine the optimal sequence length [15]. This method requires a large amount of computation, and its performance lacks stability.

  2. 2.

    We improved and corrected the above shortcomings of the fixed length command sequence detection model and proposed a variable length command sequence detection model, which is described as follows.

4.2 Algorithm flow

Figure 5 illustrates the process of crowd abnormal event detection in this paper. The purpose of crowd event detection is to classify each frame as normal or abnormal [16].

Fig. 5
figure 5

Crowd anomaly detection flowchart

The algorithm flow is as follows:

  1. 1.

    First, a grid segmentation is performed for each video frame to extract a series of particle points and each particle point is treated as an entity.

  2. 2.

    Calculate the reference interaction force of the training set. In the data set, a part of the normal video frames of the crowd is extracted and used as a training set. In most of the standard data sets, there will be an initial video segment that is normal pedestrian behavior [17]. This segment can be used as a reference. Let N be the number of samples of the training set, which is the number of particles extracted in each frame, which is the reference to the social interaction under normal conditions:

$$F_{r} = \frac{1}{N}\sum\limits_{i = 1}^{k} {F_{\text{int}} (x_{i} )}$$
(12)
  1. 3.

    For each frame of the test set, calculate the sum of the social forces of all particles, as shown in Eq. (13), and then calculate the distance between the social force of the current frame and the reference force, as shown in Eq. (14):

$$F_{t} = \sum\limits_{i = 1}^{k} {F_{\text{int}} (x_{i}^{\text{new}} )}$$
(13)
$$C_{t} = \left| {F_{t} - F_{r} } \right|$$
(14)

Finally, we use mean filtering to reduce the abrupt jump of a frame and get the filtered result after the difference of the magnitude of the interaction force calculated by formula (14). Next, each frame is divided into abnormal or normal classes according to the threshold value, as shown in formula 15:

$$L_{t} = \left\{ {\begin{array}{*{20}l} {{\text{Abnormal}},} \hfill & {{\text{if}}\;C_{t} > {\text{th}}} \hfill \\ {{\text{Normal}},} \hfill & {\text{otherwise}} \hfill \\ \end{array} } \right.$$
(15)

The whole process of crowd abnormal event detection algorithm based on variable length command sequence detection model is shown Fig. 6:

Fig. 6
figure 6

Anomaly behavior detection algorithm based on variable length command sequence detection model

Calculating the interaction force between crowds and shading is difficult. The particle flow calculation method provides another feasible method for the interaction force of the crowd. In densely populated individuals, the movement of individuals is similar to the gradual movement of a moving particle [18]. The optical flow in the crowd shows the direction of the crowd movement.

  1. 1.

    Crowd movements are captured by optical flow grids.

  2. 2.

    In the crowd, people tend to move smoothly, so the optical flow mean can be used to detect the convection of particles frequently.

A normal person in the crowd of interaction force diagram is shown in Fig. 7, in which red represents the calculated interaction force, yellow indicates the optical flow.

Fig. 7
figure 7

Interaction force of normal people in the population

5 Experimental results and analysis

The video data set selected in this paper comes from a square denser data set, which includes video frames for normal and abnormal behavior of the crowd. The data set has two different scenes, while the left of Fig. 8 represents the behavior of the normal crowd, and the left of Fig. 9 represents abnormal behavior. Normal and abnormal behaviors were tested, including normal walking and standing, abnormal behavior, and sudden agitation [19]. In order to simplify the calculation model, the grid particle resolution in the interactive force flow field is set to 20% of the total number of pixels, that is, a grid particle is extracted for every 5 pixels, and the panic factor is fixed to 0.25.

Fig. 8
figure 8

Social interaction force flow diagram under normal behavior of the population

Fig. 9
figure 9

Social interaction force flowchart under abnormal crowd behavior

First, for each video frame, the social interaction force is calculated. Figure 8 is an interactive force flow map under the normal behavior of the crowd. Blue indicates that the interaction force is smaller in the region, and the red indicates that the interaction force is larger. It can be seen that under normal conditions, the force flow is mainly distributed in one or two regions with low interaction force, which indicates that the social interaction force of the population is at a lower level.

Figure 8 shows a social interaction force flow diagram of one of the abnormal behaviors in the crowd video. It can be seen that the red area in the thermogram is more than that in Fig. 9. The red region indicates that the social interaction force is larger, that is, the activity of the crowd is more intense and frequent, and that the probability of the abnormal behavior of this video image is larger.

6 Conclusion

This paper proposes a new machine learning-based IDS video crowd and its behavior anomaly detection model, and uses Unix user’s shell command data to conduct experiments. Experimental results show that the detection performance of the new model is much better than that of Lane et al. Because the learning method and detection algorithm in the model have some adaptability to different detection data, this model can also be used for other data types (such as system call) outside the shell command (such as system call), but the specific application scope and detection performance need further research and experiment. How to further improve the speed, real-time immunity and robustness of the system is the focus of further research.