Unsupervised deep auto-encoding network-based unknown threat detection method and system for HTTP data
A self-encoding network and unknown threat technology, applied in digital transmission systems, transmission systems, neural learning methods, etc., can solve problems such as difficult to deal with attackers' attack methods, ineffective and comprehensive detection, and difficult to effectively detect unknown threats
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0064] like figure 1 As shown in the present invention, an unknown threat detection method based on unsupervised deep self-encoding network for HTTP data, comprising the following steps:
[0065] S101, data access, access HTTP request data;
[0066] S102, data cleaning, cleaning the HTTP request data;
[0067] S103, feature extraction, performing feature extraction on the cleaned HTTP data;
[0068] S104, model matching, performing model matching on the extracted feature data, where the model matching algorithm includes an HTTP request hierarchical distribution algorithm and an unsupervised deep self-encoding model algorithm;
[0069] The HTTP request hierarchical distribution algorithm can split the HTTP request data into five layers: ACTION, URL, PROTO, HEADERS, and BODY (such as Figure 5 shown), and respectively perform S103 to complete the scalar processing, and then hand over to different unsupervised deep self-encoding model algorithms respectively;
[0070] The uns...
Embodiment 2
[0107] An unknown threat detection system based on unsupervised deep self-encoding network, the specific implementation is as follows:
[0108] like figure 1 As shown in the figure, the unknown threat detection system accesses the non-abnormal data filtered by the traditional protection system, and cleans the HTTP data after receiving the HTTP data. There are a lot of duplicate HTTP data in the Internet, for example, different people visit a web page at the same time , so that the HOST, URL and other parameters of the HTTP data are basically the same. These same data will bring additional pressure to the system. It is necessary to use a whitelist method to deduplicate the duplicate data. Then the data needs to be encoded and decoded. In the process of network transmission, the data will be encoded due to the problems of privacy data and character ambiguity. Here, the encoded data needs to be restored. Recognition such as: base64 encoding, URL encoding, and then do the corresp...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


