Network security traceability semantic recognition method based on prompt self-supervised learning
A network security and semantic recognition technology, applied in neural learning methods, semantic analysis, secure communication devices, etc., can solve the problems of insufficient data set resources, high cost of supervision data, redundant memory information, etc., to achieve low cost, low cost. The effect of cost intelligent processing and low semantic extraction
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0033] Such as Figure 1 to Figure 3 As shown, this embodiment provides a method for semantic recognition of network security traceability based on prompt self-supervised learning.
[0034] This method uses the transformer model as the basic structure of the masked language model, and can carry out network security traceability semantic recognition on the text according to external information and contextual content, and can learn the hints of the masked language model for network security professional vocabulary that is not rich in data set resources. To fully mine the information of existing data, so as to achieve high-efficiency and low-cost network security traceability semantic recognition, so that machines can understand the intentions of real people.
[0035] First, given a chat record of a network security war room, the process of semantic traceability identification of this chat record is as follows:
[0036] Use all relevant text information in the professional fiel...
Embodiment 2
[0075] Such as Figure 1 to Figure 3 The entire process framework shown needs to be trained in advance, and the training phase is predicted in the same way as the testing phase, and the details are as follows:
[0076] Use the public dialogue dataset for pre-training: the pre-training task uses a mask language model based on prompt learning. When inputting text, a part of the sentence in the text is randomly covered. We set the ratio to 15%, of which the probability is 80%. Replace it with masked marks, 10% with other existing words, and 10% unchanged, and then let the deep model restore the masked replacement, and calculate the loss value between the predicted and real words.
[0077] After the pre-training is completed, the network model is fine-tuned 12,000 times with the open source dataset CMCSE (Comprehensive, Multi-Source Cyber-Security Events).
[0078] We initialize the network model with the parameters of the Chinese pre-trained Bert-base-cased released by Google, u...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More - R&D
- Intellectual Property
- Life Sciences
- Materials
- Tech Scout
- Unparalleled Data Quality
- Higher Quality Content
- 60% Fewer Hallucinations
Browse by: Latest US Patents, China's latest patents, Technical Efficacy Thesaurus, Application Domain, Technology Topic, Popular Technical Reports.
© 2025 PatSnap. All rights reserved.Legal|Privacy policy|Modern Slavery Act Transparency Statement|Sitemap|About US| Contact US: help@patsnap.com



