Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

114 results about "Learning by example" patented technology

Network violent video identification method

The invention discloses a network violent video identification method based on multiple examples and multiple characteristics. The method for identifying the network violent videos comprises the steps of grasping violent videos, non-violent videos, comments on the violent videos, comments on the non-violent videos, brief introductions of the violent videos and brief inductions of the non-violent videos from a video sharing network, and structuring a video data training set; extracting textural characteristics from textural information of the training set, forming textural characteristic vectors to train a textural pre-classifier, and screening out candidate violent videos by using the pre-classifier; using a shot segmentation algorithm based on a self-adapting dual threshold for conducting segmentation on video segments of the candidate violent videos, extracting related visual characteristics and voice frequency characteristics of each scene to express the scene, taking each scene as an example of multi-example study, and taking video segments as a package; and using an MILES algorithm for converting the package into a single example, using a characteristic vector for training a classifier model, and using the classifier model for conducting classification on the candidate violent videos. By the utilization of the network violence video identification method, bad influences that the network violent videos are broadcasted without constrain are largely lightened.
Owner:人民中科(北京)智能技术有限公司

Artificial intelligence-based multi-label classification method and system of multi-level text

The invention relates to an artificial intelligence-based multi-label classification method and system of multi-level text. The method includes: 1) utilizing a neural network to construct a multi-label classification model of the multi-level text, and obtaining text class prediction results of training text according to the model; 2) carrying out learning on parameters of the multi-label classification model of the multi-level text according to existing text class labeling information in the training text and the text class prediction results, which are of the training text and are obtained inthe step 1), to obtain a multi-label classification model of the multi-level text with determined parameters; and 3) utilizing the multi-label classification model of the multi-level text with the determined parameters to classify to-be-classified text. The method infers labels of the formed text simply through the document-level labeling information, and can be well applied to scenes where labels of formed text are difficult to collect; compared with traditional multi-instance learning (MIL) methods, the method of the invention introduces minimal assumptions, and can better fit actual data;and the method of the invention has good scalability.
Owner:INST OF INFORMATION ENG CHINESE ACAD OF SCI

Semantic propagation and mixed multi-instance learning-based Web image retrieval method

The invention belongs to the technical field of image processing and particularly provides a semantic propagation and mixed multi-instance learning-based Web image retrieval method. Web image retrieval is performed by combining visual characteristics of images with text information. The method comprises the steps of representing the images as BoW models first, then clustering the images according to visual similarity and text similarity, and propagating semantic characteristics of the images into visual eigenvectors of the images through universal visual vocabularies in a text class; and in a related feedback stage, introducing a mixed multi-instance learning algorithm, thereby solving the small sample problem in an actual retrieval process. Compared with a conventional CBIR (Content Based Image Retrieval) frame, the retrieval method has the advantages that the semantic characteristics of the images are propagated to the visual characteristics by utilizing the text information of the internet images in a cross-modal mode, and semi-supervised learning is introduced in related feedback based on multi-instance learning to cope with the small sample problem, so that a semantic gap can be effectively reduced and the Web image retrieval performance can be improved.
Owner:XIDIAN UNIV

Plate and strip steel surface defect detection method based on saliency label information propagation model

The invention relates to the technical field of industrial surface defect detection, and provides a plate strip steel surface defect detection method based on a significance label information propagation model. The method comprises the following steps of firstly, acquiring a plate strip steel surface image I; then, extracting a bounding box from the image I, and executing a bounding box selectionstrategy; then, performing super-pixel segmentation on the image I, and extracting a feature vector from each super-pixel; then, constructing a significance label information propagation model, constructing a training set based on a multi-example learning framework to train a classification model based on a KISVM, classifying a test set by using the trained model to obtain a category label matrix,calculating a smooth constraint item and a high-level prior constraint item, and optimizing and solving a diffusion function; and finally, calculating a single-scale saliency map under multiple scales, and obtaining a final defect saliency map through multi-scale fusion. The surface defects of the strip steel can be efficiently, accurately and adaptively detected, a complete defect target can beuniformly highlighted, and a non-significant background area can be effectively inhibited.
Owner:NORTHEASTERN UNIV

Face comparing verification method based on multi-instance learning

The invention discloses a face comparing verification method based on multi-instance learning, applied to a people and certificate consistency verification occasion. The method performs face comparing verification based on the thought of multi-instance learning, and comprises the following steps of: S1, face image preprocessing; S2, face multi-instance learning training; and S3, face verification. The face image preprocessing step comprises face detection, feature point locating and DoG light treatment; the face multi-instance learning training step comprises face multi-instance definition, multi-instance feature extraction and multi-instance feature fusion; and the face verification step is to perform face consistency verification according to stock equity of each instance and similarity of matched instances in the step S2. The method solves a difficult problem of change of hair style, skin color, making up, micro plastic, and so on, in face comparing verification, provides an effective algorithm and a train of thought for face verification, and improves reliability of face verification. The method provided by the invention can be widely applied to the people and certificate consistency verification occasion for checking whether certificates, such as a 2nd-generation ID card, a passport, a driving license and a student ID card, are held by owners thereof or not.
Owner:GUANGDONG MICROPATTERN SOFTWARE CO LTD

Image multi-tag marking algorithm based on multi-example package feature learning

The invention discloses an image multi-tag marking algorithm based on multi-example package feature learning, and the algorithm comprises the steps: obtaining a set of image blocks of all training images; extracting the features of a color histogram and the features of a direction gradient histogram of each image block of the set of the training images; enabling one training image to serve as an image package, and obtaining an image package structure needed by a multi-example learning framework; enabling the examples in all image packages in the set to form a projection example set, enabling each image package to be projected towards the projection example set, and obtaining the projection features of the image packages; selecting the features with the high discrimination performance as the classification features of the image packages; importing the classification features of the image packages of the learned training image set into an SVM classifier for training, obtaining the parameters of a training model, and predicting a test image tag through employing a trained SVM classifier. The algorithm is simple in implementation, and a trainer is mature and reliable. The algorithm is quick in prediction, and achieves multiple image tags better.
Owner:SHANDONG INST OF BUSINESS & TECH

Object segmentation method based on multiple-instance learning and graph cuts optimization

The present invention discloses an object segmentation method based on multiple-instance learning and graph cuts optimization. The method comprises the first step of carrying out salient model construction by adopting a multiple-instance learning method on training images, and predicting packages and instances of a testing image by using a salient model, thus to obtain a saliency testing result of the testing image; a second step of introducing the saliency testing result of the testing image into a graph-cut frame, optimizing the graph-cut frame according to instance characteristic vectors and marks of the instance packages, acquiring a second-best solution of graph cuts optimization, and obtaining precise segmentation of an object. According to the method provided by the present invention, the saliency testing model is constructed by using the multiple-instance learning method and thus is suitable for images of specific types, the saliency testing result is used into an image segmentation method based on the graph theory so as to guide image segmentation, a graph cut model frame link is optimized, an agglomerative hierarchical clustering algorithm is adopted for solving, the segmentation result can thus well accords to semantic aware output, and an accurate object segmentation result can be obtained.
Owner:CHANGAN UNIV

Weak supervision target detection method and system based on transfer learning

The invention discloses a weak supervision target detection method and system based on transfer learning; the method comprises the steps: extracting the features of an input strong supervision image and a weak supervision image through a deep convolutional neural network, extracting candidate boxes in the images through a region suggestion network, and obtaining the visual features of different candidate regions; performing feature extraction on category texts in the strong supervision data set and the weak supervision data set, establishing a semantic graph, and performing optimization by using a graph convolutional network to obtain semantic features of all category texts; employing dual-supervised average teacher network structure, which comprises a strong supervised classification and boundary regression student network, a weak supervised multi-instance learning student network and a classification and boundary regression teacher network; and aggregating bounding box information and classification information in the strong supervision data set and the weak supervision data set by using visual features and optimized semantic features, thereby performing bounding box regression and classification on candidate boxes. According to the invention, the weak supervision target detection effect is improved.
Owner:SHANGHAI JIAO TONG UNIV

Weighted extreme learning machine video target tracking method based on weighted multi-example learning

The invention discloses a weighted extreme learning machine video target tracking method based on weighted multi-example learning, solving the problem of bad tracking accuracy in the prior art. The method includes 1. initializing a Haar-like feature similar model pool and constructing a variety of feature model blocks, setting the weighted extreme learning machine network parameters; 2. extracting the training samples in the current frame and their feature blocks corresponding to the feature blocks of the different feature model blocks; 3. calculating the weighted multi-instance learning weight values; 4. constructing a plurality of networks corresponding to the different feature blocks and selecting the network with the largest similarity function value of the packet and the corresponding feature model block; 5. calculating the network global output weight values; 6. extracting the detection samples in the next frame and their corresponding feature blocks corresponding to the selected feature model blocks; 7. classifying the detection samples by means of the selected network and obtaining the target position of the next frame; and 8. repeating the above steps until the video is ended. According to the invention, the tracking accuracy is improved, and the target robustness tracking is realized.
Owner:XIDIAN UNIV

Electrocardiosignal classification method and device, electronic equipment and storage medium

The embodiment of the invention relates to an electrocardiosignal classification method and device, electronic equipment and a storage medium. The electrocardiosignal classification method disclosed by the embodiment of the invention comprises the following steps: acquiring electrocardiosignals; detecting a QRS complex wave from the electrocardiosignal; carrying out single heart beat cutting on the electrocardiosignals with the detected QRS composite waves and then packaging the electrocardiosignals into a plurality of heart beat packets; inputting the plurality of heart beat packets into a classifier model based on multi-example learning, obtaining a type identification result of each heart beat packet, wherein when the classifier model based on multi-example learning is used for classification, each heart beat packet is used as an example packet, each heart beat signal in the heart beat packet is used as an example in the example packet, and the type identification result comprises anormal rhythm type and an abnormal rhythm type. The electrocardiosignal classification method provided by the embodiment of the invention effectively reduces the time and labor cost spent in performing machine learning classification by manually labeling the heart beat.
Owner:GUANGZHOU SHIYUAN ELECTRONICS CO LTD

Three-branch convolutional network fabric defect detection method based on weak supervised learning

The invention provides a three-branch convolutional network fabric defect detection method based on weak supervised learning, and the method comprises the steps: firstly, building a multi-example learning detection network based on a mutual exclusion principle in a weak supervised network, so as to carry out the training through an image-level label; then, establishing a three-branch network framework, and adopting a long connection structure so as to extract and fuse the multi-level convolution feature map; utilizing the SE module and the cavity convolution to learn the correlation between channels and expand the convolution receptive field; and finally, calculating the positioning information of the target by using a class activation mapping method to obtain the attention mapping of thedefect image. According to the method, the problems of rich textural features and defect label missing contained in the fabric picture are comprehensively considered, and by adopting a weak supervision network mechanism and a mutual exclusion principle, the representation capability of the fabric picture is improved while the dependence on the label is reduced, so that the detection result has higher detection precision and adaptivity.
Owner:ZHONGYUAN ENGINEERING COLLEGE

Few-sample pedestrian re-identification method based on deep multi-example learning

The invention relates to a few-sample pedestrian re-identification method based on deep multi-example learning. The method comprises three stages: a network pre-training stage, a data set expansion stage and a network fine tuning stage. The method includes: after the pedestrian re-identification feature extraction sub-network is pre-trained, performing data expansion by utilizing a pedestrian keypoint feature region exchange algorithm; finely adjusting the pedestrian re-identification feature extraction sub-network and the feature aggregation sub-network by using the expanded data set; iteratively repeating data set expansion and network fine tuning until the feature extraction sub-network and the feature aggregation sub-network converge. Once the training is completed, the pedestrian re-identification model on the original domain is migrated and extended to the target domain by using few samples. On the premise of giving a small number of learning samples in the target domain, the pedestrian re-identification model can be effectively migrated and expanded to the target domain monitoring network, and the few-sample pedestrian re-identification method has the advantages of high accuracy, good robustness, good expansibility and mobility.
Owner:FUDAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products