Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

103 results about "Visual attentiveness" patented technology

Video description generation system based on graph convolution network

The invention belongs to the technical field of cross-media generation, and particularly relates to a video description generation system based on a graph convolution network. The video description generation system comprises a video feature extraction network, a graph convolution network, a visual attention network and a sentence description generation network. The video feature extraction network performs sampling processing on videos to obtain video features and outputs the video features to the graph convolution network. The graph convolution network recreates the video features accordingto semantic relations and inputs the video features into sentence descriptions to generate a recurrent neural network; and the sentence description generation network generates sentences according tofeatures of video reconstruction. The features of a frame-level sequence and a target-level sequence in the videos are reconstructed by adopting graph convolution, and the time sequence information and the semantic information in the videos are fully utilized when description statements are generated, so that the generation is more accurate. The invention is of great significance to video analysisand multi-modal information research, can improve the understanding ability of a model to video visual information, and has a wide application value.
Owner:FUDAN UNIV

Full convolutional network fabric defect detection method based on attention mechanism

The invention provides a full convolutional network fabric defect detection method based on an attention mechanism. The full convolutional network fabric defect detection method comprises the steps: firstly extracting a multi-stage and multi-scale intermediate depth feature map of a fabric image through an improved VGG16 network, and carrying out the processing through the attention mechanism, andobtaining a multi-stage and multi-scale depth feature map; then, performing up-sampling on the multi-level and multi-scale depth feature maps by utilizing bilinear interpolation to obtain multi-levelfeature maps with the same size, and performing fusion by utilizing a short connection structure to obtain a multi-level saliency map; and finally, fusing the multistage saliency maps by adopting weighted fusion to obtain a final saliency map of the defect image. According to the full convolutional network fabric defect detection method, complex defect characteristics and various backgrounds of the fabric image are comprehensively considered, and the representation capability of the fabric image is improved by simulating an attention mechanism of human visual attention cognition, and the noise influence in the image is eliminated, so that the detection result has higher adaptivity and detection precision.
Owner:ZHONGYUAN ENGINEERING COLLEGE

Key frame extraction method based on visual attention model and system

The invention discloses a key frame extraction method based on a visual attention model and a system. In a spatial domain, the extraction method uses binomial coefficients to filter the global contrast for salience detection, and uses an adaptive threshold for carrying out extraction on a target region. The algorithm can well maintain the salient target region boundary, and the salience in the region is uniform. Then, in a time domain, the method defines the motion salience, motion of the target is estimated via a homography matrix, a key point is adopted for replacing the target for salience detection, data of salience in the spatial domain is converged, and a boundary extension method based on an energy function is brought forward to acquire a bounding box to serve as the salient target region of the time domain. Finally, the method reduces richness of the video through the salient target region and an online clustering lens adaptive method is adopted for key frame extraction.
Owner:SUN YAT SEN UNIV +1

Alzheimer's disease rehabilitation training and ability evaluation system based on virtual reality

The invention provides an Alzheimer's disease rehabilitation training and ability evaluation system based on virtual reality. The system comprises a rehabilitation training module, an ability evaluation module and a feedback module, wherein the rehabilitation training module is used for conducting rehabilitation training on an Alzheimer's disease patient in the aspects of cognition and social contact; the ability evaluation module is used for collecting the visual attention information, the behavior mode information and social communication ability information generated by the rehabilitation training module, carrying out comprehensive evaluation according to the information and generating a comprehensive evaluation result; and the feedback module is used for feeding back the evaluation result to the rehabilitation training module and changing the content and process of rehabilitation training. According to the invention, the behavior mode of the testee is amplified by using the virtualreality technology, the data is analyzed and processed by combining an artificial intelligence technology, a comprehensive evaluation result can be obtained by comparing the data with a standard control group and an Alzheimer's disease control group, and the content and process of rehabilitation training can be adjusted by combining real-time feedback and stage feedback to adapt to the current state of the Alzheimer's disease patient.
Owner:JIAXING SEC VIEW INFORMATION TECH CO LTD

Text and image fused bimodal character classification method and device

ActiveCN112949622AAddress limitationsSolve the problem of not being able to capture information about cognitive differencesCharacter and pattern recognitionNeural architecturesFeature extractionImaging Feature
The invention relates to a text and image fused bimodal character classification method and device, and belongs to the technical field of artificial intelligence. The method comprises the following steps: inputting text data and image data into a pre-trained character classification network, and obtaining a character classification result, The character classification network comprises a feature extraction network, a contrast visual attention network and a contrast perception decoding network; a text feature extraction branch in the feature extraction network is used for extracting a word embedding vector of the text dataa text feature extraction branch in the feature extraction network, and an image feature extraction branch is used for extracting an image region vector of the image dataan image feature extraction branch; a basic visual attention branch in the contrast visual attention network is used for extracting an image object aligned with the text data and calculating aligned visual representation, and an inverse visual attention branch is used for extracting an image object not aligned with the text data and calculating non-aligned visual representation the inverse visual attention branch; and the comparison perception decoding network is used for predicting character categories. The problems that the classification performance is poor and cognitive difference information cannot be captured are solved.
Owner:SUZHOU UNIV

Intelligent screening method and system thereof for abnormal tongue images

The invention provides an intelligent screening method and an intelligent screening system for abnormal tongue image. The intelligent screening method and the intelligent screening system have the beneficial effects that: by constructing a knowledge platform of traditional Chinese medicine tongue diagnosis and an expert auxiliary diagnosis system, an objective and quantitative measurement means isprovided for tongue diagnosis information judgment, and the diagnosis correctness is improved; a fine-grained image classification algorithm based on an attention mechanism is adopted, and the attention mechanism reflects the perception difference of a human vision system to the surrounding environment, that is, attention is focused on a salient region in the environment; a convolutional neural network is used for simulating visual attention properties of human beings, so that information of classification targets is fully utilized, and the accuracy rate of tongue image fine-grained image classification can be improved; under the inspiration of the visual attention mechanism, spatial position features are introduced into classification and modeling is carried out in the process of utilizing the fine-grained image classification algorithm, namely, a plurality of key regions of a tongue image are positioned, modeling is carried out according to the spatial relationship among all parts of the tongue image, and the integrity and discrimination of part features can be improved.
Owner:中润普达(十堰)大数据中心有限公司

Teaching quality evaluation method, device and system based on eye movement tracking

PendingCN112070641AQuantify Learning EfficiencyAchieving Quality of LearningInput/output for user-computer interactionData processing applicationsMedicineEngineering
The invention discloses a teaching quality evaluation method based on eye movement tracking, and the method comprises the steps: calculating a visual attention score of a student according to the correlations between a visual path of the student and a cursor path in a teaching process of a teacher, a display path of course contents in the teaching process of the teacher, and a teaching content sequence in the teaching process of the teacher; and generating the attention mode of the student according to the relationship between the attention score and the preset attention score threshold, so that the attention of the student in the online learning process is visualized, and the learning efficiency of the student can be quantified. Furthermore, according to the attention mode, the quantitative index and the assessment result of the previous course of the same subject of the student, the attention mode, the quantitative index and the assessment result of the future course of the subject of the student are predicted, and prediction of future learning quality and assessment results of the student is realized. In addition, the invention also discloses a teaching quality evaluation deviceand system based on eye movement tracking, and a computer readable storage medium.
Owner:东莞市东全智能科技有限公司

Fine-grained image retrieval method based on self-attention mechanism weighting

The invention relates to the technical field of image retrieval and computer vision, in particular to a fine-grained image retrieval method based on visual attention mechanism weighting. The method comprises the following steps of: image preprocessing: setting the length of the longest side of an image to be 500 pixels; feature extraction: inputting the image into a convolutional neural network, and then selecting and outputting the features of the last convolutional layer; target feature selection: firstly, optimizing a local activation graph, and then selecting a local feature vector according to an activation graph result, so as to realize more accurate target feature selection; feature weighted aggregation: evaluating the importance degree of each feature, so as to enable the weightedfine-grained local features to still be embodied during pooling aggregation and improve the precision of fine-grained retrieval; and performing image retrieval, and calculating cosine similarity between the characteristic vectors of the queried image and a database image. An image feature extraction and coding detail graph is shown in figure 1. According to the method, fine-grained image retrievalcan be realized, and the retrieval accuracy is improved.
Owner:HUNAN UNIV

Air conditioner outdoor unit portrait intelligent detection method based on visual attention and multi-scale convolutional neural network

The invention relates to an air conditioner outdoor unit portrait intelligent detection method based on visual attention and a multi-scale convolutional neural network, and the method comprises the following steps: (1) data preprocessing: carrying out the manual classification of an air conditioner outdoor unit portrait sample, and generating a correct label and a wrong label; (2) reading the preprocessed sample image, inputting the preprocessed sample image into a visual attention network, and generating an attention distribution diagram; (3) inputting into a multi-scale network for training to obtain a deep fusion feature vector; (4) taking the deep fusion feature vector as the input of a softmax classifier model for training; (5) inputting the verification sample set into a softmax classifier model to verify the classification precision to obtain a trained softmax classifier model; and (6) inputting the test sample set into the trained softmax classifier model to obtain a correct or wrong classification result of the test sample set. And the conduction gradient is helped in the reverse process, so that a deeper model can be successfully trained, and the performance of the network is improved.
Owner:SHANDONG UNIV

Task scene associated unmanned aerial vehicle pilot visual attention distribution mode extraction method

The invention discloses a task scenario associated unmanned aerial vehicle pilot visual attention distribution mode extraction method, which comprises the steps of obtaining pilot attention distribution data through an eye tracker, dividing different task scenarios through flight parameter data, extracting pilot physiological signals to divide fatigue levels of the pilot physiological signals, anddistinguishing pilot attention distribution situations under different task scenes and different fatigue levels. An eye tracker collects eye movement data when a pilot executes a task under the simulation platform; performs segmentation clustering on the flight parameter high-dimensional time sequence, and dividing different task scenes; collects pilot electrocardiogram data for time-frequency domain analysis, and establishes a fatigue state discrimination classifier in combination with eye movement data and auditory response; attention distribution characteristics are represented through thepercentage of fixation points in the region of interest, the number of times of review and the importance degree, the attention distribution situation with the highest support degree is extracted toserve as a mode, the guidance effect can be achieved on control over pilots of the unmanned aerial vehicle, and the important significance is also achieved on optimization of interface setting of an unmanned aerial vehicle control platform.
Owner:NORTHWESTERN POLYTECHNICAL UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products