Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

3754 results about "Recognition algorithm" patented technology

The Text Recognition Algorithm Independent Evaluation (TRAIT) is being conducted to assess the capability of text detection and recognition algorithms to correctly detect and recognize text appearing in unconstrained imagery.

Emergency communications for the mobile environment

Systems and methods for two-way, interactive communication regarding emergency notifications and responses for mobile environments are disclosed. In accordance with one embodiment of the present invention, a specific geographic area is designated for selective emergency communications. The emergency communications may comprise text, audio, video, and other types of data. The emergency notification is sent to users' mobile communications devices such as in-vehicle telematics units, cellular phones, personal digital assistants (PDAs), laptops, etc. that are currently located in the designated area. The sender of the emergency message or the users' service provider(s) may remotely control cameras and microphones associated with the users' mobile communications devices. For example, a rear camera ordinarily used when driving in reverse may be used to capture images and video that may assist authorities in searching for a suspect. The users' vehicles may send photographs or video streams of nearby individuals, cars and license plates, along with real-time location information, in response to the emergency notification. Image recognition algorithms may be used to analyze license plates, vehicles, and faces captured by the users' cameras and determine whether they match a suspect's description. Advantageously, the present invention utilizes dormant resources in a highly beneficial and time-saving manner that increases public safety and national security.
Owner:MOTOROLA INC

Cross-camera pedestrian detection tracking method based on depth learning

The invention discloses a cross-camera pedestrian detection tracking method based on depth learning, which comprises the steps of: by training a pedestrian detection network, carrying out pedestrian detection on an input monitoring video sequence; initializing tracking targets by a target box obtained by pedestrian detection, extracting shallow layer features and deep layer features of a region corresponding to a candidate box in the pedestrian detection network, and implementing tracking; when the targets disappear, carrying out pedestrian re-identification which comprises the process of: after target disappearance information is obtained, finding images with the highest matching degrees with the disappearing targets from candidate images obtained by the pedestrian detection network and continuously tracking; and when tracking is ended, outputting motion tracks of the pedestrian targets under multiple cameras. The features extracted by the method can overcome influence of illuminationvariations and viewing angle variations; moreover, for both the tracking and pedestrian re-identification parts, the features are extracted from the pedestrian detection network; pedestrian detection, multi-target tracking and pedestrian re-identification are organically fused; and accurate cross-camera pedestrian detection and tracking in a large-range scene are implemented.
Owner:WUHAN UNIV

System and method for integrating and controlling audio/video devices

A processor integrating and controlling at least two A/V devices by constructing a control model, referred to as a filter graph, of the at least two A/V devices as a function of a physical connection topology of the at least two A/V devices and a desired content to be rendered by one of the at least two A/V devices. The filter graph may be constructed as a function of at least two device filters corresponding to the at least two A/V devices, in which the device filters include certain characteristics of the at least two A/V device. These characteristics may include the input or output pins for each device, the media type that the A/V device may process, the type of functions that the device may serve, etc. The desired content may be received as a user input which is entered via a keyboard, mouse or other comparable input devices. In addition, the user input may be entered as a voice command, which may be parsed by the processor using conventional speech recognition algorithms or natural language processing to extract the necessary information. Once the filter graph is constructed, the processor may control the at least two A/V devices via the filter graph by invoking predetermined operations on the filter graph resulting in the appropriate commands being sent to the at least two A/V devices, thereby results in the rendering of the desired content.
Owner:INTEL CORP

Named entities recognition method based on bidirectional LSTM and CRF

The invention discloses a named entities recognition method based on bidirectional LSTM and CRF. The named entities recognition method based on the bidirectional LSTM and CRF is improved and optimizedbased on the traditional named entities recognition algorithm in the prior art. The named entities recognition method based on the bidirectional LSTM and CRF comprises the following steps: (1) preprocessing a text, extracting phrase information and character information of the text; (2) coding the text character information by means of the bidirectional LSTM neural network to convert the text character information into character vectors; (3) using the glove model to code the text phrase information into word vectors; (4) combining the character vectors and the word vectors into a context information vector and putting the context information vector into the bidirectional LSTM neural network; and (5) decoding the output of the bidirectional LSTM with a linear chain condition random field to obtain a text annotation entity. The invention uses a deep neural network to extract text features and decodes the textual features with the condition random field, therefore, the text feature information can be effectively extracted and good effects can be achieved in the entity recognition tasks of different languages.
Owner:南京安链数据科技有限公司

3D (three-dimensional) convolutional neural network based human body behavior recognition method

InactiveCN105160310AThe extracted features are highly representativeFast extractionCharacter and pattern recognitionHuman bodyFeature vector
The present invention discloses a 3D (three-dimensional) convolutional neural network based human body behavior recognition method, which is mainly used for solving the problem of recognition of a specific human body behavior in the fields of computer vision and pattern recognition. The implementation steps of the method are as follows: (1) carrying out video input; (2) carrying out preprocessing to obtain a training sample set and a test sample set; (3) constructing a 3D convolutional neural network; (4) extracting a feature vector; (5) performing classification training; and (6) outputting a test result. According to the 3D convolutional neural network based human body behavior recognition method disclosed by the present invention, human body detection and movement estimation are implemented by using an optical flow method, and a moving object can be detected without knowing any information of a scenario. The method has more significant performance when an input of a network is a multi-dimensional image, and enables an image to be directly used as the input of the network, so that a complex feature extraction and data reconstruction process in a conventional recognition algorithm is avoided, and recognition of a human body behavior is more accurate.
Owner:XIDIAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products