Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

466 results about "False recognition" patented technology

Transparent monitoring and intervention to improve automatic adaptation of speech models

A system and method to improve the automatic adaptation of one or more speech models in automatic speech recognition systems. After a dialog begins, for example, the dialog asks the customer to provide spoken input and it is recorded. If the speech recognizer determines it may not have correctly transcribed the verbal response, i.e., voice input, the invention uses monitoring and if necessary, intervention to guarantee that the next transcription of the verbal response is correct. The dialog asks the customer to repeat his verbal response, which is recorded and a transcription of the input is sent to a human monitor, i.e., agent or operator. If the transcription of the spoken input is correct, the human does not intervene and the transcription remains unmodified. If the transcription of the verbal response is incorrect, the human intervenes and the transcription of the misrecognized word is corrected. In both cases, the dialog asks the customer to confirm the unmodified and corrected transcription. If the customer confirms the unmodified or newly corrected transcription, the dialog continues and the customer does not hang up in frustration because most times only one misrecognition occurred. Finally, the invention uses the first and second customer recording of the misrecognized word or utterance along with the corrected or unmodified transcription to automatically adapt one or more speech models, which improves the performance of the speech recognition system.
Owner:AVAYA INC

Apparatus and System for Recognizing Environment Surrounding Vehicle

In conventional systems using an onboard camera disposed rearward of a vehicle for recognizing an object surrounding the vehicle, the object is recognized by the camera disposed rearward of the vehicle. In the image recognized by the camera, a road surface marking taken by the camera appears at a lower end of a screen of the image, which makes it difficult to predict a specific position in the screen from which the road surface marking appears. Further, an angle of depression of the camera is large, and it is a short period of time to acquire the object. Therefore, it is difficult to improve a recognition rate and to reduce false recognition. Results of recognition (type, position, angle, recognition time) made by a camera disposed forward of the vehicle, are used to predict a specific timing and a specific position of a field of view of a camera disposed rearward of the vehicle, at which the object appears. Parameters of recognition logic of the rearwardly disposed camera and processing timing are then optimally adjusted. Further, luminance information of the image from the forwardly disposed camera is used to predict possible changes to be made in luminance of the field of view of the rearwardly disposed camera. Gain and exposure time of the rearwardly disposed camera are then adjusted.
Owner:HITACHI LTD

Pedestrian re-identification method and system and computer readable storage medium

The invention provides a pedestrian re-identification method and system, a computer readable storage medium. The pedestrian re-identification method comprises the following steps: obtaining a calibration data set, and training the calibration data set to form a segmentation model; acquiring a pedestrian image, and segmenting the background of the pedestrian image to obtain a foreground image and an environment image; extracting body-shaped key points of pedestrians in the foreground image containing the pedestrians, and segmenting the foreground image based on the body-shaped key points to form an ROI; extracting features of the foreground image and the ROI of the region of interest based on a feature extraction model to obtain global features and weighted features, and connecting the global features and the weighted features in series to form a multi-dimensional feature vector; and performing similarity comparison on the multi-dimensional feature vector and features extracted from thetarget pedestrian to determine whether the pedestrian is the target pedestrian. By removing background images of pedestrians captured under different cameras, redundant features during feature extraction are eliminated, recognition results of pedestrian re-recognition are only based on pure features, and the occurrence of false recognition is reduced.
Owner:艾特城信息科技有限公司

Finger vein recognition method fusing local features and global features

The invention discloses a finger vein recognition method fusing local features and global features. At present, a number of vein recognition methods adopt the local features of a vein image, so that the recognition precision of the vein recognition methods is greatly affected by the quality of the image; the phenomena of rejection and false recognition are liable to appear. The finger vein recognition method provided by the invention comprises the following steps: firstly, performing pretreatment operations such as finger area extraction of a read-in finger vein image, binarization and the like; then, according to the point set of extracted detail features, realizing the matching of the local features within a certain angle and a certain radius by virtue of a flexible matching-based local feature recognition module; using a global feature recognition module for vein image recognition to realize the matching of the global features as the global feature recognition module is used for analyzing bidirectional two-dimensional principal components and can better display a two-dimensional image data set on the whole; finally, designing weights according to the correct recognition rates of the two recognition methods, performing decision-level fusion to the results of two classifiers, and taking the fused result as a final recognition result. The method is applied to finger vein recognition.
Owner:HEILONGJIANG UNIV

Speech recognition attack defense method based on PSO algorithm

The invention discloses a speech recognition attack defense method based on a PSO algorithm. The method comprises the following steps of (1) preparing an original audio data set, and dividing the original audio data set into a pre-training data set, a test data set and a disturbance data set used for generating an adversarial sample; (2) training a speech recognition model, wherein the speech recognition model is built, related parameters of the model are initialized, the speech recognition model is trained by using the pre-training data set, and the test data set is used to test the recognition accuracy of the model; (3) attacking the speech recognition model, wherein an attack method based on the PSO algorithm is built, a fitness function and related parameters of the PSO algorithm are set, and an optimal adversarial sample generated by the attack method can be recognized by mistake and is not recognized by the human ear; and (4) performing adversarial training on the speech recognition model, wherein the adversarial sample generated in the step (3) is added into the pre-training data set, and the speech recognition model is re-trained, so that the speech recognition model has the capability of defending attack of the adversarial sample and the safety and stability of the model are improved.
Owner:ZHEJIANG UNIV OF TECH

Target positioning method and device and unmanned aerial vehicle

The invention relates to a target positioning method and device and an unmanned aerial vehicle. The method comprises the steps of acquiring a target image through an image acquisition device; acquiring position information of the target in the target image according to the target image; obtaining an original point cloud of an environment in front of the unmanned aerial vehicle through a depth sensor; obtaining a point cloud corresponding to the target in the original point cloud according to the position information of the original point cloud and the target in the target image; and acquiringposition information of the target in a three-dimensional space according to the point cloud corresponding to the target. According to the invention, the depth information of the target is obtained bydetermining the position information of the three-dimensional space of the target; a motion estimation model with higher stability and precision can also be provided for target tracking, the probability of false recognition and tracking loss is reduced, more accurate visualization of three-dimensional path planning and real-time three-dimensional reconstruction can be achieved, and meanwhile a three-dimensional map with a target object can be used for obstacle avoidance.
Owner:SHENZHEN AUTEL INTELLIGENT AVIATION TECH CO LTD

A Chinese scene text line identification method based on residual convolution and a recurrent neural network

The invention discloses a Chinese scene text line identification method based on residual convolution and a recurrent neural network. The method comprises the following steps: collecting a Chinese scene text training image, performing normalization processing on the size of the training image, performing data augmentation processing on the training image, designing a residual convolutional neuralnetwork, a residual recurrent neural network and a CTC model, training horizontal text lines and vertical text lines, and selecting a result with higher confidence as an identification result. According to the invention, the convolutional neural network and the recurrent neural network are combined; the problem of Chinese scene text line identification is solved; error recognition caused by character segmentation and error segmentation on the text lines is avoided; the training of the residual error connection acceleration model is added into the convolutional neural network and the recurrentneural network, so that a practical Chinese scene text recognition model is obtained, the robustness is high, and Chinese text lines with complex backgrounds, complex illumination and various fonts can be recognized.
Owner:SOUTH CHINA UNIV OF TECH +1
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products