Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

449results about How to "Expand the receptive field" patented technology

A retinal blood vessel image segmentation method based on a multi-scale feature convolutional neural network

The invention belongs to the technical field of image processing, in order to realize automatic extraction and segmentation of retinal blood vessels, improve the anti-interference ability to factors such as blood vessel shadow and tissue deformation, and make the average accuracy rate of blood vessel segmentation result higher. The invention relates to a retinal blood vessel image segmentation method based on a multi-scale feature convolutional neural network. Firstly, retinal images are pre-processed appropriately, including adaptive histogram equalization and gamma brightness adjustment. Atthe same time, aiming at the problem of less retinal image data, data amplification is carried out, the experiment image is clipped and divided into blocks, Secondly, through construction of a multi-scale retinal vascular segmentation network, the spatial pyramidal cavity pooling is introduced into the convolutional neural network of the encoder-decoder structure, and the parameters of the model are optimized independently through many iterations to realize the automatic segmentation process of the pixel-level retinal blood vessels and obtain the retinal blood vessel segmentation map. The invention is mainly applied to the design and manufacture of medical devices.
Owner:TIANJIN UNIV

Text detection method, system and equipment based on multi-receptive field depth characteristics and medium

The invention discloses a text detection method, system and device based on multi-receptive field depth characteristics and a medium, and the method comprises the steps: obtaining a text detection database, and taking the text detection database as a network training database; building a multi-receptive field depth network model; inputting a natural scene text picture and corresponding textbox coordinate true value data in the network training database into a multi-receptive field depth network model for training; calculating an image mask for segmentation through the trained multi-receptive field depth network model to obtain a segmentation result, and converting the segmentation region into a regression textbox coordinate; and counting the textbox size of the network training database, designing a textbox filtering condition, and screening out a target textbox according to the textbox filtering condition. The method fully utilizes the feature learning capability and classification performance of the deep network model, combines the characteristics of image segmentation, has the characteristics of high detection accuracy, high recall rate, strong robustness and the like, and has agood text detection effect in a natural scene.
Owner:SOUTH CHINA UNIV OF TECH

Multi-speaker voice separation method based on convolutional neural network and depth clustering

The invention discloses a multi-speaker voice separation method based on a convolutional neural network and depth clustering. The method comprises the following steps: 1, a training stage: respectively performing framing, windowing and short-time Fourier transform on single-channel multi-speaker mixed voice and corresponding single-speaker voice; and training mixed voice amplitude frequency spectrum and single-speaker voice amplitude frequency spectrum as an input of a neural network model; 2, a testing stage: taking the mixed voice amplitude frequency spectrum as an input of a threshold expansion convolutional depth clustering model to obtain a high-dimensional embedded vector of each time-frequency unit in the mixed frequency spectrum; using a K-means clustering algorithm to classify thevectors according to a preset number of speakers, obtaining a time-frequency masking matrix of each sound source by means of the time-frequency unit corresponding to each vector, and multiplying thematrixes with the mixed voice amplitude frequency spectrum respectively to obtain a speaker frequency spectrum; and combining a mixed voice phase frequency spectrum according to the speaker frequencyspectrum, and obtaining a plurality of separate voice time domain waveform signals by adopting short-time Fourier inverse transform.
Owner:XINJIANG UNIVERSITY

Magnetic resonance multi-channel reconstruction method based on deep learning and data self-consistence

The invention discloses a magnetic resonance multi-channel reconstruction method based on deep learning and data self-consistence and belongs to the magnetic resonance reconstruction method field. Themethod includes the following steps that: multi-channel full-sampling training set data are collected, under-sampled data are converted into wrap-around images which are used as the input of a reconstruction network, and full-sampled data are used as training mark data, and a convolution kernel is generated based the full-sampled data; 2, the data in the step 1 are inputted, on the basis of the reconstruction network constructed by means of the repeated superposition of SC layers, CNNs, and DC layers, and with the training mark data adopted as an objective, network parameters are trained withback propagation, so that mapping relationships between the input and output of the reconstruction network are obtained; and 3, test set data are inputted into the reconstruction network so as to besubjected to forward propagation, so that unknown mapping data are obtained, and the reconstruction of magnetic resonance is completed. With the method adopted, a problem that an existing resonance reconstruction method can only process single-channel data can be solved; more stable and accurate end-to-end mapping relationships can be obtained; the quality of magnetic resonance reconstruction canbe fundamentally improved; and magnetic resonance scanning time can be significantly shortened.
Owner:朱高杰

Lightweight deep network image target detection method suitable for Raspberry Pi

A lightweight deep network image target detection method applicable to Raspberry Pi belongs to the field of deep learning and target detection, and comprises the following steps: firstly, collecting an image containing a to-be-detected target, and preprocessing the collected image for network training; secondly, inputting the preprocessed image into a depth separable expansion convolutional neural network for feature extraction to obtain feature maps with different resolutions; inputting the feature maps with different resolutions into a feature pyramid network for feature fusion, and generating a fusion feature map carrying richer information; and then carrying out classification and positioning of a to-be-detected target on the fusion feature map by adopting a detection network, and finally carrying out non-maximum suppression to obtain an optimal target detection result. The image target detection method based on the deep neural network overcomes the difficulties that the image target detection method based on the deep neural network is difficult to realize on the Raspberry Pi platform and the image target detection method based on the lightweight network is low in detection accuracy on the Raspberry Pi platform.
Owner:BEIJING UNIV OF TECH

Remote sensing image terrain classification method based on lightweight semantic segmentation network

ActiveCN111079649APreserve spatial featuresMulti-context featuresScene recognitionData setTest sample
The invention discloses a remote sensing image terrain classification method based on a lightweight semantic segmentation network, and mainly solves the problems of low remote sensing image terrain classification precision and low training speed caused by insufficient utilization of image space and channel feature information and a huge model in an existing method. According to the scheme, the method includes obtaining a training sample and a test sample in a remote sensing image terrain classification data set; constructing and introducing a lightweight remote sensing image terrain classification model capable of broadening channel decomposition hole convolution, and designing an overall loss function of a concerned terrain edge; inputting a training sample into the constructed terrain classification model for training to obtain a trained model; and inputting the test sample into the trained model, and predicting and outputting a terrain classification result in the remote sensing image. According to the method of the invention, the feature expression capability is improved, the network parameters are reduced, the average precision and the training speed of remote sensing image terrain classification are improved, and the method can be used for obtaining the terrain distribution condition of a remote sensing image.
Owner:XIDIAN UNIV

Dilated causal convolution generative adversarial network end-to-end bone conduction speech blind enhancement method

The invention relates to the field of artificial intelligence and medical rehabilitation instruments, aims to provide an end-to-end bone conduction speech enhancement method, and solves the problems of the absence of high-frequency components of bone conduction speech, poor auditory perception, communication under the background of strong noise and the like. According to the dilated causal convolution generative adversarial network end-to-end bone conduction speech blind enhancement method, a bone conduction original audio sampling point is taken as input data, a pure air conduction original audio is taken as an output target of training, bone conduction speech is input into a trained dilated causal convolution generative adversarial network, the dilated causal convolution generative adversarial network comprises a generator and a discriminator, the generator adopts dilated causal convolution, and an enhanced sample is output; the discriminator inputs original audio data and the enhanced speech sample generated by the generator, and a convolution layer in the discriminator is used for extracting deep nonlinear features, thereby performing deep similarity judgment of the sample. Thedilated causal convolution generative adversarial network end-to-end bone conduction speech blind enhancement method is mainly applied to the design and manufacturing occasion of bone conduction speech enhancement equipment.
Owner:TIANJIN UNIV

Convolutional neural network-based human face expression identification method

The invention relates to a human face expression identification method, in particular to a convolutional neural network-based human face expression identification method. The method comprises the steps of firstly obtaining human face images from a video, performing scale normalization operation processing on the obtained human face images to obtain human face images same in size, and performing alignment preprocessing operation on the human face images same in size to obtain preprocessed human face images; and performing feature extraction operation on the preprocessed human face images by using a convolutional neural network to obtain features of the human face images, and performing classified identification operation on the features of the human face images by utilizing a Softmax classifier. A human face expression identification algorithm realized by utilizing the convolutional neural network is an end-to-end process. According to the method, the human face images only need to be simply preprocessed and then fed into the convolutional neural network, the feature extraction is automatically performed, and a classification result is given, so that the accuracy is greatly improved, adjustable parameters are reduced, and intermediate processing steps are simplified to a great extent.
Owner:ANHUI SUN CREATE ELECTRONICS

Blood vessel and fundus image segmentation method, device and equipment and readable storage medium

The embodiment of the invention discloses a blood vessel and eye fundus image segmentation method, device and equipment and a readable storage medium, and relates to the computer vision technology ofartificial intelligence. Specifically, the method comprises steps of acquiring a blood vessel image to be segmented, such as a fundus image; performing feature extraction on the blood vessel image such as the fundus image to obtain high-level feature information; performing dictionary learning on the high-level feature information based on a preset dictionary to obtain dictionary representation corresponding to the high-level feature information; selecting a plurality of channels of the high-level feature information according to the dictionary representation to obtain target feature information; fusing the target feature information with the high-level feature information to obtain channel attention feature information; and segmenting blood vessels in the blood vessel image, such as the fundus image, according to the channel attention feature information to obtain a blood vessel segmentation result. According to the scheme, global information loss of the characteristic blood vessel image such as the fundus image can be avoided, and the segmentation accuracy of the blood vessel image such as the fundus image is greatly improved.
Owner:腾讯医疗健康(深圳)有限公司

SAR sequence image classification method based on space-time joint convolution

ActiveCN110781830AImprove effectivenessOvercome the problem of destroying the time information of sequence imagesScene recognitionNeural architecturesTime informationGoal recognition
The invention discloses an SAR (Synthetic Aperture Radar) sequence image classification method based on space-time joint convolution, which mainly solves the problems of insufficient time informationutilization and low classification accuracy due to the fact that only single image features are utilized in the existing SAR target recognition technology. The method comprises the following steps: 1)generating a sample set, and generating a training sequence sample set and a test sequence sample set from the sample set; 2) constructing a space-time joint convolutional neural network; 3) traininga space-time joint convolutional neural network by using the training sequence sample set to obtain a trained space-time joint convolutional neural network; and 4) inputting the test sequence sampleset into the trained space-time joint convolutional neural network to obtain a classification result. According to the method, the space-time joint convolutional neural network is utilized to extractthe change characteristics of the time dimension and the space dimension of the SAR sequence image, and the accuracy of SAR target classification and recognition is improved. The method can be used for automatic target identification based on SAR sequence images.
Owner:XIDIAN UNIV

Image processing method and device and processing equipment

The invention provides an image processing method and device and processing equipment, and relates to the technical field of image recognition, and the method comprises the steps: obtaining a to-be-recognized image; inputting the image to be identified into a target identification network; wherein the target identification network comprises a plurality of convolution calculation layers and a plurality of residual error calculation layers which are connected in sequence; wherein the convolution calculation layer comprises a convolution block, and the residual calculation layer comprises a residual block; The residual block comprises at least two convolution blocks which are sequentially connected; The convolution block comprises at least one channel invariable convolution layer; when the channel invariant convolution layer calculates the input feature map, each channel of the input feature map is independently subjected to convolution transformation to obtain one channel of the output feature map; and performing posture recognition on the to-be-recognized image through the target recognition network to obtain a posture recognition result, the posture recognition result comprising the position and the mode of the target contained in the to-be-recognized image. According to the embodiment of the invention, the calculation amount can be reduced, the receptive field is increased, and the position and mode are accurately determined.
Owner:BEIJING KUANGSHI TECH

Road scene segmentation method based on residual network and expanded convolution

The invention discloses a road scene segmentation method based on a residual network and expanded convolution. The method comprises: a convolutional neural network being constructed in a training stage, and a hidden layer of the convolutional neural network being composed of ten Respondial blocks which are arranged in sequence; inputting each original road scene image in the training set into a convolutional neural network for training to obtain 12 semantic segmentation prediction images corresponding to each original road scene image; calculating a loss function value between a set formed by12 semantic segmentation prediction images corresponding to each original road scene image and a set formed by 12 independent thermal coding images processed by a corresponding real semantic segmentation image to obtain an optimal weight vector of the convolutional neural network classification training model. In the test stage, prediction is carried out by utilizing the optimal weight vector of the convolutional neural network classification training model, and a predicted semantic segmentation image corresponding to the road scene image to be subjected to semantic segmentation is obtained. The method has the advantages of low calculation complexity, high segmentation efficiency, high segmentation precision and good robustness.
Owner:ZHEJIANG UNIVERSITY OF SCIENCE AND TECHNOLOGY

Model training method and device, frame image generation method and device, frame insertion method and device, equipment and medium

The invention provides a model training method and device, a frame image generation method and device, a frame insertion method and device, equipment and a medium and relates to the technical field ofmodel training. The method is applied to a neural network model and adopts a first feature extraction module to extract local features of front and back frame images of a sample; a second feature extraction module is adopted to extract non-local features of front and back frame images of the sample; a frame synthesis module is adopted to generate a sample intermediate frame image according to thesynthesis features of the local features and the non-local features; and the neural network model is trained according to the sample intermediate frame image and the corresponding label intermediateframe image to obtain a trained neural network model. According to the neural network model obtained by training based on the mode, the receptive field is expanded, the learning ability of large changes in front and back frame images is enhanced, and when the front and back frame images are processed based on the trained neural network model, the generated middle frame image is more accurate.
Owner:NETEASE (HANGZHOU) NETWORK CO LTD

A vehicle image semantic segmentation system based on bilateral segmentation network

The invention discloses a vehicle-mounted image semantic segmentation system based on a bilateral segmentation network. The system comprises a data storage module, which is used for storing a vehicle-mounted image training set and a vehicle-mounted image to be tested; a bilateral segmentation network consisting of a spatial channel and a context channel, wherein the spatial channel is used to extract the spatial information of the vehicle image, and the context channel is used to extract the context semantic information of the vehicle image; a training module for training a bilateral segmentation network using a vehicle-mounted image training set; a semantic segmentation module used for predicting the vehicle image to be tested by using the trained bilateral segmentation network, and obtaining the class to which each pixel in the vehicle image to be tested belongs. The invention discloses a bilateral segmentation network comprising a spatial channel and a context channel, wherein the spatial channel is used for extracting spatial information of an image while retaining enough spatial information, and the context channel is used for extracting context semantic information of the image while ensuring a large enough receptive field.
Owner:HUAZHONG UNIV OF SCI & TECH

Face attribute editing method based on generative adversarial network and information processing terminal

The invention belongs to the technical field of human face attribute editing, discloses a human face attribute editing method based on a generative adversarial network and an information processing terminal, and constructs a human face attribute editing model combining the generative adversarial network and an auto-encoder. Wherein the auto-encoder is used as a generator, and the input of the model is pictures and attributes; optimizing the GAN loss by using the WGAN-GP to realize a task of editing a plurality of attributes by using a single generator; enabling the generated image to correctlyhave expected attributes by using an attribute classifier; a multi-scale discriminator is adopted to guide a generator to generate details, and detail information is captured on an original image; and combining the reconstruction loss, the attribute classification loss and the multi-scale GAN loss for face attribute editing. Experiments on a CelebA data set show that a high-quality face image isgenerated on the basis that expected attributes are correctly owned, and the method has good performance in the aspects of single-attribute face editing, multi-attribute face editing and attribute strength control.
Owner:CHINA UNIV OF PETROLEUM (EAST CHINA) +1
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products