Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

3438 results about "Pyramid" patented technology

A pyramid (from Greek: πυραμίς pyramís) is a structure whose outer surfaces are triangular and converge to a single step at the top, making the shape roughly a pyramid in the geometric sense. The base of a pyramid can be trilateral, quadrilateral, or of any polygon shape. As such, a pyramid has at least three outer triangular surfaces (at least four faces including the base). The square pyramid, with a square base and four triangular outer surfaces, is a common version.

Small target detection method based on feature fusion and depth learning

InactiveCN109344821AScalingRich information featuresCharacter and pattern recognitionNetwork modelFeature fusion
The invention discloses a small target detection method based on feature fusion and depth learning, which solves the problems of poor detection accuracy and real-time performance for small targets. The implementation scheme is as follows: extracting high-resolution feature map through deeper and better network model of ResNet 101; extracting Five successively reduced low resolution feature maps from the auxiliary convolution layer to expand the scale of feature maps. Obtaining The multi-scale feature map by the feature pyramid network. In the structure of feature pyramid network, adopting deconvolution to fuse the feature map information of high-level semantic layer and the feature map information of shallow layer; performing Target prediction using feature maps with different scales and fusion characteristics; adopting A non-maximum value to suppress the scores of multiple predicted borders and categories, so as to obtain the border position and category information of the final target. The invention has the advantages of ensuring high precision of small target detection under the requirement of ensuring real-time detection, can quickly and accurately detect small targets in images, and can be used for real-time detection of targets in aerial photographs of unmanned aerial vehicles.
Owner:XIDIAN UNIV

Image semantic division method based on depth full convolution network and condition random field

The invention provides an image semantic division method based on a depth full convolution network and a condition random field. The image semantic division method comprises the following steps: establishing a depth full convolution semantic division network model; carrying out structured prediction based on a pixel label of a full connection condition random field, and carrying out model training, parameter learning and image semantic division. According to the image semantic division method provided by the invention, expansion convolution and a spatial pyramid pooling module are introduced into the depth full convolution network, and a label predication pattern output by the depth full convolution network is further revised by utilizing the condition random field; the expansion convolution is used for enlarging a receptive field and ensures that the resolution ratio of a feature pattern is not changed; the spatial pyramid pooling module is used for extracting contextual features of different scale regions from a convolution local feature pattern, and a mutual relation between different objects and connection between the objects and features of regions with different scales are provided for the label predication; the full connection condition random field is used for further optimizing the pixel label according to feature similarity of pixel strength and positions, so that a semantic division pattern with a high resolution ratio, an accurate boundary and good space continuity is generated.
Owner:CHONGQING UNIV OF TECH

Gesture recognition method based on 3D-CNN and convolutional LSTM

The invention discloses a gesture recognition method based on 3D-CNN and convolution LSTM. The method comprises the steps that the length of a video input into 3D-CNN is normalized through a time jitter policy; the normalized video is used as input to be fed to 3D-CNN to study the short-term temporal-spatial features of a gesture; based on the short-term temporal-spatial features extracted by 3D-CNN, the long-term temporal-spatial features of the gesture are studied through a two-layer convolutional LSTM network to eliminate the influence of complex backgrounds on gesture recognition; the dimension of the extracted long-term temporal-spatial features are reduced through a spatial pyramid pooling layer (SPP layer), and at the same time the extracted multi-scale features are fed into the full-connection layer of the network; and finally, after a latter multi-modal fusion method, forecast results without the network are averaged and fused to acquire a final forecast score. According to the invention, by learning the temporal-spatial features of the gesture simultaneously, the short-term temporal-spatial features and the long-term temporal-spatial features are combined through different networks; the network is trained through a batch normalization method; and the efficiency and accuracy of gesture recognition are improved.
Owner:BEIJING UNION UNIVERSITY

PTAM improvement method based on ground characteristics of intelligent robot

The invention discloses a PTAM improvement method based on ground characteristics of an intelligent robot. The PTAM improvement method based on ground characteristics of the intelligent robot comprises the steps that firstly, parameter correction is completed, wherein parameter correction includes parameter definition and camera correction; secondly, current environment texture information is obtained by means of a camera, a four-layer Gausses image pyramid is constructed, the characteristic information in a current image is extracted by means of the FAST corner detection algorithm, data relevance between corner characteristics is established, and then a pose estimation model is obtained; two key frames are obtained so as to erect the camera on the mobile robot at the initial map drawing stage; the mobile robot begins to move in the initializing process, corner information in the current scene is captured through the camera and association is established at the same time; after a three-dimensional sparse map is initialized, the key frames are updated, the sub-pixel precision mapping relation between characteristic points is established by means of an extreme line searching and block matching method, and accurate re-positioning of the camera is achieved based on the pose estimation model; finally, matched points are projected in the space, so that a three-dimensional map for the current overall environment is established.
Owner:BEIJING UNIV OF TECH

A pedestrian and vehicle detection method and system based on improved YOLOv3

The invention discloses a pedestrian and vehicle detection method and system based on improved YOLOv3. According to the method, an improved YOLOv3 network based on Darknet-33 is adopted as a main network to extract features; the cross-layer fusion and reuse of multi-scale features in the backbone network are carried out by adopting a transmittable feature map scale reduction method; and then a feature pyramid network is constructed by adopting a scale amplification method. In the training stage, a K-means clustering method is used for clustering the training set, and the cross-to-parallel ratio of a prediction frame to a real frame is used as a similarity standard to select a priori frame; and then the BBox regression and the multi-label classification are performed according to the loss function. And in the detection stage, for all the detection frames, a non-maximum suppression method is adopted to remove redundant detection frames according to confidence scores and IOU values, and an optimal target object is predicted. According to the method, a feature extraction network Darknet-33 of feature map scale reduction fusion is adopted, a feature pyramid is constructed through feature map scale amplification migration fusion, and a priori frame is selected through clustering, so that the speed and precision of the pedestrian and vehicle detection can be improved.
Owner:NANJING UNIV OF POSTS & TELECOMM

Multiple video cameras synchronous quick calibration method in three-dimensional scanning system

A synchronous quick calibration method of a plurality of video cameras in a three-dimensional scanning system, which includes: (1) setting a regular truncated rectangular pyramid calibration object, setting eight calibration balls at the vertexes of the truncated rectangular pyramid, and respectively setting two reference calibration balls at the upper and lower planes; (2) using the video cameras to pick-up the calibration object, adopting the two-threshold segmentation method to respectively obtain the corresponding circles of the upper and lower planes, extracting centers of the circles, obtaining three groups of corresponding relationships between circle center points in the image and the centres of calibration ball in the space, solving the homography matrix to obtain the internal parameter matrix and external parameter matrix and obtaining the distortion coefficient, taking the solved video camera parameter as the initial values, and then using a non-linear optimization method to obtain the optimum solution of a single video camera parameter; (3) obtaining in sequence the external parameter matrix between a plurality of video cameras and a certain video camera in the space, using the polar curve geometric constraint relationship of the binocular stereo vision to establish an optimizing object function, and then adopting a non-linear optimization method to solve to get the optimum solution of the external parameter matrix between two video cameras.
Owner:NANTONG TONGYANG MECHANICAL & ELECTRICAL MFR +1

VR (Virtual Reality) panoramic video layout method and device and VR panoramic video presentation method and system capable of saving bandwidth

The invention provides a VR (Virtual Reality) panoramic video layout method, device and system. The method comprises the following steps: projecting a VR panoramic spherical video onto a pyramid surface which takes a sphere as an inscribed sphere; expanding the pyramid surface into a planar graph to obtain a video plane of pyramid projection; and turning the video plane to a regular shape through deformation to obtain a regular video plane. The VR panoramic video layout system comprises a VR panoramic video layout processing device, a VR panoramic video streaming media server and a VR panoramic video player, wherein the VR panoramic video player is used for playing video data from the VR panoramic video streaming media server according to a viewing angle of a user. Through adoption of the VR panoramic video layout method, device and system provided by the invention, a pyramid-based projection way is provided. In the pyramid-based projection way, the surface area of a projected video is reduced by 80 percent compared with an equirectangular projection way; the bandwidth of VR panoramic video transmission is reduced by 80 percent compared with an existing VR video technology; and a panoramic video can be played according to the viewing angle of the user.
Owner:北京金字塔虚拟现实科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products