Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

44 results about "Vision processing unit" patented technology

A vision processing unit (VPU) is (as of 2018) an emerging class of microprocessor; it is a specific type of AI accelerator, designed to accelerate machine vision tasks.

Rotor unmanned aircraft independent take-off and landing system based on three-layer triangle multi-color landing ground

Disclosed is a rotor unmanned aircraft independent take-off and landing system based on three-layer triangle multi-color landing ground. The system comprises a small rotor unmanned aircraft (SRUA), an onboard sensor, a data processing unit, a flight control system, an onboard camera, landing ground, a wireless image transmission module, a wireless data transmission module and a ground monitor station, wherein the onboard sensor comprises an inertial measurement unit, a global positioning system (GPS) receiver, a barometer, an ultrasound device and the like, the data processing unit is used for integrating sensor data, the flight control system finishes route planning to achieve high accuracy control over the SRUA, the onboard camera is used for collecting images of the landing ground, the landing ground is a specially designed landing point for an unmanned aircraft, the wireless image transmission module can transmit the images to a ground station, the wireless data transmission module can achieve communication of data and instructions between the unmanned aircraft and the ground station, and the ground station is composed of a visual processing unit and a display terminal. According to the rotor unmanned aircraft independent take-off and landing system based on the three-layer triangle multi-color landing ground, the reliability of SRUA navigation messages is guaranteed, the control accuracy of the SRUA is increased, costs are low, the application is convenient, and important engineering values can be achieved.
Owner:BEIHANG UNIV

Vision testing device with enhanced image clarity

The vision testing device with an enhanced image clarity for determining good or bad of a testing object by photographing a testing object assembled or mounted during the component assembly process and comparing the photographed image with a previously inputted target image, comprising: a stage part for fixing or transferring the testing object to a testing location; a lighting part for providing lighting to the testing object located on an upper portion of the stage part; a first camera part for obtaining a 2-dimensional image of the testing object located in a center of the lighting part; a plurality of second camera parts placed on a side section of the first camera part; a plurality of grid pattern irradiating parts placed between cameras of the second camera parts; a vision processing unit for reading the image photographed by the first camera part and the second camera parts and determining good or bad of the testing object; a control unit for controlling the stage part, the grid pattern irradiating parts, and the first and second camera parts; and a light diffusion part. The present invention enhances the uniformity of light being irradiated on the surface of the testing object. In addition, the present invention enables a more clear shooting of an image by removing a half mirror placed on the front of a camera part in the center, and enables convenient maintenance by miniaturizing the size of the device and dispersing the configuration.
Owner:MIRTEC

Vision testing device with enhanced image clarity

The vision testing device of the present invention is a vision testing device with enhanced image clarity for discriminating a satisfactory or a non-satisfactory testing object by shooting a testing object assembled or mounted during a component assembly process, and comparing the image taken with a previously inputted target image, and comprises: a stage part for fixing or transferring the testing object to the testing location; a lighting part located on the upper portion of the stage part, for providing lighting to the testing object; a first camera part located in the center of the lighting part, for obtaining a 2-dimensional image of the testing object; a second camera part placed in plurality on the side of the first camera part; a grid pattern irradiating part placed between the camera of the second camera part and a camera; a vision processing unit for reading the image taken in the first camera part and the second camera part, and discriminating whether the testing object is satisfactory or non-satisfactory; a control unit for controlling the stage part, the light irradiating part, and the first and the second camera parts; and a light diffusion part. The present invention enhances the uniformity of light being irradiated on the surface of the testing object. In addition, the present invention enables a more clear shooting of an image by removing a half mirror placed on the front of a camera part in the center, and enables convenient maintenance by miniaturizing the size of the device and dispersing the configuration.
Owner:MIRTEC

Scraper knife posture control system and method based on machine vision and land leveler

The invention discloses a scraper knife posture control system and method based on machine vision and a land leveler, and belongs to the technical field of land levelers. A scraper knife can be automatically adjusted to a set posture according to an operation instruction of a driver, and the control difficulty and labor intensity of the driver are reduced. The scraper knife posture control system comprises a plurality of cameras, a visual processing unit and a vehicle-mounted controller, wherein the cameras are installed at set positions of the land leveler and used for collecting real-time position information of a plurality of visual perception tracking points installed at the set positions of the land leveler; the visual processing unit is used for acquiring the current posture of a scraper knife according to the received real-time position information of each visual perception tracking point, and generating a target state of a scraper knife posture control mechanism in combination with the target posture, sent by a man-machine display, of the scraper knife; and the vehicle-mounted controller is used for controlling the scraper knife posture control mechanism through an electric proportional control valve according to the target state and the current state of the scraper knife posture control mechanism, so that the scraper knife posture control mechanism drives the scraper knife to a target posture.
Owner:XUZHOU XUGONG ROAD CONSTR MACHINERY +1

Water surface garbage autonomous collection method and system thereof

The invention discloses a water surface garbage autonomous collection method and a system thereof. The method comprises the steps of enabling an electronic module to be installed on a power ship for collecting water surface garbage, carrying out the autonomous recognition of a water surface object through a camera unit, and judging whether the garbage collection operation is carried out on the water surface object or not, using the camera unit, a visual processing unit, a sensing unit and a central control unit for autonomously avoiding obstacles and sweeping garbage on the water surface alongthe shore, establishing a three-dimensional model of a water area shoreline, and storing the information of the water area where the power ship works in the storage unit, monitoring the garbage collection amount in real time through the sensing unit, and when the garbage amount reaches a preset threshold value, enabling the power ship to automatically return. By the adoption of the water surfacegarbage autonomous collection method, autonomous recognition is conducted on the objects on the water surface through the camera unit, and garbage collection, autonomous obstacle avoidance, real-timemonitoring and automatic return are conducted on the objects on the water surface through the central control unit.
Owner:陕西欧卡电子智能科技有限公司

Mechanical arm visual control method and device

ActiveCN113733141AImprove gripAvoid the unstable phenomenon of adsorption and graspingGripping headsEngineeringVision processing unit
The invention discloses a mechanical arm visual control device. According to the technical scheme, the mechanical arm visual control device is characterized by comprising a mechanical arm body, partition plates, a mounting base, a shifting mechanism, a first grabbing mechanism and a second grabbing mechanism, and a mounting cylinder with a mounting cavity is assembled at the tail end of the mechanical arm body; the partition plates are arranged in the mounting cylinder and divides the mounting cavity of the mounting cylinder into a first cavity body, a second cavity body and a third cavity body, and the first cavity body, the second cavity body and the third cavity body are all arranged to be fan-shaped cavity bodies; the mounting base is arranged in the center of the bottom of the mounting cylinder and located at the joints of the partition plates, a visual processing unit is mounted on the mounting base, and the visual processing unit comprises a visual sensor, a visual processor and a distance measuring sensor; the shifting mechanism is arranged in the first cavity and used for shifting parts into the grabbing range of the mechanical arm; the first grabbing mechanism is arranged in the second cavity and used for adsorbing and grabbing the parts; and the second grabbing mechanism is arranged in the third cavity and used for clamping and grabbing the part. The mechanical arm visual control device has the beneficial effects of being stable in grabbing and high in working effect.
Owner:QUZHOU COLLEGE OF TECH

A Grassland Drought Monitoring System and Method Based on Computer Vision and Internet of Things

The invention discloses a grassland drought status monitoring system based on computer vision and the Internet of things and a grassland drought status monitoring method. The system comprises an image acquisition unit, an Internet of things control unit, a communication unit and a computer vision processing unit, wherein the image acquisition unit and the communication unit are respectively in communication connection with the Internet of things control unit, the communication unit communicates with the computer vision processing unit through the network, the Internet of things control unit controls the image acquisition unit for acquiring images of plants from different angles, the images are transmitted through the communication unit to the computer vision processing unit, and the images acquired from different angles are analyzed and processed by the computer vision processing unit to acquire a drought status of the plants. Through the system and the method, the image processing speed and accuracy can be improved, under the common action of the Internet of things module and the computer vision module, grassland degradation and the drought status can be warned in advance, and thereby remote drought monitoring can be realized.
Owner:CHINA AGRI UNIV

Machining track deformation compensation method and system

PendingCN112435350AEasy to edit laterFlexible deformation compensation capabilityImage enhancementImage analysisComputer graphics (images)Engineering
The invention provides a machining track deformation compensation method and system, and the method comprises the following steps: obtaining an initial track; acquiring an image; performing preprocessing operations such as splicing and calibration on the acquired images; setting a plurality of anchoring point measuring boxes, and calculating in a corresponding range to obtain anchoring points; generating track point measurement boxes at the initial track point positions, and setting track point coordinate calculation modes corresponding to the track point measurement boxes; obtaining an anchoring point; updating the position of the corresponding anchored track point measurement box according to the obtained coordinates of the anchoring point; calculating coordinates in the range of the track point measuring box in a set mode and updated to the corresponding track points, and obtaining a correction track; directly executing track correction; completing the steps through an initial trackacquisition unit, an image acquisition unit, a visual processing unit and a track execution unit of the machining track deformation compensation system, and deviation rectification or positioning ofa large complex track can be achieved in a simple mode.
Owner:深圳群宾精密工业有限公司

An autonomous take-off and landing system for rotary-wing unmanned aerial vehicles based on a three-layer character-shaped multi-color landing pad

Disclosed is a rotor unmanned aircraft independent take-off and landing system based on three-layer triangle multi-color landing ground. The system comprises a small rotor unmanned aircraft (SRUA), an onboard sensor, a data processing unit, a flight control system, an onboard camera, landing ground, a wireless image transmission module, a wireless data transmission module and a ground monitor station, wherein the onboard sensor comprises an inertial measurement unit, a global positioning system (GPS) receiver, a barometer, an ultrasound device and the like, the data processing unit is used for integrating sensor data, the flight control system finishes route planning to achieve high accuracy control over the SRUA, the onboard camera is used for collecting images of the landing ground, the landing ground is a specially designed landing point for an unmanned aircraft, the wireless image transmission module can transmit the images to a ground station, the wireless data transmission module can achieve communication of data and instructions between the unmanned aircraft and the ground station, and the ground station is composed of a visual processing unit and a display terminal. According to the rotor unmanned aircraft independent take-off and landing system based on the three-layer triangle multi-color landing ground, the reliability of SRUA navigation messages is guaranteed, the control accuracy of the SRUA is increased, costs are low, the application is convenient, and important engineering values can be achieved.
Owner:BEIHANG UNIV

Visual carrier landing guiding method and system for unmanned rotorcraft

The invention belongs to the technical field of aerospace, and particularly relates to a rotor unmanned aerial vehicle visual carrier landing guiding method and system. Comprising an airborne intelligent terminal and a ground auxiliary identifier. The airborne intelligent terminal comprises a visual processing unit, a visible light camera and an infrared camera, and the visible light camera or the infrared camera is used for obtaining ground auxiliary identification image information in real time and sending the ground auxiliary identification image information to the visual processing unit; the visual processing unit compares the received image information with a target image library in the visual processing unit, obtains the speed quantity required by the landing of the unmanned aerial vehicle corresponding to the current image acquisition moment, and feeds back the speed quantity to the flight control system; the flight control system judges whether the current speed of the unmanned aerial vehicle meets the speed quantity required by landing of the unmanned aerial vehicle; according to the invention, based on visual servo control of the image, autonomous landing of the unmanned aerial vehicle under supervision control is realized.
Owner:CHINA HELICOPTER RES & DEV INST
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products