Eureka-AI is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Eureka AI

2625 results about "Edge detection" patented technology

Edge detection includes a variety of mathematical methods that aim at identifying points in a digital image at which the image brightness changes sharply or, more formally, has discontinuities. The points at which image brightness changes sharply are typically organized into a set of curved line segments termed edges. The same problem of finding discontinuities in one-dimensional signals is known as step detection and the problem of finding signal discontinuities over time is known as change detection. Edge detection is a fundamental tool in image processing, machine vision and computer vision, particularly in the areas of feature detection and feature extraction.

Systems and methods for automated screening and prognosis of cancer from whole-slide biopsy images

InactiveUS20140233826A1Accurate and unambiguous measureReduce dependenceImage enhancementMedical data miningFeature setProstate cancer
The invention provides systems and methods for detection, grading, scoring and tele-screening of cancerous lesions. A complete scheme for automated quantitative analysis and assessment of human and animal tissue images of several types of cancers is presented. Various aspects of the invention are directed to the detection, grading, prediction and staging of prostate cancer on serial sections/slides of prostate core images, or biopsy images. Accordingly, the invention includes a variety of sub-systems, which could be used separately or in conjunction to automatically grade cancerous regions. Each system utilizes a different approach with a different feature set. For instance, in the quantitative analysis, textural-based and morphology-based features may be extracted at image- and (or) object-levels from regions of interest. Additionally, the invention provides sub-systems and methods for accurate detection and mapping of disease in whole slide digitized images by extracting new features through integration of one or more of the above-mentioned classification systems. The invention also addresses the modeling, qualitative analysis and assessment of 3-D histopathology images which assist pathologists in visualization, evaluation and diagnosis of diseased tissue. Moreover, the invention includes systems and methods for the development of a tele-screening system in which the proposed computer-aided diagnosis (CAD) systems. In some embodiments, novel methods for image analysis (including edge detection, color mapping characterization and others) are provided for use prior to feature extraction in the proposed CAD systems.

Deinterlacing of video sources via image feature edge detection

ActiveUS7023487B1Reduce artifactsPreserves maximum amount of vertical detailImage enhancementTelevision system detailsInterlaced videoProgressive scan
An interlaced to progressive scan video converter which identifies object edges and directions, and calculates new pixel values based on the edge information. Source image data from a single video field is analyzed to detect object edges and the orientation of those edges. A 2-dimensional array of image elements surrounding each pixel location in the field is high-pass filtered along a number of different rotational vectors, and a null or minimum in the set of filtered data indicates a candidate object edge as well as the direction of that edge. A 2-dimensional array of edge candidates surrounding each pixel location is characterized to invalidate false edges by determining the number of similar and dissimilar edge orientations in the array, and then disqualifying locations which have too many dissimilar or too few similar surrounding edge candidates. The surviving edge candidates are then passed through multiple low-pass and smoothing filters to remove edge detection irregularities and spurious detections, yielding a final edge detection value for each source image pixel location. For pixel locations with a valid edge detection, new pixel data for the progressive output image is calculated by interpolating from source image pixels which are located along the detected edge orientation.

Regional depth edge detection and binocular stereo matching-based three-dimensional reconstruction method

The invention discloses a regional depth edge detection and binocular stereo matching-based three-dimensional reconstruction method, which is implemented by the following steps: (1) shooting a calibration plate image with a mark point at two proper angles by using two black and white cameras; (2) keeping the shooting angles constant and shooting two images of a shooting target object at the same time by using the same camera; (3) performing the epipolar line rectification of the two images of the target objects according to the nominal data of the camera; (4) searching the neighbor regions of each pixel of the two rectified images for a closed region depth edge and building a supporting window; (5) in the built window, computing a normalized cross-correlation coefficient of supported pixels and acquiring the matching price of a central pixel; (6) acquiring a parallax by using a confidence transmission optimization method having an acceleration updating system; (7) estimating an accurate parallax by a subpixel; and (8) computing the three-dimensional coordinates of an actual object point according to the matching relationship between the nominal data of the camera and the pixel and consequently reconstructing the three-dimensional point cloud of the object and reducing the three-dimensional information of a target.
Owner:江苏省华强纺织有限公司 +1

Autonomously identifying and capturing method of non-cooperative target of space robot

The invention relates to an autonomously identifying and capturing method of a non-cooperative target of a space robot, comprising the main steps of (1) pose measurement based on stereoscopic vision, (2) autonomous path planning of the target capture of the space robot and (3) coordinative control of a space robot system, and the like. The pose measurement based on the stereoscopic vision is realized by processing images of a left camera and a right camera in real time, and computing the pose of a non-cooperative target star relative to a base and a tail end, wherein the processing comprises smoothing filtering, edge detection, linear extraction, and the like. The autonomous path planning of the target capture of the space robot comprises is realized by planning the motion tracks of joints in real time according to the pose measurement results. The coordinative control of the space robot system is realized by coordinately controlling mechanical arms and the base to realize the optimal control property of the whole system. In the autonomously identifying and capturing method, a self part of a spacecraft is directly used as an identifying and capturing object without installing a marker or a comer reflector on the target star or knowing the geometric dimension of the object, and the planned path can effectively avoid the singular point of dynamics and kinematics.

Vision guiding AGV (Automatic Guided Vehicle) system and method of embedded system

The invention discloses a vision guiding AGV (Automatic Guided Vehicle) system and a method of an embedded system. Two cameras fixed on a trolley are used for acquiring guiding path information in real time, wherein the cameras and the ground form a certain angle to be inclined forwards and is used for acquiring prospect images; and the cameras are arranged at the middle part and the front part of the interior of the trolley, are vertical to the ground and are used for secondary exact location. The vision guiding AGV method comprises the following step of: embedding an anti-metal radio frequency mark on the ground surface of a key position, wherein a vehicle-mounted radio frequency card reader obtains information in the mark when the trolley passes through the upper part of the mark. When a laser scanner scans a front obstacle in real time, obstacle avoidance detection and obstacle avoidance are realized. A control box of the embedded system disclosed by the invention is as an inner kernel for image collection, image treatment and policy control, the acquired image is subjected to Gauss high-pass filtering, edge detection and two-step Hough transformation so that position deviation and angle deviation of the trolley relative to a current path can be calculated, and two-dimensional deviation value of the AGV currently corresponding to a location reference point is fed back at a station point.

Extraction method for characters in form document image

The invention relates to the field of image processing and computer vision technologies, in particular to an extraction method for characters in a form document image. The extraction method includes a first step of extracting line segments in the image through edge detection and Hough transformation algorithm, a second step of estimating an inclined angle of the whole image according to direction distribution of the line segments and carrying out inclination correction on the image, a third step of connecting the line segments in the horizontal direction and the perpendicular direction, and locating table cells of a form, a fourth step of carrying out image binaryzation and segmenting a full line of characters in the table cells through a maximum between-cluster variance method, and extracting the characters in the table cells through a window sliding method, and a fifth step of carrying out restoration on deletion of strokes of the characters according to statistics features of frame lines of the table cells. The extraction method is good in flexibility and capable of effectively solving the problems of adhesion between the characters and overlap between the characters and form lines, and greatly reduces the influence of the adhesion and overlap on optical character recognition (OCR).
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products