Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

117 results about "Large matrices" patented technology

High speed and efficient matrix multiplication hardware module

A matrix multiplication module and matrix multiplication method are provided that use a variable number of multiplier-accumulator units based on the amount of data elements of the matrices are available or needed for processing at a particular point or stage in the computation process. As more data elements become available or are needed, more multiplier-accumulator units are used to perform the necessary multiplication and addition operations. To multiply an N×M matrix by an M×N matrix, the total (maximum) number of used MAC units is “2*N−1”. The number of MAC units used starts with one (1) and increases by two at each computation stage, that is, at the beginning of reading of data elements for each new row of the first matrix. The sequence of the number of MAC units is {1, 3, 5, . . . , 2*N−1} for computation stages each of which corresponds to reading of data elements for each new row of the left hand matrix, also called the first matrix. For the multiplication of two 8×8 matrices, the performance is 16 floating point operations per clock cycle. For an FPGA running at 100 MHz, the performance is 1.6 Giga floating point operations per second. The performance increases with the increase of the clock frequency and the use of larger matrices when FPGA resources permit. Very large matrices are partitioned into smaller blocks to fit in the FPGA resources. Results from the multiplication of sub-matrices are combined to form the final result of the large matrices.
Owner:HARRIS CORP

High-precision matrix-vector multiplication on a charge-mode array with embedded dynamic memory and stochastic method thereof

Analog computational arrays for matrix-vector multiplication offer very large integration density and throughput as, for instance, needed for real-time signal processing in video. Despite the success of adaptive algorithms and architectures in reducing the effect of analog component mismatch and noise on system performance, the precision and repeatability of analog VLSI computation under process and environmental variations is inadequate for some applications. Digital implementation offers absolute precision limited only by wordlength, but at the cost of significantly larger silicon area and power dissipation compared with dedicated, fine-grain parallel analog implementation. The present invention comprises a hybrid analog and digital technology for fast and accurate computing of a product of a long vector (thousands of dimensions) with a large matrix (thousands of rows and columns). At the core of the externally digital architecture is a high-density, low-power analog array performing binary-binary partial matrix-vector multiplication. Digital multiplication of variable resolution is obtained with bit-serial inputs and bit-parallel storage of matrix elements, by combining quantized outputs from one or more rows of cells over time. Full digital resolution is maintained even with low-resolution analog-to-digital conversion, owing to random statistics in the analog summation of binary products. A random modulation scheme produces near-Bernoulli statistics even for highly correlated inputs. The approach has been validated by electronic prototypes achieving computational efficiency (number of computations per unit time using unit power) and integration density (number of computations per unit time on a unit chip area) each a factor of 100 to 10,000 higher than that of existing signal processors making the invention highly suitable for inexpensive micropower implementations of high-data-rate real-time signal processors.
Owner:GENOV ROMAN A +1

High Speed and Efficient Matrix Multiplication Hardware Module

A matrix multiplication module and matrix multiplication method are provided that use a variable number of multiplier-accumulator units based on the amount of data elements of the matrices are available or needed for processing at a particular point or stage in the computation process. As more data elements become available or are needed, more multiplier-accumulator units are used to perform the necessary multiplication and addition operations. To multiply an N×M matrix by an M×N matrix, the total (maximum) number of used MAC units is “2*N−1”. The number of MAC units used starts with one (1) and increases by two at each computation stage, that is, at the beginning of reading of data elements for each new row of the first matrix. The sequence of the number of MAC units is {1, 3, 5, . . . , 2*N−1} for computation stages each of which corresponds to reading of data elements for each new row of the left hand matrix, also called the first matrix. For the multiplication of two 8×8 matrices, the performance is 16 floating point operations per clock cycle. For an FPGA running at 100 MHz, the performance is 1.6 Giga floating point operations per second. The performance increases with the increase of the clock frequency and the use of larger matrices when FPGA resources permit. Very large matrices are partitioned into smaller blocks to fit in the FPGA resources. Results from the multiplication of sub-matrices are combined to form the final result of the large matrices.
Owner:HARRIS CORP

Underdetermined blind source separation (UBSS) method based on maximum matrix diagonal rate

The invention discloses an underdetermined blind source separation (UBSS) method based on a maximum matrix diagonal rate. The method comprises the following steps of: constructing inverse matrixes of C*M / N M*M-dimensional sub matrixes of a mixed matrix (wherein M and N are respectively the number of sensors and the number of source signals); multiplying the inverse matrixes by observation signal vectors to acquire initial estimation signal vectors; and sequentially calculating the covariance matrix, the solid matrix, the absolute value matrix and the diagonal rate of each initial estimation signal vector, selecting the initial estimation signal vector corresponding to the maximum diagonal rate as estimation of a source signal vector, and thus realizing underdetermined separation of sourcesignals. By the method, the requirement for source signal sparseness is reduced, aliasing of road source signals is allowed at each time frequency point at most, and the underdetermined separation problem of music signals and noise signals is solved. The requirement for the statistical property of the source signals is low, and the underdetermined separation problem of Gaussian signals and related signals is solved. In addition, by the method, processing of each time frequency point and each sub matrix can be executed in parallel, and hardware implementation is facilitated.
Owner:DALIAN UNIV OF TECH

Matrix convolution calculation method, interface, coprocessor and system based on RISC-V architecture

The invention discloses a set based on RISC-. According to the method and system complete mechanism of the instruction, the interface and the coprocessor for matrix convolution calculation of the V instruction set architecture, traditional matrix convolution calculation is efficiently achieved in a software and hardware combined mode, and RISC-is utilized. Extensibility of V instruction sets, a small number of instructions and a special convolution calculation unit (namely a coprocessor) are designed; the memory access times and the execution period of a matrix convolution calculation instruction are reduced, the complexity of application layer software calculation is reduced, the efficiency of large matrix convolution calculation is improved, the calculation speed of matrix convolution isincreased, flexible calling of upper-layer developers is facilitated, and the coding design is simplified. Meanwhile, RISC-is utilized. The processor designed by the V instruction set also has greatadvantages in power consumption, size and flexibility compared with ARM, X86 and other architectures, can adapt to different application scenes, and has a wide prospect in the field of artificial intelligence.
Owner:NANJING HUAJIE IMI TECH CO LTD

Front-face-compensation-operator-based multi-pose human face recognition method

The invention discloses a simple and effective principal component analysis (PCA) algorithm-based multi-pose human face recognition technique, which compensating a multi-pose human face by using a front face compensation operator and using the compensated human face for multi-pose human face recognition. The attached figure in the abstract of the description is the whole multi-pose human face recognition flow chart. In the invention, when the PCA algorithm is used for human face recognition, the front face compensation operator compensates for front face profile information, namely feature face information corresponding to the large feature value resolved by the PCA algorithm, lacking in a multi-pose human face to be recognized, and reduces part of multi-pose human face information interfering with the PCA algorithm. A multi-pose human face usually lacks a front face profile which is more important information for PCA algorithm. Compared with the prior art, the technique has the advantages that: calculating by using an average face and avoiding using the method for training by forming a large matrix with human faces from a human face library, the calculation amount is reduced; thehuman face normalization requirement is low; it is easy to choose a human face library; and fewer faces are required to be trained. In addition, the algorithm used by the technique is simple, a good recognition effect can be achieved by simple addition and subtraction operation, and the technique can be used for recognizing human faces in all poses. In the invention, the method for compensating the multi-pose human face by using the front face compensation operator provides a new way for improving multi-pose human face recognition rate.
Owner:CHONGQING UNIV

Three-phase decoupling load flow calculation method of power distribution network based on path matrix

The invention discloses a three-phase decoupling load flow calculation method of a power distribution network based on a path matrix. The method comprises the following steps: firstly, adopting a symmetrical component method to perform sequence component decoupling on a three-phase unbalanced power distribution network to obtain zero sequence, a power distribution sequence network with positive sequence and negative sequence, and adopting a loop-analysis method based on the path matrix to perform one-phase-sequence component load flow calculation to obtain load flows of the three-sequence networks; and secondly, transforming sequence network load flows in a phase component mode by an inverse transformation principle of the symmetrical component method to obtain three-phase load flows. By using the method, a three-phase unbalanced power distribution network system is decoupled into zero sequence, positive sequence and negative sequence networks, so that large matrix manipulation in the three-phase load flow calculation is avoided, the calculated amount is decreased, and the calculation efficiency is improved. The method has the advantages of clear calculation process, simple programming and fast calculation speed. Finally, a 6-busbar test example verifies the correctness and good convergence; and the method is good in generality and practical applicability.
Owner:HOHAI UNIV

Matrix transposition method in SAR imaging system based on DSP chip

InactiveCN103412284ASave storage spaceOvercome the disadvantage of taking a lot of time to moveRadio wave reradiation/reflectionRadarAlgorithm
The invention discloses a matrix transposition method in an SAR imaging system based on a DSP chip. The problems that radar real-time imaging storage space is small, and the requirement for real-time performance is high are mainly solved. The implementation process of the matrix transposition method comprises the steps that (1) an original matrix is transversely divided; (2) a rough classification matrix is longitudinally divided; (3) small unit matrixes are obtained; (4) storage space is created; (5) the small unit matrixes are classified; (6) transposition is carried out on the small diagonal unit matrixes; (7) transposition is carried out on the small non-diagonal unit matrixes; (8) the transpositioned matrixes are combined. The matrix transposition method aims at the transposition operation of a large-scale radar return original data matrix, the large matrix is roughly and finely divided into two classes of small unit matrixes, different transposition methods are applied to processing the matrixes, a large amount of storage space is saved, and operation efficiency is improved. The matrix transposition method is simple, easy to implement and suitable for the transposition operation of various radar real-time imaging systems and other systems.
Owner:XIDIAN UNIV

LDPC (Low Density Parity Check) encoding method based on FPGA (Field Programmable Gate Array) and used in CMMB (China Mobile Multimedia Broadcasting) exciter

The invention discloses an LDPC (Low Density Parity Check) encoding method based on FPGA (Field Programmable Gate Array) and used in CMMB (China Mobile Multimedia Broadcasting) exciter. The method comprises the following steps: firstly, processing a verification matrix H of an LDPC system code on an MATLAB (Matrix Laboratory) platform, so as to generate verification matrixes Hp and Hs which correspond to a verification bit column vector P and an information bit column vector S; carrying out LU (Logical Unit) decomposition to the verification matrixes so as to obtain a lower triangular matrix L and an upper triangular matrix U; and realizing the LDPC encoding on the FPGA platform, which mainly involves the storage of a large matrix, matrix multiplication, forward iteration and backward iteration. According to the encoding method, an encoding mode based on a LU decomposition verification matrix is adopted; the logic calculation that large matrixes are multiplied is avoided; and the problem of large requirement on FPGA internal storage resource caused by large data quantity storage is solved, thereby simplifying the logic calculation operation, saving the storage space, and being beneficial for the realization of the LDPC encoding of the CMMB system.
Owner:ALLWIN TELECOMM

Active and passive combined fixed-wing aircraft gust alleviation control method

The invention discloses an active and passive combined fixed-wing aircraft gust alleviation control method, and belongs to the technical field of aircraft control. Firstly, a two-dimensional coordinate system OZX is constructed, a wing comprises an inner side wing body and a full-moving wing tip, the wing is discretized in the OZX, and a vibration differential equation of the inner side wing bodyand the all-moving wing tip is listed; aerodynamic force acting on the inner side wing body is obtained through an aerodynamic force coefficient matrix, and discrete gust is calculated; the aerodynamic force and a model of the gust are substituted into the vibration differential equation, shear force merging is eliminated to obtain a kinetic equation in a complete machine large matrix form, and the kinetic equation is converted into an aircraft gust response dynamics equation; the acceleration and displacement of the wing mass center of a fixed-wing aircraft are selected as an objective function of the optimal control design, the deformation of the wing is worked out, and the efficiency of gust alleviation is verified. According to the method, under the same gust intensity and overload requirements, a required rudder deflection angle is reduced, the efficiency of the gust alleviation is improved, and a good weakening effect on the overload and deformation of the wing tip is achieved.
Owner:BEIHANG UNIV

Score prediction method for constructing local matrix based on graph random walk

The invention discloses a score prediction method for constructing a local matrix based on graph random walk, and belongs to the field of personalized recommendation. Depending on user-article scoringmatrix, a user-article bipartite graph is constructed, random walk is carried out on the bipartite graph, and A users with the maximum node convergence probability after walk and articles are selected to form A anchor points; and for each anchor point, a random walk algorithm with restart is adopted to obtain a correlation between each node and the anchor point so as to distribute each node intoa corresponding anchor point neighborhood. Each anchor point and the neighborhood thereof form a local matrix, and score prediction is carried out in each local matrix by using a matrix decompositionmethod. The prediction scores of the A local matrixes are averaged to obtain a final prediction result. According to the method, anchor points are selected and neighborhoods of the anchor points are constructed based on graph random walk, so that errors caused by a traditional distance calculation process are avoided; starting from nodes, the nodes are distributed to different anchor point neighborhoods, and complete coverage of a large matrix can be achieved.
Owner:HUAZHONG UNIV OF SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products