Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

65results about How to "Improve I/O efficiency" patented technology

Data page caching method for file system of solid-state hard disc

The invention discloses a data page caching method for a file system of a solid-state hard disc, which comprises the following implementation steps of: (1) establishing a buffer link list used for caching data pages in a high-speed cache; (2) caching the data pages read in the solid-state hard disc in the buffer link list for access, classifying the data pages in the buffer link list into cold clean pages, hot clean pages, cold dirty pages and hot dirty pages in real time according to the access states and write access states of the data pages; (3) firstly searching a data page as a page to be replaced in the buffer link list according to the priority of the cold clean pages, the hot clean pages, the cold dirty pages and the hot dirty pages, and replacing the page to be replaced with a new data page read from the solid-state hard disc when a free space does not exist in the buffer link list. In the invention, the characteristics of the solid-state hard disc can be sufficiently utilized, the performance bottlenecks of the external storage can be effectively relieved, and the storage processing performance of the system can be improved; moreover, the data page caching method has the advantages of good I/O (Input/Output) performance, low replacement cost for cached pages, low expense and high hit rate.
Owner:NAT UNIV OF DEFENSE TECH

Method with verification for intelligently crawling network information in distributed way

The invention puts forward a method with verification for intelligently crawling network information in a distributed way. The method comprises the following steps that: when a judgement result shows that the target page data of a website can be obtained after login verification is carried out, obtaining corresponding login information from a database, carrying out automatic login through a browser, and submitting verification information; starting a timed task, using cookie to access the webpage of the timed task, and carrying out keep-alive processing; starting a network package capture detector, accessing a corresponding target page according to business requirements, carrying out HTTP (Hyper Text Transport Protocol) message analysis, carrying out customization on a crawler script, and determining a task crawling data size; and emitting a broadcast by a main node, notifying a corresponding task node, distributing the crawler script, starting the task node, applying for a task from a main node task queue, carrying out data crawling according to the applied task, and storing the crawled target data into the queue so as to store the crawled target data into the database in batches. By use of the method, a protected page can be automatically logged in and accessed, and a quick and expandable distributed webpage crawler integrated framework capable of mining the script is automatically generated.
Owner:北京京拍档科技股份有限公司

Hash table concurrent access performance optimization method under multi-core environment

The invention discloses a hash table concurrent access performance optimization method under a multi-core environment. The method comprises the steps that aiming at the concurrent connection treatment of high concurrent access of a hash table, a half-synchronous half-asynchronous network connection treatment mechanism is adopted; aiming at the concurrent data treatment of high concurrent access of the hash table, the hash table is divided into multiple independent sub hash tables according to the number of CPU cores, working threads correspond to the independent sub hash tables in a one-to-one mode, and each working thread is only in charge of the data of the corresponding sub hash table; a main thread selects the corresponding working threads to process by adopting a consistent hash strategy according to the Key of each datum; each sub hash table maintains one LRU queue, when a memory space is insufficient, cold data are deleted, and new elements are inserted. The method improves the concurrent connection treatment capacity and the concurrent access capacity of the hash table, and the synchronization overhead problem and the cache consistency overhead problem are solved when the shared hash table is visited by multiple threads.
Owner:HUAZHONG UNIV OF SCI & TECH

Small computer system interface (SCSI) fault-tolerant optimization method and device based on hadoop distributed file system (HDFS)

The invention provides a small computer system interface (SCSI) fault-tolerant optimization method and device based on a hadoop distributed file system (HDFS). The method comprises the following steps: a request which is sent by a client end and is about a first data block duplicate is received; the request is sent to an SCSI bottom layer through an SCSI middle layer; the SCSI bottom layer processes the request and returns a processed result to the SCSI middle layer; the SCSI middle layer executes calling-back according to the processed result returned by the SCSI bottom layer so as to judge whether the request is successful or not, and the SCSI bottom layer needs to be retried or is failed; and retrying times are set or retrying is given up when the fact that the SCSI bottom layer needs to be retried is judged. According to the SCSI fault-tolerant optimization method and device based on the HDFS, when the request of the client end is judged to be the fact that the SCSI bottom layer needs to be retried according to the result returned by the SCSI bottom layer, the retrying times are set or giving-up retrying is set, and the error processing method which does not influence the actual request of the client end is optimized. Thus, IO efficiency is improved, and hardware failure rate is lowered.
Owner:BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD

Spatial data storage organization method and system in consideration of load balancing and disc efficiency

The invention discloses a spatial data storage organization method and system in consideration of load balancing and disc efficiency. The method comprises the steps of counting an average access concurrency degree according to times of spatial data files accessed and requested by a user; distributively storing the spatial data files to a distributed geographical information system server according to the load balancing requirement of a distributed geographical information system and the average access concurrency degree; calculating the average access continuity of the spatial data files according to the average access concurrency degree and an average access spacing distance of the spatial data files; and carrying out consecutive storage organization on the spatial data files in the same server according to the average access continuity of the spatial data files. According to the spatial data storage organization method and the spatial data storage organization system in consideration of load balancing and disc efficiency, effective batch reading of continuous accessed space data is realized and the storage efficiency of the geographical information system is guaranteed while the load balancing requirement is satisfied. The project is completed by funding of Natural Science Foundation of China (Fund No: 41671382, 41271398).
Owner:WUHAN UNIV

A multi-task external memory schema graph processing method based on I/O scheduling

The invention discloses a multi-task external memory mode diagram processing method based on I / O scheduling, includes streaming partitioning graph data to obtain graph partition, evenly placing graphpartition in multiple external storage devices, selecting target external storage devices from multiple external storage devices based on I / O scheduling, and taking graph partition in the target external storage device that has not been accessed by graph processing task as designated partition; Judging whether the synchronization field of the designated partition is not mapped into the memory according to the synchronization field of the designated partition, if so, mapping the designated partition from the external storage device into the memory, and updating the synchronization field of thedesignated partition; Otherwise, the graph partition data is accessed directly through the address information mapped to memory by the specified partition. Through I / O scheduling, the invention selects the external storage device with the least number of tasks to access, thereby controlling the sequence of accessing the data of the external storage diagram partition and balancing the I / O pressure.By setting the synchronization field to realize the data sharing of graph partition, the repeated loading of the same graph partition is reduced, so as to reduce the total I / O bandwidth and improve I / O efficiency.
Owner:HUAZHONG UNIV OF SCI & TECH

Method for controlling parallelism degree of program capable of sensing band width of storage device

The invention discloses a method for controlling the parallelism degree of a program capable of sensing the band width of a storage device. By means of the method, the parallelism degree of the application program is dynamically adjusted according to the comprehensive performance and real-time loading conditions of the storage device of an operating platform of the application program in a self-adaptive mode, namely the number of courses or threads, so that the application program keeps and obtains the parallelism degree with the optimal I/O efficiency. According to the method, the information of the I/O efficiency of the application program is monitored and recorded in real time and is used as feedback information to control and adjust the parallelism degree of the application program. The parallelism degree of the program is probed step by step and increased gradually until the parallelism degree reaches the inflection point of the actual I/O efficiency, and then fine adjustment is carried out to obtain the optimal parallelism degree for different application program platforms; further according to the real-time loading conditions of the different application program platforms, periodic, dynamic and self-adaptive operation combining active adjustment and passive adjustment is carried out to obtain the optimal parallelism degree of the application program.
Owner:凯习(北京)信息科技有限公司

Data updating system and method based on resource mutual exclusion scheduling model

The invention provides a data updating system and method based on a resource mutual exclusion scheduling model, which can be applied to the field of data processing and finance, and the system comprises a resource cache module which obtains market information push data, and stores the market information data into a preset cache queue according to the receiving time of the market information push data and the data category of each piece of market information data in the market information push data, a coordination thread module which obtains the resource type of each product, and constructs characteristic data corresponding to each product according to the resource type, acquires market data in the cache queue, and constructs a product set according to resource types and characteristic data of the market data and performs mutual exclusion scheduling according to resource types of the product set, distributes market information data to different execution queues, and calculates resource updating data of each product through working threads corresponding to the execution queues, and a synthesis calculation module which obtains resource update data calculated by each working thread, and updates corresponding resources in each product through the resource update data.
Owner:INDUSTRIAL AND COMMERCIAL BANK OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products