Eureka-AI is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Eureka AI

4265 results about "Source data" patented technology

Source data is raw data (sometimes called atomic data) that has not been processed for meaningful use to become Information.

Method and apparatus for data communication

A data acquisition and delivery system for performing data delivery tasks is disclosed. This system uses a computer running software to acquire source data from a selected data source, to process (e.g. filter, format convert) the data, if desired, and to deliver the resulting delivered data to a data target. The system is designed to access remote and/or local data sources and to deliver data to remote and/or local data targets. The data target might be an application program that delivers the data to a file or the data target may simply be a file, for example. To obtain the delivered data, the software performs processing of the source data as appropriate for the particular type of data being retrieved, for the particular data target and as specified by a user, for example. The system can communicate directly with a target application program, telling the target application to place the delivered data in a particular location in a particular file. The system provides an external interface to an external context. If the external context is a human, the external interface may be a graphical user interface, for example. If the external context is another software application, the external interface may be an OLE interface, for example. Using the external interface, the external context is able to vary a variety of parameters to define data delivery tasks as desired. The system uses a unique notation that includes a plurality of predefined parameters to define the data delivery tasks and to communicate them to the software.

Managing a codec engine for memory compression/decompression operations using a data movement engine

A system and method for managing a functional unit in a system using a data movement engine. An exemplary system may comprise a CPU coupled to a memory controller. The memory controller may include or couple to a data movement engine (DME). The memory controller may in turn couple to a system memory or other device which includes at least one functional unit. The DME may operate to transfer data to/from the system memory and/or the functional unit, as described herein. In one embodiment, the DME may also include multiple DME channels or multiple DME contexts. The DME may operate to direct the functional unit to perform operations on data in the system memory. For example, the DME may read source data from the system memory, the DME may then write the source data to the functional unit, the functional unit may operate on the data to produce modified data, the DME may then read the modified data from the functional unit, and the DME may then write the modified data to a destination in the system memory. Thus the DME may direct the functional unit to perform an operation on data in system memory using four data movement operations. The DME may also perform various other data movement operations in the computer system, e.g., data movement operations that are not involved with operation of the functional unit.

System and method for database conversion

A database conversion engine comprising a method and system to convert business information residing on one system to another system. A generic, extensible, scalable conversion engine may perform conversion of source data to target data as per mapping instructions/specifications, target schema specifications, and a source extract format specification without the need for code changes to the engine itself for subsequent conversions. A scheduler component may implement a scalable architecture capable of voluminous data crunching operations. Multi-level validation of the incoming source, data may also be provided by the system. A mechanism may provide data feeds to third-party systems as a part of business data conversion. An English-like, XML-based (extensible markup language), user-friendly, extensible data markup language may be further provided to specify the mapping instructions directly or via a GUI (graphical user interface). The system and method employs a business-centric approach to data conversion that determines the basic business object that is the building block of a given conversion. This approach facilitates identification of basic minimum required data for conversion leading to efficiencies in volume of data, performance, validations, reusability, and conversion turnover time.

Data processing apparatus and method for performing rearrangement operations

A data processing apparatus and method are provided for performing rearrangement operations. The data processing apparatus has a register data store with a plurality of registers, each register storing a plurality of data elements. Processing circuitry is responsive to control signals to perform processing operations on the data elements. An instruction decoder is responsive to at least one but no more than N rearrangement instructions, where N is an odd plural number, to generate control signals to control the processing circuitry to perform a rearrangement process at least equivalent to: obtaining as source data elements the data elements stored in N registers of said register data store as identified by the at least one re-arrangement instruction; performing a rearrangement operation to rearrange the source data elements between a regular N-way interleaved order and a de-interleaved order in order to produce a sequence of result data elements; and outputting the sequence of result data elements for storing in the register data store. This provides a particularly efficient technique for performing N-way interleave and de-interleave operations, where N is an odd number, resulting in high performance, low energy consumption, and reduced register use when compared with known prior art techniques.

Modularized extraction, transformation, and loading for a database

Techniques exporting data and metadata from a source database environment to a target database environment are provided. The techniques include the steps of analyzing metadata that describes one or more items, the data for which is in a source database, where the one or more items include at least one of a view, a sequence, a dimension, a cube, an ETL mapping, and any database object for which the metadata is stored outside of the source and target databases. The data for each item resides in a data file associated with the source database. The data for each item is incorporated into the target database based on the metadata by providing the target database access to an incorporated data file, where the incorporated data file is the data file or a copy thereof. Techniques are also provided for exporting database data from the source database. The techniques include extracting metadata that describes one or more items, the data for which is in the source database, where the one or more items include at least one of the structures described above. The data for each item resides in a data file associated with the source database. After the exporting, a database server that manages the target database is provided access to an incorporated data file, where the incorporated data file is the data file or a copy thereof.

Using shrinkable read-once snapshots for online data backup

The present invention discloses a method and system for snapshot-based online backup operations permitting reduced requirements to storage capacity and computational overhead for snapshot function.
At the beginning of an online backup operation, the backup software system creates a snapshot of source data storage. The snapshot includes a watch-list used for identifying blocks of a source storage which are watched by snapshot management means for update. If a block included into the watch-list was requested for update, the snapshot management means preserve original contents of that block in a retention container for the purpose of temporary store. The retention container includes a set of temporal stores dedicated for transient storing of blocks until they are backed up.
The essence of the invention is enabling to exclude blocks from the watch-list and the retention container at any moment within the period of snapshot operation. Therefore it is possible to exclude unnecessary blocks from the scope of blocks managed by the snapshot management means, for the purpose of preserving point-in-time data.
Backed up blocks can be operatively excluded from the snapshot so that unchanged blocks are excluded from the watch-list and updated blocks are removed from the retention container. In the latter case temporal stores are shrunk as well. This technique allows to reduce progressively storage expenses and computational overheads required for maintenance of a snapshot being used in the online backup routine.
When a volume-level online backup is performed the snapshot is switched to the read-once mode at the beginning of the data copying stage. A backup utility performs sequential read of blocks from the snapshot. The snapshot management means automatically exclude requested blocks from the scope of managed blocks.
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products