Offline analysis method for massive data
A technology for off-line analysis and massive data, applied in the field of off-line analysis, can solve problems such as increasing data robustness and cleanliness, achieve the effect of solving the bottleneck of data collection, improving collection efficiency, and improving efficiency
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Example Embodiment
[0041] Example
[0042] The invention utilizes the distributed storage of massive data of the Hadoop platform cluster and the efficient and fast parallel computing capability, specifically:
[0043] Data collection and preprocessing. For different power big data types, diversified data collection modes can be used. For streaming data, the Kafka collection tool is used to aggregate the streaming data into the Kafka cluster, and then store and process it by HBase. For the relational database, the Sqoop data exchange tool is used, combined with the customizable data cleaning module, and the distributed import of relational data to HDFS or HBase is realized through the Map-Reduce distributed computing framework. At the same time, Sqoop provides an incremental import function of data. For big data files, you need to import them into Hadoop using FTP protocol or localized upload. After the electric power big data is stored in HDFS, the data can be cleaned, converted, filtered and oth...
PUM
Abstract
Description
Claims
Application Information
- R&D Engineer
- R&D Manager
- IP Professional
- Industry Leading Data Capabilities
- Powerful AI technology
- Patent DNA Extraction
Browse by: Latest US Patents, China's latest patents, Technical Efficacy Thesaurus, Application Domain, Technology Topic.
© 2024 PatSnap. All rights reserved.Legal|Privacy policy|Modern Slavery Act Transparency Statement|Sitemap