Abstract | ||
---|---|---|
Research on big data analytics is entering in the new phase called fast data where multiple gigabytes of data arrive in the big data systems every second. Modern big data systems collect inherently complex data streams due to the volume, velocity, value, variety, variability, and veracity in the acquired data and consequently give rise to the 6Vs of big data. The reduced and relevant data streams are perceived to be more useful than collecting raw, redundant, inconsistent, and noisy data. Another perspective for big data reduction is that the million variables big datasets cause the curse of dimensionality which requires unbounded computational resources to uncover actionable knowledge patterns. This article presents a review of methods that are used for big data reduction. It also presents a detailed taxonomic discussion of big data reduction methods including the network theory, big data compression, dimension reduction, redundancy elimination, data mining, and machine learning methods. In addition, the open research issues pertinent to the big data reduction are also highlighted. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1007/s41019-016-0022-0 | Data Science and Engineering |
Keywords | Field | DocType |
Big data, Data compression, Data reduction, Data complexity, Dimensionality reduction | Open research,Data science,Data mining,Data stream mining,Dimensionality reduction,Computer science,Complex data type,Curse of dimensionality,Data compression,Big data,Data reduction | Journal |
Volume | Issue | ISSN |
1 | 4 | 2364-1541 |
Citations | PageRank | References |
6 | 0.44 | 30 |
Authors | ||
6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Muhammad Habib Ur Rehman | 1 | 154 | 13.92 |
Chee Sun Liew | 2 | 82 | 7.78 |
Assad Abbas | 3 | 83 | 9.72 |
Prem Prakash Jayaraman | 4 | 7 | 1.15 |
Teh Ying Wah | 5 | 126 | 9.82 |
Samee U. Khan | 6 | 1572 | 83.04 |