UCLA researchers in the Department of Computer Science have developed a new algorithm that approximates large frequency moments in big datasets with pick-and-drop sampling for analysis.
With increasing data volume, the ability to analyze the data becomes challenging. In some cases, the data is generated by a single event and stored for analysis, e.g. large simulations (financial or scientific). In other instances, the data is generated by singular simultaneous events, such as daily sales data from online purchases/retailers. While each day's data may be efficiently analyzed, the size the combined data is likely too big for practical in-depth analysis. Approximate frequency moments could be used to analyze retailers weekly or yearly sales figures when analysis of the data becomes impractically large to handle with conventional analysis.
UCLA researcher Rafail Ostrovsky has developed an algorithm to estimate higher frequency moments of a given data stream. The algorithm provides useful statistics on the data set when the incoming data is too big to store or efficiently analyze.
Researchers have created and validated the algorithm.
|United States Of America||Issued Patent||9,158,822||10/13/2015||2013-367|
big data, analytics, big data statistics, sales analytics, big data tools, big data algorithm, pick and drop sampling, pick and drop, algorithm