A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit <a rel="external noopener" href="https://businessperspectives.org/images/pdf/applications/publishing/templates/article/assets/12930/PPM_2019_04_Rashid.pdf">the original URL</a>. The file type is <code>application/pdf</code>.
<i title="LLC CPC Business Perspectives">
<a target="_blank" rel="noopener" href="https://fatcat.wiki/container/lq35a5722nfr5e4owcnngrsbzy" style="color: black;">Problems and Perspectives in Management</a>
The term "big data" characterizes the massive amounts of data generation by the advanced technologies in different domains using 4Vs – volume, velocity, variety, and veracity - to indicate the amount of data that can only be processed via computationally intensive analysis, the speed of their creation, the different types of data, and their accuracy. High-dimensional financial data, such as time-series and space-time data, contain a large number of features (variables) while having a small<span class="external-identifiers"> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.21511/ppm.17(4).2019.28">doi:10.21511/ppm.17(4).2019.28</a> <a target="_blank" rel="external noopener" href="https://fatcat.wiki/release/76yr472o6rf7vm3torvgnxfcnm">fatcat:76yr472o6rf7vm3torvgnxfcnm</a> </span>
more »... r of samples, which are used to measure various real-time business situations for financial organizations. Such datasets are normally noisy, and complex correlations may exist between their features, and many domains, including financial, lack the al analytic tools to mine the data for knowledge discovery because of the high-dimensionality. Feature selection is an optimization problem to find a minimal subset of relevant features that maximizes the classification accuracy and reduces the computations. Traditional statistical-based feature selection approaches are not adequate to deal with the curse of dimensionality associated with big data. Cooperative co-evolution, a meta-heuristic algorithm and a divide-and-conquer approach, decomposes high-dimensional problems into smaller sub-problems. Further, MapReduce, a programming model, offers a ready-to-use distributed, scalable, and fault-tolerant infrastructure for parallelizing the developed algorithm. This article presents a knowledge management overview of evolutionary feature selection approaches, state-of-the-art cooperative co-evolution and MapReduce-based feature selection techniques, and future research directions.
<a target="_blank" rel="noopener" href="https://web.archive.org/web/20200807212320/https://businessperspectives.org/images/pdf/applications/publishing/templates/article/assets/12930/PPM_2019_04_Rashid.pdf" title="fulltext PDF download" data-goatcounter-click="serp-fulltext" data-goatcounter-title="serp-fulltext"> <button class="ui simple right pointing dropdown compact black labeled icon button serp-button"> <i class="icon ia-icon"></i> Web Archive [PDF] <div class="menu fulltext-thumbnail"> <img src="https://blobs.fatcat.wiki/thumbnail/pdf/3b/fe/3bfebbd1f7488e95aaf52d818e05298f56923263.180px.jpg" alt="fulltext thumbnail" loading="lazy"> </div> </button> </a> <a target="_blank" rel="external noopener noreferrer" href="https://doi.org/10.21511/ppm.17(4).2019.28"> <button class="ui left aligned compact blue labeled icon button serp-button"> <i class="unlock alternate icon" style="background-color: #fb971f;"></i> Publisher / doi.org </button> </a>