earticle

논문검색

A New Data Mining Algorithm based on MapReduce and Hadoop

초록

영어

The goal of data mining is to discover hidden useful information in large databases. Mining frequent patterns from transaction databases is an important problem in data mining. As the database size increases, the computation time and required memory also increase. Base on this, we use the MapReduce programming mode which has parallel processing ability to analysis the large-scale network. All the experiments were taken under hadoop, deployed on a cluster which consists of commodity servers. Through empirical evaluations in various simulation conditions, the proposed algorithms are shown to deliver excellent performance with respect to scalability and execution time.

목차

Abstract
 1. Introduction
 2. Fast Newman Parallel Algorithm
  2.1. The Newman Algorithm with Modularity
  2.2. Newman Parallel Algorithm and Modularity
 3. The Comparison Algorithm with this Article
  3.1. Partitioning Around Medoids
  3.2. Clustering Large Applications
 4. The Simulation and Conclusion
 References

저자정보

  • Xianfeng Yang School of Information Engineering, Henan Institute of Science and Technology, Xinxiang, Henan, P.R.CHINA
  • Liming Lian Department of Mechanical and Electrical Engineering, Xinxiang University, Xinxiang Henan, P.R.CHINA

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      ※ 원문제공기관과의 협약기간이 종료되어 열람이 제한될 수 있습니다.

      0개의 논문이 장바구니에 담겼습니다.