Optimization of Distributed Crawler under Hadoop

Web crawler is an important link in the data acquisition of the World Wide Web. It is necessary to optimize traditional methods so as to meet the current needs in the face of the explosive growth of data. This paper introduces the process and the model of the current distributed crawler based on Had...

Full description

Bibliographic Details
Main Authors: Zhang Xiaochen, Xian Ming
Format: Article
Language:English
Published: EDP Sciences 2015-01-01
Series:MATEC Web of Conferences
Subjects:
Online Access:http://dx.doi.org/10.1051/matecconf/20152202029