| Abundant information sources are provided by web, which have made the web source mining fairly difficult because of huge volume of data and its complexity, extreme dynamic and all kinds of clients and so on. Owing to the self-organized and semi-structured characteristics of web, classical information retrieve and data base technologies can not applied effectively. Hyperlink is the special component and the bond of linking data sources of web. Link analysis is an important approach to improve the equality of web source mining.This paper analyzes web data mining methods, structure of web search engine, characteristics of web link structure and characteristics and relevant questions of main link analysis algorithms in detail. PageRank and HITS algorithms are two classic link analysis ranking algorithms. But PageRank treats all links equally without considering the importance of updating time of old web pages and the influence of navigation web pages. And HITS exists the multiple reinforcement relationship, the phenomenon of topic drift, the unreasonable results and the unsatisfied information requirement of user site granularity and so on. According to the link character of web and browsing process of web walker, it can found that random walker will jump to the similar pages influenced by the current page content but not jump to the out-link pages with equal probability. Based on that, this paper introduces SimRank algorithm by which the similarity between web pages can be calculated and consequently proposes a new score assignment scheme based on similarity between web pages, which would assign large jumping probability to the largely relevant links. Then we propose a new ranking algorithm based on distributed factor (RAD).This paper designs the platform of RAD algorithm including experiment data sets, data base, pruning disposal, SimRank and RAD implementation. Experiment results show that RAD algorithm executes more efficiently than standard PageRank algorithm in relative precision and mean average precision. |