Ranking research

A new approach to evaluating research papers exploits social bookmarking tools to extract relevance. Details are reported in the latest issue of the International Journal of Internet Technology and Secured Transactions.

Social bookmarking systems are almost indispensible. Very few of us do not use at least one system whether it's Delicious, Connotea, Trunk.ly, Digg, Reddit or any of countless others. For Academics and researchers CiteULike is one of the most popular and has been around since November 2004. CiteUlike allows users to bookmark references but also embeds more conventional bibliographic management. As users of such systems quickly learn the only way to make them useful for others is to ensure that you tag your references comprehensively, but selectively.

On the whole, is very useful but it could be even more so if, rather than using similarity or query-dependent ranking for generating search results if it had a better ranking system.

Researchers in Thailand have now proposed "CiteRank", a combination of a similarity ranking with a static ranking. "Similarity ranking measures the match between a query and a research paper index," they explain. "While a static ranking, or a query-independent ranking, measures the quality of a research paper." Siripun Sanguansintukul of Chulalongkorn University in Bangkok and colleagues have used a group of factors including number of groups citing the posted paper, year of publication, research paper post date, and priority of a research paper to determine a static ranking score, which is then combined with the query-independent measure to give the CiteRank.

The team tested their new ranking by asking literature researchers to rate the results it produced in ranking obtained from the search engines based on an index that uses, TTA, tag-title-abstract. The weighted algorithm CiteRank 80:20 in which a combination of similarity ranking 80% and static ranking 20% was most effective. They found that many literature researchers preferred to read more recent paper or just-posted papers but they also rated highly classic papers that emerged in the results if they were posted across different user groups or communities. Users found good papers based on priority rating but TTA was still important.

"CiteRank combines static ranking with similarity ranking to enhance the effectiveness of the ranking order," explains Sanguansintukul. "Similarity ranking measures the similarity of the text (query) with the document. Static ranking employed the factors posted on paper. Four factors used are: year of publication, posted time, priority rating and number of groups that contained the posted paper."

"Improving indexing not only enhances the performance of academic paper searches, but also all document searches in general. Future research in the area consists of extending the personalization; creating user profiling and recommender system on research paper searching." the team says. The experimental factors that emerged from the study, can help in the optimization of the algorithm to adjust rankings and to improve search results still further.

More information: "CiteRank: combination similarity and static ranking with research paper searching" in International Journal of Internet Technology and Secured Transactions, 2011, 3, 161-177

Citation: Ranking research (2011, May 3) retrieved 19 March 2024 from https://phys.org/news/2011-05-ranking-research.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

IST researchers categorize social media searches

0 shares

Feedback to editors