Samiksha Jaiswal (Editor)

Expertise finding

Updated on
Edit
Like
Comment
Share on FacebookTweet on TwitterShare on LinkedInShare on Reddit

Expertise finding is the use of tools for finding and assessing individual expertise, with particular focus on scientific expertise. In recruitment industry, e.g., LinkedIn, expertise finding is the problem of searching for hiring candidates given certain required skills as well as other types of information need, such as, location and companies.

Contents

Importance of expertise

It can be argued that human expertise is more valuable than capital, means of production or intellectual property. Contrary to expertise, all other aspects of capitalism are now relatively generic: access to capital is global, as is access to means of production for many areas of manufacturing. Intellectual property can be similarly licensed. Furthermore, expertise finding is also a key aspect of institutional memory, as without its experts an institution is effectively decapitated. However, finding and “licensing” expertise, the key to the effective use of these resources, remain much harder, starting with the very first step: finding expertise that you can trust.

Until very recently, finding expertise required a mix of individual, social and collaborative practices, a haphazard process at best. Mostly, it involved contacting individuals one trusts and asking them for referrals, while hoping that one’s judgment about those individuals is justified and that their answers are thoughtful.

In the last fifteen years, a class of knowledge management software has emerged to facilitate and improve the quality of expertise finding, termed “expertise locating systems”. These software range from social networking systems to knowledge bases. Some software, like those in the social networking realm, rely on users to connect each other, thus using social filtering to act as “recommender systems”.

At the other end of the spectrum are specialized knowledge bases that rely on experts to populate a specialized type of database with their self-determined areas of expertise and contributions, and do not rely on user recommendations. Hybrids that feature expert-populated content in conjunction with user recommendations also exist, and are arguably more valuable for doing so.

Still other expertise knowledge bases rely strictly on external manifestations of expertise, herein termed “gated objects”, e.g., citation impacts for scientific papers or data mining approaches wherein many of the work products of an expert are collated. Such systems are more likely to be free of user-introduced biases (e.g., ResearchScorecard ), though the use of computational methods can introduce other biases.

More recently, LinkedIn Expertise Search introduces a hybrid approach based on user-generated data (e.g., member profiles), community-based signals (e.g., recommendations and skill endorsements) and personalized signals (e.g., social connection between searcher and results). Given required skills and other types of information need like location and industries, the system allows recruiters to search for hiring candidates amongst more than 450 million LinkedIn members. A recent survey shows that 94% of recruiters use LinkedIn to search for candidates

Examples of the systems outlined above are listed in Table 1.

Table 1: A classification of expertise location systems

Technical problems

A number of interesting problems follow from the use of expertise finding systems:

  • The matching of questions from non-expert to the database of existing expertise is inherently difficult, especially when the database does not store the requisite expertise. This problem grows even more acute with increasing ignorance on the part of the non-expert due to typical search problems involving use of keywords to search unstructured data that are not semantically normalized, as well as variability in how well an expert has set up their descriptive content pages. Improved question matching is one reason why third-party semantically normalized systems such as ResearchScorecard and BiomedExperts should be able to provide better answers to queries from non-expert users.
  • Avoiding expert-fatigue due to too many questions/requests from users of the system (ref. 1).
  • Finding ways to avoid “gaming” of the system to reap unjustified expertise credibility.
  • Infer expertise on implicit skills. Since users typically do not declare all of the skills they have, it is important to infer their implicit skills that are highly related their explicit ones. The inference step can significantly improve recall in expertise finding.
  • Expertise ranking

    Means of classifying and ranking expertise (and therefore experts) become essential if the number of experts returned by a query is greater than a handful. This raises the following social problems associated with such systems:

  • How can expertise be assessed objectively? Is that even possible?
  • What are the consequences of relying on unstructured social assessments of expertise, such as user recommendations?
  • How does one distinguish authoritativeness as a proxy metric of expertise from simple popularity, which is often a function of one's ability to express oneself coupled with a good social sense?
  • What are the potential consequences of the social or professional stigma associated with the use of an authority ranking, such as used in Technorati and ResearchScorecard)?
  • How to make expertise ranking personalized to each individual searcher? This is particularly important for recruiting purpose since given the same skills, recruiters from different companies, industries, locations might have different preferences on candidates
  • Approaches for creating expertise content

  • Manual, either by experts themselves (e.g., Skillhive) or by a curator (Expertise Finder)
  • Automated, e.g., using software agents (e.g., MIT's ExpertFinder and the ExpertFinder initiative) or a combination of agents and human curation (e.g., ResearchScorecard )
  • In industrial expertise search engines (e.g., LinkedIn), there are many signals coming into the ranking functions, such as, user-generated content (e.g., profiles), community-generated content (e.g., recommendations and skills endorsements) and personalized signals (e.g., social connections). Moreover, user queries might contain many other aspects rather required expertise, such as, locations, industries or companies. Thus, traditional information retrieval features like text matching are also important. Learning to rank is typically used to combine all of these signals together into a ranking function
  • Interesting expertise systems over the years

    In no particular order...

  • GuruScan
  • Autonomy's IDOL
  • AskMe
  • Expertise Finder
  • Tacit Knowledge Systems' ActiveNet
  • Triviumsoft's SEE-K
  • MIT’s ExpertFinder (ref 3)
  • MITRE’s (ref 1) Expert Finder
  • MITRE’s XpertNet
  • Arnetminer (ref 2)
  • Dataware II Knowledge Directory
  • Thomson’s tool
  • Hewlett-Packard’s CONNEX
  • Microsoft’s SPUD project
  • ProFinda
  • Xperscore
  • Skillhive
  • WhoKnows
  • LinkedIn
  • Conferences

    1. The ExpertFinder Initiative

    References

    Expertise finding Wikipedia