ir and nlp

6
Combining the Best of Two Worlds: NLP and IR for Intranet Search Suma Adindla and Udo Kruschwitz School of Computer Science and Electronic Engineering University of Essex Wivenhoe Park, Colchester, C04 3SQ, UK {sadind,udo}@essex.ac.uk  Abstract—Na tural language processing (NLP) is becoming muc h more robus t and app lic abl e in re ali sti c app lic ati ons . One ar ea in which NLP has st il l not been full y expl oi te d is info rmat ion ret riev al (IR). In parti cular we are interested in search ov er int ranets and oth er loc al Web sit es. We see dialogue-driven search which is based on a largely automated kno wle dge extrac tio n pro cess as one of the nex t big ste ps. Instead of replying with a set of documents for a user query the syste m woul d allo w the user to navigat e thro ugh the extra cted knowledge base by making use of a simple dialogue manager. Here we support this idea with a rst task-based evaluation that we condu cted on a univ ersit y intra net. We automat ically ex- tracted entities like person names, organizations and locations as well as relations between entities and added visual graphs to the search results whenever a user query could be mapped into this knowledge base. We found that users are willing to inter act and use those visual inter faces. We also found that users prefered such a system that guides a user through the resul t set over a basel ine approac h. The results repres ent an impor tant rst step towa rds full NLP- driv en intra net sear ch.  Keywords -na tur al lan gua ge pro cessing; inf ormation re- trieval; dialogue; domain knowledge; visualization; I. MOTIVATION Imagine we could interact with a university intranet search engine just like with a human person in a natural dialogue. The search engine would automatically extract knowledge from the Web site so that a searcher can be assisted in nding the information required. A student who asks for a particular course can be directed to the most recent lecture notes or the contact details of the lecturer. An external searcher typing in “PhD NLE” could be assisted by allowing him to explore the space of experts and projects available in the area of natura l langu age engine ering . Obvi ously , this infor matio n can change any day and the idea is to have always the most up-to-date facts and relations available to assist a searcher. Currently, we do not have systems which support this type of interaction. However, our aim is to automatically acquire knowledge (a domain model) from the document collection and employ that in an interactive search system. One motivation for a system that guides a user through the search space is the problem of “too many results”. Even queries in document collections of limited size often return a large number of documents, many of them not relevant to the query. Part of the problem is the fact that both on the Web and in intranet search queries tend to be short and short queries always pose ambiguity and uncertanity issues for information retrieval systems [1]. Some form of dialogue based on fee dba ck from the system cou ld be ver y useful in helping the user nd the right results. This combination of NLP and IR we assume is par tic ula rly promi sing and scalable in smaller domains like university intranets or local Web sites. Obviously, most queries can be answered by a standard search engine but the use of NLP tools to extract knowledge can help address ambiguous queries as well as those where there might be only a single relevant document (which is common in an intranet setting). In order to employ a dialogue system we would ideally have access to a domain knowledge base because dialogue systems work well with structure d knowledge bases. Web sites do ha ve some int ernal str uct ure but unl ike produc t catalogues and onli ne shoppi ng si tes they ar e not full y structured and the rst question we face is: How can we acquire suitable knowledge from a document collection to suppor t sys tem-guide d sea rch? We are not int ere ste d in man ual ly extrac ting suc h kno wle dge , but we would like to automa te that process so that we can appl y the same approach to a new document collection without expensive manual customization. A related question is: What kind of knowledge should this knowledge base (the domain model) contain? We pr opose a system that guides a user in the sear ch process which relies on a database automatically populated by proces sing the document collection and extra cting pieces of kno wle dge from the se doc ume nts . Alo ng wit h named ent iti es, rel ati ons tha t exi st bet we en those entities are es- sentia l in va rious pra cti cal applic ations [2]. We use NLP tec hni que s to par se all sentences in the input doc ume nts, extract relati ons (such as subjec t-ver b-obje ct triple s) and then map user queries against these relations. Such relati ons often involve named entities (as objects or subjects or both). Named entities have been found to play a key role in Web and intranet search, e.g. named entities can be used to deal with page ranking problems [3], they have also been found to play a key role in corporate and university search logs [4], [5]. Although we are using a knowledge base, we do not move away from the standard search paradigm. While displaying the results to a user, we combine the domain model with 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology 978-0-7695-4 513-4/11 $26.00 © 2011 IEEE DOI 10.1109/WI-IAT.20 11.187 483 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology 978-0-7695-4 513-4/11 $26.00 © 2011 IEEE DOI 10.1109/WI-IAT.20 11.187 483

Transcript of ir and nlp

7/29/2019 ir and nlp

http://slidepdf.com/reader/full/ir-and-nlp 1/6

Combining the Best of Two Worlds: NLP and IR for Intranet Search

Suma Adindla and Udo Kruschwitz

School of Computer Science and Electronic EngineeringUniversity of Essex

Wivenhoe Park, Colchester, C04 3SQ, UK 

{sadind,udo}@essex.ac.uk 

 Abstract—Natural language processing (NLP) is becomingmuch more robust and applicable in realistic applications.One area in which NLP has still not been fully exploitedis information retrieval (IR). In particular we are interestedin search over intranets and other local Web sites. We seedialogue-driven search which is based on a largely automatedknowledge extraction process as one of the next big steps.Instead of replying with a set of documents for a user query the

system would allow the user to navigate through the extractedknowledge base by making use of a simple dialogue manager.Here we support this idea with a first task-based evaluation thatwe conducted on a university intranet. We automatically ex-tracted entities like person names, organizations and locationsas well as relations between entities and added visual graphsto the search results whenever a user query could be mappedinto this knowledge base. We found that users are willing tointeract and use those visual interfaces. We also found thatusers prefered such a system that guides a user through theresult set over a baseline approach. The results represent animportant first step towards full NLP-driven intranet search.

 Keywords-natural language processing; information re-trieval; dialogue; domain knowledge; visualization;

I. MOTIVATION

Imagine we could interact with a university intranet search

engine just like with a human person in a natural dialogue.

The search engine would automatically extract knowledge

from the Web site so that a searcher can be assisted in finding

the information required. A student who asks for a particular

course can be directed to the most recent lecture notes or the

contact details of the lecturer. An external searcher typing in

“PhD NLE” could be assisted by allowing him to explore

the space of experts and projects available in the area of 

natural language engineering. Obviously, this information

can change any day and the idea is to have always the most

up-to-date facts and relations available to assist a searcher.

Currently, we do not have systems which support this typeof interaction. However, our aim is to automatically acquire

knowledge (a domain model) from the document collection

and employ that in an interactive search system.

One motivation for a system that guides a user through

the search space is the problem of “too many results”. Even

queries in document collections of limited size often return

a large number of documents, many of them not relevant

to the query. Part of the problem is the fact that both on

the Web and in intranet search queries tend to be short and

short queries always pose ambiguity and uncertanity issues

for information retrieval systems [1]. Some form of dialogue

based on feedback from the system could be very useful

in helping the user find the right results. This combination

of NLP and IR we assume is particularly promising and

scalable in smaller domains like university intranets or local

Web sites. Obviously, most queries can be answered by a

standard search engine but the use of NLP tools to extractknowledge can help address ambiguous queries as well as

those where there might be only a single relevant document

(which is common in an intranet setting).

In order to employ a dialogue system we would ideally

have access to a domain knowledge base because dialogue

systems work well with structured knowledge bases. Web

sites do have some internal structure but unlike product

catalogues and online shopping sites they are not fully

structured and the first question we face is: How can we

acquire suitable knowledge from a document collection to

support system-guided search? We are not interested in

manually extracting such knowledge, but we would like

to automate that process so that we can apply the sameapproach to a new document collection without expensive

manual customization. A related question is: What kind of 

knowledge should this knowledge base (the domain model)

contain?

We propose a system that guides a user in the search

process which relies on a database automatically populated

by processing the document collection and extracting pieces

of knowledge from these documents. Along with named

entities, relations that exist between those entities are es-

sential in various practical applications [2]. We use NLP

techniques to parse all sentences in the input documents,

extract relations (such as subject-verb-object triples) and

then map user queries against these relations. Such relationsoften involve named entities (as objects or subjects or both).

Named entities have been found to play a key role in Web

and intranet search, e.g. named entities can be used to deal

with page ranking problems [3], they have also been found

to play a key role in corporate and university search logs

[4], [5].

Although we are using a knowledge base, we do not move

away from the standard search paradigm. While displaying

the results to a user, we combine the domain model with

2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology

978-0-7695-4513-4/11 $26.00 © 2011 IEEE

DOI 10.1109/WI-IAT.2011.187

483

2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology

978-0-7695-4513-4/11 $26.00 © 2011 IEEE

DOI 10.1109/WI-IAT.2011.187

483

7/29/2019 ir and nlp

http://slidepdf.com/reader/full/ir-and-nlp 2/6

the results of a local search engine. Our work is a first step

towards a full dialogue search system. Here we use visual

graphs to present the relations and corresponding terms for

a given query. What we are investigating here is the general

validity of our approach.

I I . RELATED WOR K

Early work on dialogue systems focused on human com-

puter interaction, e.g. ELIZA [6]. Since then a variety of 

task oriented applications have been developed in various

domains. Intially, many of these dialogue systems assisted

the users in travel domains. Examples include ATIS [7]

and PHILIPS [8]. Generally speaking, one can distinguish

different types of dialogue systems, e.g. a well defined

structured database can be considered as Type I and systems

which lack knowledge bases or deal with unstructured data

as Type II [9]. Dialogue systems based on ontologies would

fall under Type I [10], our work would start from Type II.

Another possible way of guiding the user in the navigation

of search results is through faceted search. A lot of researchhas been done in this area and we can also see commercial

online sites and even libraries1 support this feature. However,

the difference is that faceted search systems typically rely

on well-structured databases, in other words they make use

of rich structure in the knowledge bases [11].

Question-answering (QA) systems are related to our work 

as they tend to rely on similar NLP techniques that we apply

although the main idea of a QA system is to return an

answer rather than a list of documents. The first question

answering systems were only natural language interfaces to

structured databases [12]. Progress in Information Extraction

has more recently contributed much to the success of factoid

based question answering systems. To support interaction, anelement of dialogue has been added to a number of question

answering systems, e.g. [13]. An example of high quality

interaction question answering system is HITIQA [14].

We extract knowledge entirely from unstructured data

available on (e.g.) a university website. This makes our work 

different from the above mentioned dialogue systems.

In recent years, Web search algorithms have matured

significantly by adapting to the users’ information needs.

An example is named entity recognition. Named entities are

becoming increasingly popular in Web search. A study has

shown that 71% of Web queries constitute named entities

and identifying entities in a query further improves retrieval

performance [15]. Analysing the search logs we have beencollecting at the University confirms this observation. The

sort of entities people search for might not coincide with

typically identified ones such as dates, organisations and

locations. In our logs it was found that queries like person

names, room numbers, labs, course titles etc were routinely

searched for. In addition to that 10% of our search queries

1http://search.trln.org/search.jsp

Figure 1. Overview of the system components

consist of person names. This evidence indeed supports our

work and also recommends the need for query type identi-

fication. Like on the Web, we can categorize user queries

into some general types: information needs, browsing and

transactional etc. [16]. The use of Web search engines haswitnessed quite a bit of progress in that respect, compared to

that intranet users still experience poor search results [17].

It has however been shown that understanding a query type

(who, where, when) would be quite useful in an intranet

domain [4].

Another area that is worth exploring in information re-

trieval is visualization. Information visualization is an im-

portant aspect in information retrieval systems. Also, visual

interfaces are excellent tools for interacting and exploring

search results. Various studies have been conducted to test

the significance of visualization for information retrieval

systems [18], [19] suggest the use of various visualization

methods for information organization. For information re-trieval systems, the presentation of search results is still a

challenging issue [20]. One example of search system which

supports entity level search is [21] EntityCube2. Another

example is Google’s3 Wonder Wheel.

III. TOWARDS DIALOGUE-DRIVEN INTRANET SEARCH

Our system consists of two parts: offline knowedge

extraction and an online mapping process that maps the

query into the extracted knowledge. With the help of NLP

tools and information extraction techniques, we process the

document collection automatically to build a domain model.

In an offline extraction process we extract named entities

and predicate argument structures from all documents of 

the local Web site at hand. We thus turn the university

Web document collection into a usable knowledge base

by populating it with named entities and simple facts. To

identify entities (person, organization and location names),

we use the Annie IE system that is part of the Gate 4 NLP

2http://entitycube.research.microsoft.com/index.aspx3http://www.googlewonderwheel.com/ 4http://gate.ac.uk/ 

484484

7/29/2019 ir and nlp

http://slidepdf.com/reader/full/ir-and-nlp 3/6

toolkit. We use GATE, but any similar NLP tool could be

employed. For extracting simple facts, we use the Stanford

parser and our extraction methodology is similar to [22].

The extracted triples are represented in terms of subject-

verb-object pairs.

Along with triplet relations, we have also extracted de-

pendency/predicate relations from the sentences. We will

consider different ways of aiding users by suggesting various

query options. This knowledge can then be used to guide the

dialogue manager and our extracted relations are similar to

the ones presented by [23]. Figure 1 shows an overview of 

the system architecture.

In the second part, we try to map a user query against the

knowledge base. The key component of our system is the

dialogue manager and it is also responsible for the online

mapping process. Whenever a user submits a query, the

dialogue manager tries to map the query against the domain

model and simultaneously submits it to search engine. For

any user query which can be found in the knowledge base

the dialogue manager allows the user to navigate through

the knowledge base by presenting the relations that map

the user query in some way (e.g. if the user query is a

named entity that has relations with other named entities in

the database). When displaying search results to a user, we

combine the extracted domain knowledge with the results

of a local search engine. Figure 2 shows a screenshot of 

our dialogue system. Results from the search engine are

presented alongside a graph of extracted knowledge related

to the query. In the figure the query is shown in the centre

and edges present the dialogue manager suggestions for

the above query. We use various colour codes to illustrate

different types of terms. The green ones are the entities and

the red colour terms indicate the relations. When a user

clicks on any one of those entities the corresponding search

box automatically updates with the clicked entity. With

this interface also a user could interact during information

searching. We have used JIT5 for visualization purposes.

Semantic graphs are starting to get used in assisted search,

e.g. in question answering [24].

 A. Domain Model

In the first stage we identified named entities such as per-

son names, organization names, and locations. We capture

simple facts (relation between entities) from the sentences by

using the Stanford parser and populate the database with this

knowledge. This gives us a structured database, a network of related terms and the corresponding relations. A user query

can then be matched against any part of this knowledge base.

If the user query was a person name (very common in our

domain), the dialogue manager would come up with various

suggestions (department, role, contact details, other people,

projects etc.). By using this piece of information, we frame

5http://thejit.org/ 

questions, generate answers and vice versa as shown in the

motivating example, similar to [25] but more generic.

We have two separate tables one for the entities and the

other for relations. While query mapping, we extract the

terms that match the user query from both the tables. With

the proposed dialogue system (which will eventually go

beyond a graphical interaction) a user could also engage

in a dialogue, but the user is obviously not required to do

so.

IV. EVALUATION

We conducted this first evaluation to explore the potential

of the outlined idea.

The methodology we employed to evaluate our dialogue

system is a task-based evaluation. We followed the TREC6

interactive track guidelines for comparing two systems. Here

we are comparing our system against a baseline system. The

two systems can be characterized as follows:

1) System A is the baseline system which is the search

engine currently installed at the local university Web

site.

2) System B is our system that works based on the

automatically extracted domain model to guide a user

in the navigation of search results. Here the domain

entities and relations are represented in visual graphs

with various colour codes.

Both systems index the same document collection (they

both use Nutch as a backend system.) System A is the

UKSearch system [26], [27]. This system also suggests

some query modification terms to refine and relax the query

(presented as a flat list of links). We assume this is a

fair comparison because it has been shown in previous

experiments on the same university Web site that usersclearly prefer a search engine that makes suggestions over a

Google-style search engine [27]. We therefore consider the

current search engine a sensible baseline to compare against.

We will now explain the experimental procedure and

later we will discuss the results. While conducting the

evaluation, users were not told anything about the underlying

differences between the both systems. For the task-based

evaluation we used the questionnaires introduced by the

TREC-9 interactive series. These four questionnaires were

employed:

1) Entry Questionnaire

2) Postsearch Questionnaire

3) Postsystem Questionnaire4) Exit Questionnaire

 A. Procedure

According to TREC interactive track guidelines at least

16 participants and 8 search tasks are required to conduct

an evaluation that compares two systems in a task-based

6http://www-nlpir.nist.gov/projects/t9i/qforms.html

485485

7/29/2019 ir and nlp

http://slidepdf.com/reader/full/ir-and-nlp 4/6

Figure 2. Dialogue system screenshot

evaluation. We recruited 16 students from the university

population (actual target users in this context). Search tasks

were designed based on query logs obtained on the uni-

versity’s intranet search engine. The terms in parentheses

are the queries found in the logs based on which the tasks

were constructed. These terms were not included in the

instructions for the subjects. Two sample tasks are:

• Task 1 (course) Imagine you are an undergraduate

student and wish to study for a master in economics at

Essex. Find a document that provides details of various

MSc programs in economics.

• Task 5 (car parking): Imagine you are attending a

seminar at the university. Please find a document which

gives details about visitor car parking areas and appli-

cable charges.

We explained the experimental setup and showed one

example on both systems before the evaluation process.

Initially, users started with the entry questionnaire. Each

subject was then asked to perform 4 search tasks on System

A and the remaining four on System B (or the other wayround). Tasks, systems and subjects were permutated based

on the Latin square matrix used by [28]. Subjects were

given 5 minutes to perform each task. After performing

each search task the users had to fill in the postsearch

questionnaire. Along with the questionnaire, they were asked

to submit the answer and rate their task success. When all

the four tasks were finished on one system, users were given

the postsystem questionnaire to be filled in. In the end, users

filled in the exit questionnaire.

V. RESULTS AND DISCUSSION

Of our 16 participants 13 were male and 3 were fe-

male studying in various departments with a range of age

(between 19 and 32) and experience (e.g. online search

experience between 4 to 12 years). Most of the users were

postgraduate students studying for a Masters degree or a PhD

and the remaining were undergraduates. For the question on

searching behavior, the majority of subjects (13) selected 5

and the remaining ones selected 4 (where 5 indicates “daily”

and 4 indicates “weekly”). Among our participants, 8 users

agreed that they enjoy carrying out information searches.

After completion of each task, users filled in the post-

search questionnaire. The following questions with 5-point

Likert scale ratings were used for both systems (where 1

indicates ”not at all” and 5 indicates ”extremely”). To study

the significance, t-tests have been conducted for comparison

wherever necessary:

1) Are you familiar with this topic?2) Was it easy to get started on this search?

3) Was it easy to do the search on this topic?

4) Are you satisfied with your search results?

5) Did you have enough time to do an effective search?

For the question Was it easy to get started on this

search? users prefered System B (but without statistical

significance). In regards to the question Was it easy to do the

search on this topic? users also found it easier to search on

486486

7/29/2019 ir and nlp

http://slidepdf.com/reader/full/ir-and-nlp 5/6

     S   y   s    t   e   m

     F   a   m     i     l     i   a   r

     E   a   s   y    t   o     S    t   a   r    t

     E   a   s   y    t   o     S   e   a   r   c     h

     S   a    t     i   s     f   a   c    t     i   o   n

     E   n   o   u   g     h     T     i   m   e

A 3.32 4.09 4.09 4.53 4.68

B 3.09 4.26 4.45 4.57 4.68

Table IPOSTSEARCH QUESTIONNAIRE

Parameter System-A System-B No Difference

Easier to learn to use 4 6 6

Easier to use 1 11 4

Best 4 10 2

Table IIEXI T QUESTIONNAIRE (SYSTEM PREFERENCE)

System B. This value is statistically significant when all the

8 search tasks were considered (p <0.01). This value clearlyindicates the usefulness of domain model suggestions. Also

for the next question Are you satisfied with your search

results? users were more satisfied with the results returned

by System B. Table I summarizes the results. For most of the

above questions System B was slightly better than System A

(though not always statistically significant). The table also

illustrates that sufficient time was allocated for the tasks.

Users filled in the postsystem questionnaire after complet-

ing four search tasks on one system. When we compared

the values on both systems, the differences between them

are marginal.

Finally, users submitted an exit questionnaire. For the

question Which of the two systems did you find easier touse? 11 users picked System B and only one user opted

for System A. Furthermore, 10 users selected System B as

the best system overall, 4 users selected System A and 2

users did not find any difference between them. Table II

demonstrates that System B scored overall better than the

baseline system. The results of the exit questionnaire clearly

demonstrate the potential that this type of guided search

offers in the context of a university intranet.

We also asked for additional user feedback in the exit

questionnaire. Users liked the idea of visualizing search

terms in a graph and most of the users did in fact select the

query options suggested by System B despite the varying

quality of the extracted knowledge (this noise was alsocommented on by a user). However, in this evaluation we did

not target the quality of the extracted knowledge (which is a

separate issue). In our evaluation we represented the first 10

entities that matched the user query from the database and

did not make use of any frequency or ranking parameter.

One user commented It is useful to know not to always stick 

to the same search engine as there are others that could be

 just as useful.

This evaluation is our first validation of the outlined idea

that promotes the use of deep NLP in order to extract facts

and relations from document collections which can then be

used to guide a user who searches this collection. Users

were overall more satisfied with a system that makes use

of extracted facts and relations when communicating results

to the user. We found that the idea of NLP-based search

in document collections appears to be a promising route

based on this simple task-based evaluation reported here.

Furthermore, we see a lot of potential in combining NLP

techniques with state-of-the-art visualization methods.

VI. CONCLUSIONS AND FUTURE WOR K

We presented a task-based evaluation assessing the use-

fulness of incorporating a dialogue component in a search

system. We particularly targeted local Web sites such as

university intranets. The sort of dialogue system we applied

makes use of small pieces of knowledge extracted from the

document collection (and linked in a simple term network)

that can then be mapped against the query. We found that thegeneral idea of such guided search offers a lot of potential.

This work can obviously only be a first step. There are a

number of limitations in such a study and we will take the

findings as a guideline for future work. We will investigate

a variety of routes. First of all, the system we investigated

used a visual representation. We will continue doing so but

will enrich the dialogue by adding more of a real NLP

dialogue paraphrasing the knowledge found in the database.

We also aim at putting a prototype of the search engine

online so that we address a number of limitations that user

studies such as the one presented here face. Finally, the

knowledge extraction process is still not perfect and we are

still working on finding the right balance between the qualityand the quantity of relations and entities extracted from the

documents.

ACKNOWLEDGMENTS

We would like to thank the anonymous reviewers for

very helpful feedback on an earlier version of the pa-

per. This work is partially supported by the AutoAdapt7

research project. AutoAdapt is funded by EPSRC grants

EP/F035357/1 and EP/F035705/1.

REFERENCES

[1] B. Sun, P. Liu, and Y. Zheng, “Short query refinement with

query derivation,” in Proceedings of the 4th Asia informa-tion retrieval conference on Information retrieval technology.Berlin, Heidelberg: Springer-Verlag, 2008, pp. 620–625.

[2] D. Bollegala, Y. Matsuo, and M. Ishizuka, “Unsupervisedextraction of semantic relations between entities on the web,”in 19th Int’l World Wide Web Conf. (WWW 2010) , Raleigh,North Carolina, USA, 2010, pp. 151–160.

7http://autoadaptproject.org

487487

7/29/2019 ir and nlp

http://slidepdf.com/reader/full/ir-and-nlp 6/6

[3] A. Caputo, P. Basile, and G. Semeraro, “Boosting a semanticsearch engine by named entities,” in Proceedings of the18th International Symposium on Foundations of Intelligent Systems. Berlin, Heidelberg: Springer-Verlag, 2009, pp. 241–250.

[4] H. Li, Y. Cao, J. Xu, Y. Hu, S. Li, and D. Meyerzon, “A newapproach to intranet search based on information extraction,”

in Proceedings of CIKM’05, Bremen, Germany, 2005, pp.460–468.

[5] R. Sutcliffe, K. White, and U. Kruschwitz, “Named entityrecognition in an intranet query log,” in Proceedings of the

 LREC Workshop on Semitic Languages, Valletta, Malta, 2010,pp. 43–49.

[6] J. Weizenbaum, “Eliza - a computer program for the study of natural language communication between man and machine,”Commun. ACM , vol. 9, no. 1, pp. 36–45, January 1966.

[7] S. Seneff, L. Hirschman, and V. W. Zue, “Interactive problemsolving and dialogue in the atis domain,” in Proceedings of the workshop on Speech and Natural Language. Stroudsburg,PA, USA: Association for Computational Linguistics, 1991,

pp. 354–359.

[8] H. Aust, M. Oerder, F. Seide, and V. Steinbiss, “The philipsautomatic train timetable information system,” Speech Com-mun., vol. 17, no. 3-4, pp. 249–262, November 1995.

[9] Y. cheng Pan and L. shan Lee, “Type-ii dialogue systemsfor information access from unstructured knowledge sources,”in IEEE Automatic Speech Recognition and UnderstandingWorkshop, Kyoto, Japan, December 2007.

[10] P. Quaresma and I. P. Rodrigues, “Using dialogues to accesssemantic knowledge in a web ir system,” in Proceedings of the 6th international conference on Computational processingof the Portuguese language, 2003, pp. 201–205.

[11] O. Ben-Yitzhak, N. Golbandi, N. Har’El, R. Lempel, A. Neu-mann, S. Ofek-Koifman, D. Sheinwald, E. Shekita, B. Sz-najder, and S. Yogev, “Beyond basic faceted search,” inProceedings of the international conference on Web searchand web data mining. New York, NY, USA: ACM, 2008,pp. 33–44.

[12] B. F. Green, A. K. Wolf, C. Chomsky, and K. Laughery,“Baseball: an automatic question-answerer,” in Papers pre-sented at the May 9-11, 1961, western joint IRE-AIEE-ACM computer conference. New York, NY, USA: ACM, 1961,pp. 219–224.

[13] S. Gandhe, A. S.Gordon, and D. Traum, “Improving question-answering with linking dialogues,” in Proceedings of the 11thinternational conference on Intelligent user interfaces. New

York, NY, USA: ACM, 2006, pp. 369–371.

[14] S. Small and T. Strzalkowski, “Hitiqa: High-quality intelli-gence through interactive question answering,” Natural Lan-guage Engineering, vol. 15, no. 1, pp. 31–54, 2009.

[15] J. Guo, G. Xu, X. Cheng, and H. Li, “Named entity recogni-tion in query,” in 32nd international ACM SIGIR conferenceon Research and development in information retrieval (SIGIR’09). New York, USA: ACM, 2009.

[16] M. Kellar, C. Watters, and M. Shepherd, “A field studycharacterizing web-based information-seeking tasks,” Journalof the American Society for Information Science and Technol-ogy, vol. 58, no. 7, pp. 999–1018, 2007.

[17] H. Zhu, S. Raghavan, S. Vaithyanathan, and A. Loser, “Nav-igating the intranet with high precision,” in Proceedings of the 16th international conference on World Wide Web. New

York, NY, USA: ACM, 2007, pp. 491–500.

[18] S. Koshman, “Testing user interaction with a prototypevisualization-based information retrieval system,” Journal of the American Society for Information Science and Technol-ogy, vol. 56, no. 8, pp. 824–833, 2005.

[19] X. Yuan, X. Zhang, and A. Trofimovsky, “Testing visualiza-tion on the use of information systems,” in Proceedings of thethird symposium on Information interaction in context IIiX ’10. New York, USA: ACM, 2010.

[20] O. Turetken and R. Sharda, “Visualization of web spaces:state of the art and future directions,” SIGMIS Database,vol. 38, no. 13, pp. 51–81, July 2007.

[21] J. Zhu, Z. Nie, X. Liu, B. Zhang, and J. rong Wen, “Statisticalmethods statsnowball: a statistical approach to extractingentity relationships,” WWW 2009 MADRID, 2009.

[22] D. Rusu, L. Dali, B. Fortuna, M. Grobelnik, and D. Mladeni,“Triplet extraction from sentences,” in Proceedings of the 10th

 International Multiconference,Information Society-IS 2007 ,2007.

[23] A. Akbik and J. Bross, “Wanderlust: Extracting semantic rela-tions from natural language text using dependency grammarpatterns,” in Workshop on Semantic Search in Conjunctionwith the 18th Int. World Wide Web Conference (WWW2009) ,Madrid, Spain, 2009.

[24] L. Dali, D. Rusu, B. Fortuna, D. Mladenic, and M. Grobelnik,“Question answering based on semantic graphs,” in Workshopon Semantic Search in Conjunction with the 18th Int. World Wide Web Conference (WWW2009), Madrid, Spain, 2009.

[25] P. Adolphs, X. Cheng, T. Klwer, H. Uszkoreit, and F. Xu,“Question answering biographic information and social net-work powered by the semantic web,” in Proceedings of theSeventh conference on International Language Resources and 

 Evaluation (LREC’10). Valletta, Malta: European LanguageResources Association (ELRA), 2010, pp. 19–21.

[26] U. Kruschwitz and H. Al-Bakour, “Users want more sophis-ticated search assistants: Results of a task-based evaluation,”

 Journal of the American Society for Information Science and Technology, vol. 56, no. 13, pp. 1377–1393, November 2005.

[27] U. Kruschwitz, Intelligent Document Retrieval:Exploiting Markup Structure, ser. The Information Retrieval Series.Springer, 2005, vol. 17.

[28] W. Hersh and P. Over, “Trec-9 interactive track report,” inProceedings of the Ninth Text Retrieval Conference (TREC-9), 2001, pp. 41–50.

488488