Overview of NTCIR

16
Overview of NTCIR Charlie Clarke, Noriko Kando NTCIR-14 General Chairs 1 Noriko kando @ ntcir-14 Kickoff

Transcript of Overview of NTCIR

Overview of NTCIR

Charlie Clarke, Noriko Kando

NTCIR-14 General Chairs

1 Noriko kando @ ntcir-14 Kickoff

Project started in late 1997 18 months Cycle

■A series of evaluation workshops designed to enhance research in information-access technologies by providing an infrastructure for evaluations. ■Data sets, evaluation methodologies, forum

NII Testbeds and Community for Information access Research Research Infrastructure for Evaluating Information Access

4146 Users! Data sets (Test collections or TCs)

Scientific, news, patents, web, CQA, Wikipedia, exams, spoken, microblogs, lifelogs

Chinese, Korean, Japanese, and English

Tasks (Research Areas) IR: cross-lingual, patents, web, geo, time,

math, Spoken, recepe, microblogs, lifelogs QA:monolingual, cross-lingual, interactive,

IR for QA, complex, exams summarization, trend info., patent maps,

Inference, 2-layer summarization for mobile opinion analysis, text mining, Intent, link

discovery, Visual, info extraction, Noriko kando @ ntcir-14 Kickoff 2

104

186

109

172

135

112

157

105

28

36

65

74

77

85

82

65

99

108

80

91

65

6

8

9

10

15

12

15

17

17

16

20

20

20

0 50 100 150 200

NTCIR-1

NTCIR-2

NTCIR-3

NTCIR-4

NTCIR-5

NTCIR-6

NTCIR-7

NTCIR-8

NTCIR-9

NTCIR-10

NTCIR-11

NTCIR-12

NTCIR-13

# of countries

# of active participating teams

# of registered teams

Task Participants

3

Tasks in NTCIRs

NTCIR th 1 2 3 4 5 6 7 8 9 10 11 12 Tasksomain Year '99 '01 '02 '04 '05 '07 '08 '10 '11 '13 '14 '16

■ Microblog■ Community QA

■ ■ ■ Opinion Analysis■ IR + Inference

■ ■ QA + IR■ Lifelog images

■ Recepe■ ■ ■ MedNLP■ ■ ■ Math Retrieval

■ ■ ■ ■ Exam (Entailment, QA)□ □ ■ ■ Spoken Q& Doc Retireval

■ ■ Temporalia■ ■ GeoTime Retrieval

■ ■ ■ ■ □ □ Patent Retrieval■ ■ Exam (multi-choice, essay

■ ■ ■ □ Any types of Qs■ ■ Dialogue

■ ■ ■ ■ Cross-Lingual IR■ ■ ■ ■ Factoid, List

IE & IR ■ ■ Link DiscoverySemantic ■ ■ ■ Inference

■ ■ 2-layer (mobile)■ ■ ■ ■ ■ Text Mining/

■ ■ ■ Trend Info Visualization■ ■ ■ Text Summarization

■ Micoblog■ ■ Spoken Query

■ VisualizationMobile ■ ■ MobileClick

■ ■ ■ ■ Intent, Diversification■ ■ SERPT Quality

■ ■ ■ Web Search■ ■ ■ ■ Machine Translation

■ ■ ■ ■ ■ ■ ■ ■ Cross-Lingual IR■ ■ ■ ■ ■ ■ ■ ■ Non-EnglishIR

TetxtRetrieval ■ ■ ■ ■ ■ ■ ■ ■ Ad Hoc, IR for QA

Cross-Lingual

Module-Based

Summarization

Web

Interactive

SocialMedia・UGC

SpecialDomain

QuesitonAnswering

Noriko kando @ ntcir-14 Kickoff

4

Tasks in NTCIRs

Noriko kando @ ntcir-14 Kickoff

MedWeb QA Lab

Lifelog STC AKG ECA NAILS WWW

IR Evaluations

• TREC: Text Retrieval Conference • CLEF: Cross-Language Evaluation Forum Conferece and Labs of the Evaluation Forum

• NTCIR: NII Testbeds and Community for

Information access Research

Noriko kando @ ntcir-14 Kickoff 5

Structure of NTCIR

Noriko kando @ ntcir-14 Kickoff 6

NTCIR

Tasks EVIA

General Co-chairs Steering Committee

EVIA Co-chairs Program Committee Authers

Program Co-chairs Program Committee Task Organizers

Participants

NTCIR changed name 3 times

• Nacsis Test Collection for IR research – NTCIR-1

• NII Test Collection for IR research – NTCIR-2~-8

• NII Testbeds and Community for Information access Research – NTCIR-9

• NTCIR is a Community-Lead Activities for Creating Future Values so, Let’s work Together!

Noriko kando @ ntcir-14 Kickoff 7

Procedures in NTCIRs • Call for Task Proposals

– Selection of Task Proposals by Program Committee – Discussion about Task Design in Each Task

• Task – Registration to Task(s)

• Deliver Training Data (Documents, Topics, Answers) – Experiments and Tuning by Each Participants

• Deliver Test Data (Documents and Topics) – Experiments by Each Participants

– Submission of Experimental Results – Pooling the Answer Candidates from the Submissions, and Conduct Manual

Judgments – Return Answers (Relevance Judgments) and Evaluation Results – Analysis of the results, and improving the systems, etc. – Paper writing for the Conference proceedings

• Conference – Presenting the achievements of this Round – Discussion for the Next Round – Networking

• Test Collection Release for non-participants 8 Noriko kando @ ntcir-14 Kickoff

Kickoff

Pls join Discussion!!

9

Past NTCIR Conference & EVIA

Noriko kando @ ntcir-14 Kickoff

Past NTCIR Conference & EVIA

EVIA has attracted top experts on IR Evaluation

Noriko kando @ ntcir-14 Kickoff 10

11

Past NTCIR Conference Satelite Lifelog Workshop in Glasgow at NTCIR-12 Conference, June 2012

Noriko kando @ ntcir-14 Kickoff

Past NTCIR Conference & EVIA

Travel Supports are provided for Oversee Task/EVIA Young Authors

Noriko kando @ ntcir-14 Kickoff 12

Uses of the Test Collections by Community

13

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 1924 65 134 165 317 412 518 811

1376 1366 1623 1877 2060 2228 2452 2737 2980 3283 3418

101

208 107 97

246 246

371 371

519 721

716 746 # of users (who have effective licenses)

■Research Purpose ■Task Participants

6 16 29 4 38 22 16 42 18

68 92 68 39 87 116 116 132

211

82 18 25

40 27

114 73 90

91 99

82

150

93 107

49

108 138 111

92

53

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19

International Users vs Domestic Users

■International ■Domestic

4,164

Noriko kando @ ntcir-14 Kickoff

LNCS

Courses of a Task

Noriko kando @ ntcir-14 Kickoff 14

Idea Intensive Discussion

Task Proposal

Task at NTCIR

Paper Submissions to

Other Coferences and

Journals

Special Sessions at Conference

Paper at NTCIR Conference

Paper at EVIA

Journal Special Issues

- SWIRL, Dagstuhl, NII Shonan - Post Conference WSs, etc

So far, IP&M, ACM TALIP, IRJ, etc

NII Testbeds and Community for Information access Research Why?

15

-Data sets / Users’ Information Seeking Tasks -Evaluation Methodology -Reusable vs Reproducibility -User-Centered Evaluation -Experimental Platforms -Open Advancement -Advanced NLP Knowledge- or Semantic-based -Diversified IA Applications in the Real World -Best Practice for a technology Best Practice for Evaluation Methodology -Big Data (Documents + Behaviour data)

Noriko kando @ ntcir-14 Kickoff

Mid-term Vision

16

Datasets Large-Scale Reusable TCs

Integrating NLP・IR

Documents -> Answer, Information

Opinion/ Attitudes

Cross-lingual Users’ Information Seeking

Tasks + Genres

Intension

Users Task and Context

Weaving out Knowledge

Module-based

NTCIR-1 NTCIR-3 NTCIR-6 NTCIR-9 NTCIR-12

Interaction & Visualization

Module-based Platform (Open Advancement)

Mobile

Interaction & Live Evaluation

Create Future Value

Conversation

Big Data

Reproducibility

Noriko kando @ ntcir-14 Kickoff NTCIR-14