Pergi ke luar talian dengan aplikasi Player FM !
Task-aware Retrieval with Instructions
Manage episode 355037182 series 3446693
Andrew Yates (Assistant Prof at University of Amsterdam) and Sergi Castella (Analyst at Zeta Alpha) discuss the paper "Task-aware Retrieval with Instructions" by Akari Asai et al. This paper proposes to augment a conglomerate of existing retrieval and NLP datasets with natural language instructions (BERRI, Bank of Explicit RetRieval Instructions) and use it to train TART (Multi-task Instructed Retriever).
π Paper: https://arxiv.org/abs/2211.09260
π» BEIR benchmark: https://arxiv.org/abs/2104.08663
π LOTTE (Long-Tail Topic-stratified Evaluation, introduced in ColBERT v2): https://arxiv.org/abs/2112.01488
Timestamps:
00:00 Intro: "Task-aware Retrieval with Instructions"
02:20 BERRI, TART, X^2 evaluation
04:00 Background: recent works in domain adaptation
06:50 Instruction Tuning 08:50 Retrieval with descriptions
11:30 Retrieval with instructions
17:28 BERRI, Bank of Explicit RetRieval Instructions
21:48 Repurposing NLP tasks as retrieval tasks
23:53 Negative document selection
27:47 TART, Multi-task Instructed Retriever
31:50 Evaluation: Zero-shot and X^2 evaluation
39:20 Results on Table 3 (BEIR, LOTTE)
50:30 Results on Table 4 (X^2-Retrieval)
55:50 Ablations
57:17 Discussion: user modeling, future work, scale
21 episod
Manage episode 355037182 series 3446693
Andrew Yates (Assistant Prof at University of Amsterdam) and Sergi Castella (Analyst at Zeta Alpha) discuss the paper "Task-aware Retrieval with Instructions" by Akari Asai et al. This paper proposes to augment a conglomerate of existing retrieval and NLP datasets with natural language instructions (BERRI, Bank of Explicit RetRieval Instructions) and use it to train TART (Multi-task Instructed Retriever).
π Paper: https://arxiv.org/abs/2211.09260
π» BEIR benchmark: https://arxiv.org/abs/2104.08663
π LOTTE (Long-Tail Topic-stratified Evaluation, introduced in ColBERT v2): https://arxiv.org/abs/2112.01488
Timestamps:
00:00 Intro: "Task-aware Retrieval with Instructions"
02:20 BERRI, TART, X^2 evaluation
04:00 Background: recent works in domain adaptation
06:50 Instruction Tuning 08:50 Retrieval with descriptions
11:30 Retrieval with instructions
17:28 BERRI, Bank of Explicit RetRieval Instructions
21:48 Repurposing NLP tasks as retrieval tasks
23:53 Negative document selection
27:47 TART, Multi-task Instructed Retriever
31:50 Evaluation: Zero-shot and X^2 evaluation
39:20 Results on Table 3 (BEIR, LOTTE)
50:30 Results on Table 4 (X^2-Retrieval)
55:50 Ablations
57:17 Discussion: user modeling, future work, scale
21 episod
Semua episod
Γ
1 AGI vs ASI: The future of AI-supported decision making with Louis Rosenberg 54:42

1 EXAONE 3.0: An Expert AI for Everyone (with Hyeongu Yun) 24:57

1 Zeta-Alpha-E5-Mistral: Finetuning LLMs for Retrieval (with Arthur CΓ’mara) 19:35

1 ColPali: Document Retrieval with Vision-Language Models only (with Manuel Faysse) 34:48

1 Using LLMs in Information Retrieval (w/ Ronak Pradeep) 22:15

1 Designing Reliable AI Systems with DSPy (w/ Omar Khattab) 59:57

1 The Power of Noise (w/ Florin Cuconasu) 11:45

1 Benchmarking IR Models (w/ Nandan Thakur) 21:55

1 Baking the Future of Information Retrieval Models 27:05

1 Hacking JIT Assembly to Build Exascale AI Infrastructure 38:04

1 The Promise of Language Models for Search: Generative Information Retrieval 1:07:31

1 Task-aware Retrieval with Instructions 1:11:13

1 Generating Training Data with Large Language Models w/ Special Guest Marzieh Fadaee 1:16:14

1 ColBERT + ColBERTv2: late interaction at a reasonable inference cost 57:30

1 Evaluating Extrapolation Performance of Dense Retrieval: How does DR compare to cross encoders when it comes to generalization? 58:30

1 Open Pre-Trained Transformer Language Models (OPT): What does it take to train GPT-3? 47:12

1 Few-Shot Conversational Dense Retrieval (ConvDR) w/ special guest Antonios Krasakis 1:23:11

1 Transformer Memory as a Differentiable Search Index: memorizing thousands of random doc ids works!? 1:01:40

1 Learning to Retrieve Passages without Supervision: finally unsupervised Neural IR? 59:10

1 The Curse of Dense Low-Dimensional Information Retrieval for Large Index Sizes 54:13

1 Shallow Pooling for Sparse Labels: the shortcomings of MS MARCO 1:07:17
Selamat datang ke Player FM
Player FM mengimbas laman-laman web bagi podcast berkualiti tinggi untuk anda nikmati sekarang. Ia merupakan aplikasi podcast terbaik dan berfungsi untuk Android, iPhone, dan web. Daftar untuk melaraskan langganan merentasi peranti.