Skip to content

Latest commit

 

History

History
204 lines (147 loc) · 22 KB

File metadata and controls

204 lines (147 loc) · 22 KB


logo of awesome repository

awesome-pretrained-models-for-information-retrieval

A curated list of awesome papers related to pre-trained models for information retrieval (a.k.a., pretraining for IR). If there are any papers I missed, please let me know! And any feedback and contribution are welcome!

Pretraining for IR

We also include the recent Multimodal Pre-training works whose pre-trained models fine-tuned on the cross-modal retrieval tasks such as text-image retrieval in their experiments.

For people who want to acquire some basic&advanced knowledge about neural models for information retrieval and try some neural models by hand, we refer readers to the below awesome NeuIR survey and the text-matching toolkit MatchZoo-py:

Survey Paper

First Stage Retrieval

Neural term weighting framework

Document expansion for Sparse representation

Decouple the dense representation encoding of query and document

Late interaction

Negative sampling

Knowledge distillation

Design pre-training tasks

Joint learn retrieval and index

Dense retrieval in open domain QA

Re-ranking Stage

Pre-trained models for reranking

Straightforward applications

Process long documents

Utilize generative pre-trained models

Efficient Training and query expansion

Weak supervision and pre-training for reranking

Model acceleration

Cross-lingual retrieval

Multimodal Retrieval

Unified Single-stream Architecture

Multi-stream Architecture Applied on Input

Other Resources

Some Retrieval Toolkits

Other Resources About Pre-trained Models in NLP

Surveys About Efficient Transformers