Skip to content

A curated list of awesome papers related to pre-trained models for information retrieval (a.k.a., pretraining for IR).

Notifications You must be signed in to change notification settings

csshali/awesome-pretrained-models-for-information-retrieval

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 

Repository files navigation


logo of awesome repository

awesome-pretrained-models-for-information-retrieval

A curated list of awesome papers related to pre-trained models for information retrieval (a.k.a., pretraining for IR). If there are any papers I missed, please let me know! And any feedback and contribution are welcome!

Pretraining for IR

We also include the recent Multimodal Pre-training works whose pre-trained models fine-tuned on the cross-modal retrieval tasks such as text-image retrieval in their experiments.

For people who want to acquire some basic&advanced knowledge about neural models for information retrieval and try some neural models by hand, we refer readers to the below awesome NeuIR survey and the text-matching toolkit MatchZoo-py:

Survey Paper

First Stage Retrieval

Neural term weighting framework

Document expansion for Sparse representation

Decouple the dense representation encoding of query and document

Late interaction

Negative sampling

Knowledge distillation

Design pre-training tasks

Joint learn retrieval and index

Dense retrieval in open domain QA

Re-ranking Stage

Pre-trained models for reranking

Straightforward applications

Process long documents

Utilize generative pre-trained models

Efficient Training and query expansion

Weak supervision and pre-training for reranking

Model acceleration

Cross-lingual retrieval

Multimodal Retrieval

Unified Single-stream Architecture

Multi-stream Architecture Applied on Input

Other Resources

Some Retrieval Toolkits

Other Resources About Pre-trained Models in NLP

Surveys About Efficient Transformers

About

A curated list of awesome papers related to pre-trained models for information retrieval (a.k.a., pretraining for IR).

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published