This represents a fundamentally different way of thinking about IR systems. Within the index-retrieve-then-rank paradigm, modeling work (e.g., query understanding, document understanding, retrieval, ranking, etc.) is done on top of the index itself. This results in modern IR systems being comprised of a disparate mix of heterogeneous models (e.g.,... See more
Building such domain experts would likely require developing an artificial general intelligence, which is beyond the scope of this paper. Instead, by “domain expert” we specifically mean that the system is capable of producing results (with or without actual “understanding”) that are of the same quality as a human expert in the given domain.
If all of these research ambitions were to come to fruition, the resulting system would be a very early version of the system that we envisioned in the introduction. That is, the resulting system would be able to provide domain expert answers to a wide range of information needs in a way that neither modern IR systems, question answering systems,... See more
We envision using the same corpus model as a multi-task learner for multiple IR tasks. To this end, once a corpus model has been trained, it can of course be used for the most classical of all IR tasks – document retrieval. However, by leveraging recent advances in multi-task learning, such a model can very likely be applied to a diverse range of... See more
[Curator's note: there are numerous other technical challenges addressed with alternative prescriptions throughout the paper. These highlights are narrative-centric, and I invite you to review the paper if you are a keen technologist looking for answers to the following:- Zero- and Few-Shot Learning - Response Generation- Arithmetic, Logical,... See more