Transformer-Encoder-Based Mathematical Information Retrieval

EXPERIMENTAL IR MEETS MULTILINGUALITY, MULTIMODALITY, AND INTERACTION (CLEF 2022)(2022)

引用 0|浏览1
暂无评分
摘要
Mathematical Information Retrieval (MIR) deals with the task of finding relevant documents that contain text and mathematical formulas. Therefore, retrieval systems should not only be able to process natural language, but also mathematical and scientific notation to retrieve documents. In this work, we evaluate two transformer-encoder-based approaches on a Question Answer retrieval task. Our pre-trained ALBERT-model demonstrated competitive performance as it ranked in the first place for p'@10. Furthermore, we found that separating the pre-training data into chunks of text and formulas improved the overall performance on formula data.
更多
查看译文
关键词
Mathematical Language Processing, Information Retrieval, BERT-based Models, ARQMath Lab
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要