mPMR: A Multilingual Pre-trained Machine Reader at Scale
Weiwen Xu, Xin Li, Wai Lam, Lidong Bing
Main: Multilingualism and Cross-Lingual NLP Main-poster Paper
    Poster Session 1: Multilingualism and Cross-Lingual NLP (Poster)
    
Conference Room: Frontenac Ballroom and Queen's Quay 
    Conference Time: July 10, 11:00-12:30 (EDT) (America/Toronto)
    Global Time: July 10, Poster Session 1 (15:00-16:30 UTC)
    
    
  
          Keywords:
          cross-lingual transfer, multilingual pre-training
        
        
        
        
          TLDR:
          We present multilingual Pre-trained Machine Reader (mPMR), a novel method for multilingual machine reading comprehension (MRC)-style pre-training. mPMR aims to guide multilingual pre-trained language models (mPLMs) to perform natural language understanding (NLU) including both sequence classificatio...
        
  
    You can open the
    #paper-P713
    channel in a separate window.
  
  
    
            Abstract:
            We present multilingual Pre-trained Machine Reader (mPMR), a novel method for multilingual machine reading comprehension (MRC)-style pre-training. mPMR aims to guide multilingual pre-trained language models (mPLMs) to perform natural language understanding (NLU) including both sequence classification and span extraction in multiple languages. To achieve cross-lingual generalization when only source-language fine-tuning data is available, existing mPLMs solely transfer NLU capability from a source language to target languages. In contrast, mPMR allows the direct inheritance of multilingual NLU capability from the MRC-style pre-training to downstream tasks. Therefore, mPMR acquires better NLU capability for target languages. mPMR also provides a unified solver for tackling cross-lingual span extraction and sequence classification, thereby enabling the extraction of rationales to explain the sentence-pair classification process.