Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model

View through CrossRef
Recent research indicates that pretraining cross-lingual language models on large-scale unlabeled texts yields significant performance improvements over various cross-lingual and low-resource tasks. Through training on one hundred languages and terabytes of texts, cross-lingual language models have proven to be effective in leveraging high-resource languages to enhance low-resource language processing and outperform monolingual models. In this paper, we further investigate the cross-lingual and cross-domain (CLCD) setting when a pretrained cross-lingual language model needs to adapt to new domains. Specifically, we propose a novel unsupervised feature decomposition method that can automatically extract domain-specific features and domain-invariant features from the entangled pretrained cross-lingual representations, given unlabeled raw texts in the source language. Our proposed model leverages mutual information estimation to decompose the representations computed by a cross-lingual model into domain-invariant and domain-specific parts. Experimental results show that our proposed method achieves significant performance improvements over the state-of-the-art pretrained cross-lingual language model in the CLCD setting.
Title: Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model
Description:
Recent research indicates that pretraining cross-lingual language models on large-scale unlabeled texts yields significant performance improvements over various cross-lingual and low-resource tasks.
Through training on one hundred languages and terabytes of texts, cross-lingual language models have proven to be effective in leveraging high-resource languages to enhance low-resource language processing and outperform monolingual models.
In this paper, we further investigate the cross-lingual and cross-domain (CLCD) setting when a pretrained cross-lingual language model needs to adapt to new domains.
Specifically, we propose a novel unsupervised feature decomposition method that can automatically extract domain-specific features and domain-invariant features from the entangled pretrained cross-lingual representations, given unlabeled raw texts in the source language.
Our proposed model leverages mutual information estimation to decompose the representations computed by a cross-lingual model into domain-invariant and domain-specific parts.
Experimental results show that our proposed method achieves significant performance improvements over the state-of-the-art pretrained cross-lingual language model in the CLCD setting.

Related Results

Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
Hubungan Perilaku Pola Makan dengan Kejadian Anak Obesitas
<p><em><span style="font-size: 11.0pt; font-family: 'Times New Roman',serif; mso-fareast-font-family: 'Times New Roman'; mso-ansi-language: EN-US; mso-fareast-langua...
RESEARCHING WRITTEN MONUMENTS IN THE CONTEXT OF CHANGING SCIENTIFIC PARADIGMS
RESEARCHING WRITTEN MONUMENTS IN THE CONTEXT OF CHANGING SCIENTIFIC PARADIGMS
The scientific paradigm of the 21st century has acquired anthropocentric drift. In modern linguistic studies, the anthropocentric approach also occupies a dominant position: the re...
Deep Unsupervised Domain Adaptation with Time Series Sensor Data: A Survey
Deep Unsupervised Domain Adaptation with Time Series Sensor Data: A Survey
Sensors are devices that output signals for sensing physical phenomena and are widely used in all aspects of our social production activities. The continuous recording of physical ...
Mental practice of lingual resistance and cortical plasticity in older adults: An exploratory fNIRS study
Mental practice of lingual resistance and cortical plasticity in older adults: An exploratory fNIRS study
Purpose: Mental practice using motor imagery (MP) improves motor strength and coordination in the upper and lower extremities in clinical patient populations. Its ...
Lingual nerve injury during extraction of mandibular third molars: Are we extracting correctly?
Lingual nerve injury during extraction of mandibular third molars: Are we extracting correctly?
Introduction: Neurological injuries including lingual nerve damage are among the most devastating complications of mandibular third molar extraction. Various causes of lingual nerv...
FRENECTOMIA LINGUAL NA PEDIATRIA
FRENECTOMIA LINGUAL NA PEDIATRIA
A frenectomia lingual é um procedimento cirúrgico que consiste na remoção do freio lingual, que é uma pequena prega de tecido localizada na parte inferior da língua. Essa cirurgia ...
Electromyographic analysis of the suprahyoid muscles in infants based on the lingual fraenulum attachment during breastfeeding
Electromyographic analysis of the suprahyoid muscles in infants based on the lingual fraenulum attachment during breastfeeding
Introduction Muscle electrical activity analysis can aid in the identification of oral motor dysfunctions such as those resulting from altered lingual fraenulum which, in turn, imp...

Back to Top