Abstract

Discriminative learning methods are widely used in natural language processing. These methods work best when their training and test data are drawn from the same distribution. For many NLP tasks, however, we are confronted with new domains in which labeled data is scarce or non-existent. In such cases, we seek to adapt existing models from a resource-rich source domain to a resource-poor target domain. We introduce structural correspondence learning to automatically induce correspondences among features from different domains. We test our technique on part of speech tagging and show performance gains for varying amounts of source and target training data, as well as improvements in target domain parsing accuracy using our improved tagger.

Keywords

Computer scienceDiscriminative modelDomain adaptationArtificial intelligenceParsingDomain (mathematical analysis)Natural language processingAdaptation (eye)Machine learningTest dataLabeled dataTraining setTransfer of learning

Affiliated Institutions

Related Publications

Publication Info

Year
2006
Type
article
Pages
120-120
Citations
1550
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1550
OpenAlex

Cite This

John Blitzer, Ryan McDonald, Fernando Pereira (2006). Domain adaptation with structural correspondence learning. , 120-120. https://doi.org/10.3115/1610075.1610094

Identifiers

DOI
10.3115/1610075.1610094