Adapting Event Extractors to Medical Data: Bridging the Covariate Shift

Aakanksha Naik, Jill Fain Lehman, Carolyn Rose

Information Extraction and Text Mining Long paper Paper

Gather-3A: Apr 23, Gather-3A: Apr 23 (13:00-15:00 UTC) [Join Gather Meeting]

You can open the pre-recorded video in separate windows.

Abstract: We tackle the task of adapting event extractors to new domains without labeled data, by aligning the marginal distributions of source and target domains. As a testbed, we create two new event extraction datasets using English texts from two medical domains: (i) clinical notes, and (ii) doctor-patient conversations. We test the efficacy of three marginal alignment techniques: (i) adversarial domain adaptation (ADA), (ii) domain adaptive fine-tuning (DAFT), and (iii) a new instance weighting technique based on language model likelihood scores (LIW). LIW and DAFT improve over a no-transfer BERT baseline on both domains, but ADA only improves on notes. Deeper analysis of performance under different types of shifts (e.g., lexical shift, semantic shift) explains some of the variations among models. Our best-performing models reach F1 scores of 70.0 and 72.9 on notes and conversations respectively, using no labeled target data.
NOTE: Video may display a random order of authors. Correct author list is at the top of this page.

Connected Papers in EACL2021

Similar Papers

We Need To Talk About Random Splits
Anders Søgaard, Sebastian Ebert, Jasmijn Bastings, Katja Filippova,
Clinical Outcome Prediction from Admission Notes using Self-Supervised Knowledge Integration
Betty van Aken, Jens-Michalis Papaioannou, Manuel Mayrdorfer, Klemens Budde, Felix Gers, Alexander Loeser,
Few-shot learning through contextual data augmentation
Farid Arthaud, Rachel Bawden, Alexandra Birch,