Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-3 of 3
Jordan Boyd-Graber
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
Trick Me If You Can: Human-in-the-Loop Generation of Adversarial Examples for Question Answering
Open AccessPublisher: Journals Gateway
Transactions of the Association for Computational Linguistics (2019) 7: 387–401.
Published: 01 July 2019
FIGURES
| View All (10)
Abstract
View articletitled, Trick Me If You Can: Human-in-the-Loop Generation of Adversarial Examples for Question Answering
View
PDF
for article titled, Trick Me If You Can: Human-in-the-Loop Generation of Adversarial Examples for Question Answering
Adversarial evaluation stress-tests a model’s understanding of natural language. Because past approaches expose superficial patterns, the resulting adversarial examples are limited in complexity and diversity. We propose human- in-the-loop adversarial generation, where human authors are guided to break models. We aid the authors with interpretations of model predictions through an interactive user interface. We apply this generation framework to a question answering task called Quizbowl, where trivia enthusiasts craft adversarial questions. The resulting questions are validated via live human–computer matches: Although the questions appear ordinary to humans, they systematically stump neural and information retrieval models. The adversarial questions cover diverse phenomena from multi-hop reasoning to entity type distractors, exposing open challenges in robust question answering.
Journal Articles
Publisher: Journals Gateway
Transactions of the Association for Computational Linguistics (2017) 5: 1–16.
Published: 01 January 2017
Abstract
View articletitled, Evaluating Visual Representations for Topic Understanding and Their
Effects on Manually Generated Topic Labels
View
PDF
for article titled, Evaluating Visual Representations for Topic Understanding and Their
Effects on Manually Generated Topic Labels
Probabilistic topic models are important tools for indexing, summarizing, and analyzing large document collections by their themes. However, promoting end-user understanding of topics remains an open research problem. We compare labels generated by users given four topic visualization techniques—word lists, word lists with bars, word clouds, and network graphs—against each other and against automatically generated labels. Our basis of comparison is participant ratings of how well labels describe documents from the topic. Our study has two phases: a labeling phase where participants label visualized topics and a validation phase where different participants select which labels best describe the topics’ documents. Although all visualizations produce similar quality labels, simple visualizations such as word lists allow participants to quickly understand topics, while complex visualizations take longer but expose multi-word expressions that simpler visualizations obscure. Automatic labels lag behind user-created labels, but our dataset of manually labeled topics highlights linguistic patterns (e.g., hypernyms, phrases) that can be used to improve automatic topic labeling algorithms.
Journal Articles
Online Adaptor Grammars with Hybrid Inference
Open AccessPublisher: Journals Gateway
Transactions of the Association for Computational Linguistics (2014) 2: 465–476.
Published: 01 October 2014
Abstract
View articletitled, Online Adaptor Grammars with Hybrid Inference
View
PDF
for article titled, Online Adaptor Grammars with Hybrid Inference
Adaptor grammars are a flexible, powerful formalism for defining nonparametric, unsupervised models of grammar productions. This flexibility comes at the cost of expensive inference. We address the difficulty of inference through an online algorithm which uses a hybrid of Markov chain Monte Carlo and variational inference. We show that this inference strategy improves scalability without sacrificing performance on unsupervised word segmentation and topic modeling tasks.