Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
TocHeadingTitle
Date
Availability
1-2 of 2
Véronique Hoste
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
We Usually Don’t Like Going to the Dentist: Using Common Sense to Detect Irony on Twitter
Open AccessPublisher: Journals Gateway
Computational Linguistics (2018) 44 (4): 793–832.
Published: 01 December 2018
FIGURES
| View All (7)
Abstract
View articletitled, We Usually Don’t Like Going to the Dentist: Using Common Sense to Detect Irony on Twitter
View
PDF
for article titled, We Usually Don’t Like Going to the Dentist: Using Common Sense to Detect Irony on Twitter
Although common sense and connotative knowledge come naturally to most people, computers still struggle to perform well on tasks for which such extratextual information is required. Automatic approaches to sentiment analysis and irony detection have revealed that the lack of such world knowledge undermines classification performance. In this article, we therefore address the challenge of modeling implicit or prototypical sentiment in the framework of automatic irony detection. Starting from manually annotated connoted situation phrases (e.g., “flight delays,” “sitting the whole day at the doctor’s office”), we defined the implicit sentiment held towards such situations automatically by using both a lexico-semantic knowledge base and a data-driven method. We further investigate how such implicit sentiment information affects irony detection by assessing a state-of-the-art irony classifier before and after it is informed with implicit sentiment information.
Journal Articles
Publisher: Journals Gateway
Computational Linguistics (2016) 42 (3): 457–490.
Published: 01 September 2016
FIGURES
| View All (10)
Abstract
View articletitled, All Mixed Up ? Finding the Optimal Feature Set for General Readability Prediction and Its Application to English and Dutch
View
PDF
for article titled, All Mixed Up ? Finding the Optimal Feature Set for General Readability Prediction and Its Application to English and Dutch
Readability research has a long and rich tradition, but there has been too little focus on general readability prediction without targeting a specific audience or text genre. Moreover, although NLP-inspired research has focused on adding more complex readability features, there is still no consensus on which features contribute most to the prediction. In this article, we investigate in close detail the feasibility of constructing a readability prediction system for English and Dutch generic text using supervised machine learning. Based on readability assessments by both experts and crowdsourcing, we implement different types of text characteristics ranging from easy-to-compute superficial text characteristics to features requiring deep linguistic processing, resulting in ten different feature groups. Both a regression and classification set-up are investigated reflecting the two possible readability prediction tasks: scoring individual texts or comparing two texts. We show that going beyond correlation calculations for readability optimization using a wrapper-based genetic algorithm optimization approach is a promising task that provides considerable insights in which feature combinations contribute to the overall readability prediction. Because we also have gold standard information available for those features requiring deep processing, we are able to investigate the true upper bound of our Dutch system. Interestingly, we will observe that the performance of our fully automatic readability prediction pipeline is on par with the pipeline using gold-standard deep syntactic and semantic information.