Home > Events > Master Thesis Defense - Andrew Wood

Master Thesis Defense - Andrew Wood

Start: 3/19/2019 at 1:00PM
End: 3/19/2019 at 4:00PM
Location: 100 Stinson Remick
Attendees: Faculty and students are welcome to attend the presentation portion of the defense.
Add to calendar:
iCal vCal

Andrew Wood

Master Thesis Defense

March 19, 2019        1:00 pm        100 Stinson 

Adviser:  Dr. Collin McMillan 

Committee Members:   

Dr. David Chiang      Dr. Jane Cleland-Huang

Title:

"LOW DATA DIALOGUE ACT CLASSIFICATION FOR VIRTUAL AGENTS DURING DEBUGGING"

 Abstract

A “dialogue act” is a written or spoken action during a conversation. Dialogue acts are usually only a few words long, and are divided by researchers into a relatively small set (often less than 10) of dialogue act types, such as eliciting information, expressing an opinion, or making a greeting. Research interest into automatic classification of dialogue acts has grown recently due to the proliferation of Virtual Agents (VA) e.g. Siri, Cortana, Alexa. But unfortunately, the gains made into VA development in one domain are generally not applicable to other domains, since the composition of dialogue acts differs in different conversations. In this thesis, I target the problem of dialogue act classification for a VA assistant to software engineering repairing bugs in a low data setting. A problem in the SE domain is that very little sample data exists. Therefore, I present a transfer learning approach to learn on a much larger data set for general business conversations, and apply the knowledge to a manually created corpus of debugging conversations collected from 3 0 professional developers in a “Wizard of Oz” experiment and manually annotated with a predetermined dialogue act set. In experiments, we observe between 8% and 20% improvements over two key baselines. Additionally, I present a separate dialogue act classifier on the manually collected data set that uses a manually discovered SE specific dialogue act set which achieves on average 69% precision and 50% recall over 5-fold cross validation.