LEXICAL SEMANTICS
I am very much interested in statistical models of Natural Language. I am fascinated by the semantics that can be discovered from the company a word keeps. Although word distributions are extremely poor approximations of the mental lexicon, they tend to be very useful! I worked on quality-quantity trade-offs for distributional semantics models. Now, I am working on fusing statistical and human-compiled semantic dictionaries to yield better models of language.
<REPORT> Sridharan, S. “Using Google Books Ngrams for Neurolinguistic Decoding”
MULTIMODAL LANGUAGE UNDERSTANDING
I am excited about the utility of multimodal evidence streams for dialog agents. I think there is a lot more to communication that we discard by just looking at speech. I am working on fusing gestural, speech and space cues to enhance human-agent interaction. I am studying how the mutual information between these multiple streams can be used for robust input understanding. I am also exploring how Electroencephalography(EEG) can be used to decode dialog relevant mental states that the dialog system can react to. In other words, how can we leverage non-verbal cues?
SITUATED DIALOG
I built embodied conversational agents that use the Microsoft Kinect, a far-field microphone array and an advanced vision sensor for audio source localization, user awareness and voice activity detection. I am interested in the challenges posed by public environments for social dialog agents. This is ongoing work with my peers at the LACS lab. I believe that fluid interaction with robots in public spaces is an hard problem that needs to be attacked with multi-sensory perception.
MACHINE LEARNING (FOR NLP)
I worked on question quality prediction in online QA communities using both offline and online models of the question posted. I modeled the abstract measure of quality using multiple concrete metrics that I annotated. I also explored how we can better estimate the quality measure in the online scenario using co-training, where a large number of labelled instances were exchanged between the two models for training, thereby sharing knowledge.
MACHINE LEARNING (FOR SIGNALS)
I worked on adapting an acoustic model (for Speech Recognition) to a particular microphone/microphone array. I modeled the channel of the microphone by learning a transformation function from the recording of a frequency chirp signal that would give the original acoustic signal. This was used to learn an adapted acoustic models by distorting pre-recorded audio data.
<MODEL> Sridharan, S.: An acoustic model trained on the WSJ audio, adapted to the Kinect sensor.
SPEECH
I built an intelligent calendar assistant that let users manage (add, delete and query) calendar events using speech, where the input language is designed to be unrestricted. I used the Olympus Framework to build the system and the UIUC QA corpus to adapt the dialog. In another study, I explored whether disfluencies in speech can be used to improve word predictability in spontaneous speech. Using the Switchboard corpus I showed that speech disfluencies do carry information that can help decrease language perplexity.
<REPORT> Sridharan, S., Yun, W.: Personal Calendar Assistant
<REPORT> Sridharan, S.: Do Speech Disfluencies help Word Predictability?
PAGE UNDER CONSTRUCTION!
© 2012 Seshadri Sridharan.