Computational Cognitive Science

A list of potential topics for PhD students in the area of Computational Cognitive Science.

Neural Network Models of Human Language and Visual Processing

Supervisor: Frank Keller

Recent neural models have used attention mechanisms as a way of focusing the processing of a neural networks on certain parts of the input. This has proved successful for diverse applications such as image description, question answering, or machine translation. Attention is also a natural way of understanding human cognitive processing: during language processing, humans attend words in a certain order; during visual processing, they view image regions in a certain sequence. Crucially, human attention can be captured precisely using an eye-tracker, a device that measures which parts of the input the eye fixates, and for how long. Projects within this area will leverage neural attention mechanisms to model aspects of human attention. Examples include reading: when reading text, humans systematically skip words, spend more time on difficult words, and sometimes re-read passages. Another example is visual search: when looking for a target, human make a sequence of fixations which depend a diverse range of factors, such as visual salience, scene type, and object context. Neural attention models that capture such behaviors need to combine different types of knowledge, while also offering a cognitively plausible story how such knowledge is acquired, often based on only small amounts of training data.

 

Topics in morphology (NLP or cognitive modelling)

Supevisor:  Sharon Goldwater

Many NLP systems developed for English ignore the morphological structure of words and (mostly) get away with it. Yet morphology is far more important in many other languages. Handling morphology appropriately can reduce sparse data problems in NLP, and understanding human knowledge of morphology is a long-standing scientific question in cognitive science. New methods in both probabilistic modeling and neural networks have the potential to improve word representations for downstream NLP tasks and perhaps to shed light on human morphological acquisition and processing. Projects in this area could involve working to combine distributional syntactic/semantic information with morphological information to improve word representations for low-resource languages or sparse datasets, evaluating new or existing models of morphology against human behavioral benchmarks, or related topics.

Topics in unsupervised speech processing and/or modelling infant speech perception

Supervisor: Sharon Goldwater

Work in unsupervised (or 'zero-resource') speech processing encompasses areas such as learning sub-word representations that better capture phonetic or other linguistic information; and identifying repeated units such as phones or words in a target language without any transcribed training data. Systems that solve these tasks could be useful as a way to index and summarize speech files in under-resourced languages. In some cases, they may also be useful as cognitive models, allowing us to investigate how human infants might begin to identify words in the speech stream of their native language.  There are many open questions, so various projects are possible.  Projects could focus mainly on speech technology or mainly on modeling language acquisition; specific research questions will depend on this choice.  Here are just two possibilities: (1) recent work on self-supervised learning (a type of unsupervised learning) has shown that SSL models can learn representations for speech that factor out some of the speaker and contextual variability. However, the nature of these representations, and whether they handle such variability in similar ways to humans, is not well understood. Investigate these questions by developing new ways to analyze and compare SSL representations to each other and to infant or adult speech perception results. (2) existing work in both speech processing and cognitive modeling suggests that trying to learn either words or phones alone may be too difficult and in fact we need to develop *joint learners* that simultaneously learn at both levels.  Investigate models that can be used to do this and evaluate how joint learning can improve performance.