IMAGINE: Improving Multi-modal lAnguage Generation wIth world kNowledgE is a research project which main goal is to investigate how to incorporate world knowledge into vision & language tasks within natural language generation. I am a Marie Skwodówska-Curie Global Fellow.
I am spending ~2 years in New York University’s Courant Institute for Mathematical Sciences where I will work with Kyunghyun Cho, followed by 3 months in Paris visiting Antoine Bordes in Facebook Artificial Intelligence Research (FAIR). I will finally return to the Institute for Logic, Language and Computation (ILLC) in the University of Amsterdam, where I will continue collaborating with Raquel Fernández.
Concretely, I investigate how to:
- Gather world-knowledge (semi-)automatically from publicly available multi-modal knowledge bases.
- Learn representations for a knowledge base that encompasses both text and images.
- Integrate this knowledge into multi-modal language generation tasks, such as multi-modal machine translation, visual question answering and image description generation.
Please get in touch with you would like to collaborate on any of these research topics!