Hello! I am a second year PhD student in the Natural Language Processing Lab at Bar-Ilan University, supervised by prof. Yoav Goldberg. I’m working on bridging the gap between modalities, and then revealing the capabilities multimodal models have when considering more than one modality at a time. Currently, my research is focused on image-text alignment and controlled generation (what kind of changes do we need to make to generate images that are more faithful to the input prompt?). I find it particularly interesting to work on better architectures, which can be faster (not attention based?) or simply more scalable than current ones.
Previously, I obtained my Masters in Computer Science, and was jointly supervised by Yoav Goldberg and Reut Tsarfaty. Then, I was driven by fascination with underspecified language. Why do models often misinterpret ambiguous language, and we don’t? How can we make implicit information more explicit? My thesis work aimed to answer these questions in the context of verbal omissions in coordination structures. As a fun detour, I stumbled upon intriguing behavior by DALL-E when fed with ambiguous language, where it depicts multiple interpretations in a single image, instead of settling on one. We detailed this behavior in a short paper.
My CV is available here.
PhD in Computer Science (in-progress)
Bar Ilan University
MSc in Computer Science
Bar Ilan University