This site is currently broken for new queries. It seems like some bot hit it and we ran out of exa credits.
Enter your personal website to find others like you. Powered by Exa's "find similar". Built on Val Town.
I am a Research Scientist at Google working primarily on transformers and video understanding.
Uses a language model to learn a sequence of actions as temporal context for egocentric action recognition.
same video, and in spite of its simplicity, it works surprisingly well across: (i) different unsupervised
I'm interested in computer vision and its relations to natural language and deep learning, with a focus on learning visual
Hi, I am Yin Cui (崔崟 in Chinese, pronounced as /yin tsui/), a senior research scientist at Google.
I'm interested in computer vision, self-supervised and multi-modal learning as well as causality and privacy in AI.
Ego4d: Around the world in 3,000 hours of egocentric video
Having control over the synthesized person images in terms of pose and style is an important requisite for applications such as e-commerce, virtual r…
I am an assistant professor of computer science at Brown University, where I direct the PALMð´ research lab, studying computer vision, machine lear…