Zongyu (Johnson) Lin「林宗裕」
I am a second-year CS Ph.D Student at UCLA, co-advised by Prof. Yizhou Sun and Prof. Kaiwei Chang. Before coming to UCLA, I have spent a year in Moonshot.AI (one of the earliest core members), working as a full-time research engineer working on LLM and VideoGen. I was one of the major contributor of pre-training large language models with extremely long context, leading to KIMI CHAT 2023, achieving the state-of-the-art performance on many long context tasks compared with GPT4 and Claude2 at that time. I completed my bachelor's degree of Electronic Enginnering at Tsinghua University. I worked with Prof. Zhilin Yang from 2021 to 2023.
My research interest lies broadly in scalable deep generative architecture and LLM reasoning. Most Recently, I am interested in (1) studying the self-evolution paradigm of large foundation models as well as (2) exploring scalable architectures and recipes for multi-modal generation. Feel free to contact me for chat or discussion if you are also interested in these topics.
Email: lzyxx17 [at] gmail.com
Twitter  
Google Scholar  
Linkedin
|
|
News
2024-12 Check our latest paper from Apple on transparent video generation recipe towards Sora is available now at preprint.
2024-12 Check our latest paper about egocentric and long video understanding: MM-Ego: Towards Building Egocentric Multimodal LLMs is available now at preprint.
2024-11 Our recent paper on benchmarking physical commonsense alignment of video generation: VideoPhy: Evaluating Physical Commonsense In Video Generation, get accepted by Vision-Language@NIPS 2024 as **Oral**, please check our preprint.
2024-6 Our new paper on contradiction retrieval: SPARSECL: Sparse Contrastive Learning for
Contradiction Retrieval is available now at preprint.
|
Experience
Research Intern, Apple, 2024
• Main contributor of STIV, a large videogen model, outperforming PIKA, GEN-3, KLING on VBench.
Research Engineer, Moonshot AI 2023,
• Long-context scaling, videogen, main contributor of KIMI CHAT 2023.
Quant Researcher Intern, Ubiquant, Top Hedge Fund in China. 2022
Research Intern, Sensetime, China, 2021
|
Research Topic
My research interest lies broadly in natural language processing and general machine learning. Most Recently, I am interested in
(1) exploring scalable architectures and recipes for multi-modal generation;
(2) improving the self-evolution of LLMs / VLMs;
(3) improve better alignment with the physical world for vision generative models and vision language models
|
Recent Work
|
|
STIV: Scalable Text and Image Conditioned Video Generation
Zongyu Lin*,Wei Liu*, Chen Chen, Jiasen Lu, Wenze Hu, Tsu-Jui Fu, Jesse Allardice, Zhengfeng Lai, Liangchen Song, Bowen Zhang, Cha Chen, Yiran Fei, Yifan Jiang, Lezhi Li, Yizhou Sun, Kai-Wei Chang, Yinfei Yang
*Equal Contribution
arXiv, 2024
preprint, huggingface (#1 Paper of the day)
Apple's latest recipe and studies for scalable video generation models🔥🔥🔥. In this work, we aim at providing a transparent and detailed recipe 📖 for model architecture, training strategy and data for scalable text-image conditioned video generation. The pretrained T2V and TI2V models outperform SOTA open-sourced, close-sourced models like Gen-3, PIKA, KLING and CogVideoX-5B on VBench.
|
|
VideoPhy: Evaluating Physical Commonsense In Video Generation
Hritik Bansal*,
Zongyu Lin*,
Jing Zhou,
Tianyi Xie, Zeshun Zong, Michal Yarom, Yonatan Bitton, Chenfanfu Jiang, Yizhou Sun, Kai-Wei Chang, Aditya Grover
*Equal Contribution
arXiv, 2024, Oral paper accepted by VL@NIPS, 2024
preprint, website, code
We present VideoPhy, a benchmark designed to assess whether the generated videos follow physical commonsense for real-world activities (e.g. marbles will roll down when placed on a slanted surface). Specifically, we curate a list of 688 captions that involve interactions between various material types in the physical world (e.g., solid-solid, solid-fluid, fluid-fluid).
|
|
MM-Ego: Towards Building Egocentric Multimodal LLMs
Hanrong Ye, Haotian Zhang, Erik Daxberger, Lin Chen, Zongyu Lin, Yanghao Li, Bowen Zhang, Haoxuan You, Dan Xu, Zhe Gan, Jiasen Lu, Yinfei Yang
arXiv, 2024
preprint
This research aims to comprehensively explore building a multimodal foundation model for egocentric video understanding, covering training data construction, model and evaluation benchmark.
|
|
SPARSECL: Sparse Contrastive Learning for
Contradiction Retrieval
Haike Xu*,
Zongyu Lin*,Yizhou Sun, Kai-Wei Chang, Piotr Indyk
*Equal Contribution
arXiv, 2024
preprint, website, code
We propose a novel sparsity-aware contrastive method to solve contradiction retrieval problem, outperforming traditional bi-encoder and cross encoder baselines, which can benefit fact checking and pretraining data filtering.
|
|
VDebugger: Harnessing Execution Feedback for Debugging Visual Programs
Xueqing Wu, Zongyu Lin, Songyan Zhao, Te-Lin Wu, Pan Lu, Nanyun Peng, Kai-Wei Chang
EMNLP findings, 2024
preprint, code
We train a generative judge (critique model) to improve visual programs of CodeLLMs.
|
|
A Universal Discriminator for Zero-Shot Generalization
Haike Xu,
Zongyu Lin,
Jing Zhou,
Yanan Zheng, Zhilin Yang
ACL Oral, 2023
preprint, code
We are the very pioneering work to study how to train the LLM as a reward model and study its zero-shot generalization on different NLP tasks.
|
|
NOT ALL TASKS ARE BORN EQUAL: UNDERSTANDING ZERO-SHOT GENERALIZATION
Jing Zhou,
Zongyu Lin,
Yanan Zheng, Zhilin Yang
ICLR Spotlight, 2023
paper, code
We are the very pioneering work to study the task transfer between different NLP tasks using LLMs. We propose several methods to improve the generalization of LLMs on zero-shot tasks from a data-centric perspective.
|
|
Learning to Detect Noisy Labels Using Model-Based Features
Zhihao Wang*,
Zongyu Lin*,
Peiqi Liu, Guidong Zheng, Junjie Wen, Xianxin Chen, Yujun Chen, Zhilin Yang
(* First Co-Authors)
Findings of EMNLP, 2023
preprint, code
We are the early work to do self-training on LLMs, and we propose a novel model-based feature to detect noisy labels in the self-training process.
|
|
|
Hagen: Homophily-aware graph convolutional recurrent network for crime forecasting
Zongyu Lin*,
Chenyu Wang*,
Guozhen Zhang,
Xiaochen Yang,
Jiao Sun, Mingxuan Yue, Cyrus Shahabi
(* First Co-Authors)
Proceedings of the AAAI Conference on Artificial Intelligence, 2022
paper, code
We propose an end-to-end graph convolutional recurrent network called HAGEN with several novel designs for crime prediction. Specifically, our framework could jointly capture the crime correlation between regions and the temporal crime dynamics by combining an adaptive region graph learning module with the Diffusion Convolution Gated Recurrent Unit (DCGRU). Based on the homophily assumption of GNN (i.e., graph convolution works better where neighboring nodes share the same label), we propose a homophily-aware constraint to regularize the optimization of the region graph so that neighboring region nodes on the learned graph share similar crime patterns.
|
|
Vehicle Trajectory Recovery on Road Network Based on Traffic Camera Video Data
Zongyu Lin,
Guozhen Zhang,
Zhiqun He,
Jie Feng,
Wei Wu,
Yong Li
Proceedings of the 29th International Conference on Advances in Geographic Information Systems, 2021
paper
We propose a general system to recover vehicle trajectories at the level of the road intersection, where a novel iterative framework is developed to combine both vehicle clustering and trajectory recovery tasks.
|
|
HealthWalks: Sensing Fine-grained Individual Health Condition via Mobility Data
Zongyu Lin,
Shiqing Lyu,
Hancheng Cao,
Yuqiong Wei,
Pan Hui,
Hanan Samet,
Yong Li
In ACM International Joint Conference on Pervasive and Ubiquitous Computing (UBICOMP), 2020
paper
We propose a DFA-based model which can generate interpretable features automatically from raw mobility data for fine-grained health sensing.
|
|
SUME: Semantic-enhanced Urban Mobility Network Embedding for User Demographic Inference
Fengli Xu*,
Zongyu Lin*,
Tong Xia,
Diansheng Guo,
Yong Li
(* Equal Contributions)
In ACM International Joint Conference on Pervasive and Ubiquitous Computing (UBICOMP), 2020
paper
We propose a semantic-enhanced urban mobility embedding model for user profiling, and reveal meaningful patterns in all spatial, temporal and urban structure domains.
|
|
CrimeForecaster: Crime Prediction by Exploiting the Neighborhoods’ Spatiotemporal Dependencies
Jiao Sun,
Mingxuan Yue,
Zongyu Lin,
Xiaochen Yang,
Gabe Kahn,
Luciano Nocera,
Cyrus Shahabi
The European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD) , 2020
We introduce a new end-to-end spatiotemporal learning framework dubbed CrimeForecaster that: 1) represents the geographical extents of neighborhoods and their correlations in a graph; 2) uses graph convolution to predict crimes.
|
Selected Awards
Comprehensive Outstanding Scholarship(~10/280), Tsinghua University. 2020
Excellent Technology Innovation Scholarship, Tsinghua University. 2020
First Prize in Software Design Contest, Department of Electronic Enginnering, Tsinghua University. 2018
|
Hobbies
Sports! I really enjoy playing ballgames like football and tennis. I am a big fan of Lionel Messi, Rafael Nadal and Stephen Curry! Also, I love running, swimming and hiking.
|
Updated at June.2024. Thanks Jon Barron for this concise and beautiful template.
|