Multimodal Agents
Designing systems that can search, perceive, and reason over heterogeneous digital environments, with emphasis on grounded multimodal interaction.
Master's student in Artificial Intelligence (MSAI) at Nanyang Technological University, working on multimodal and agent-related research with Associate Professor Ziwei Liu at MMLab@NTU.
My current research focuses on multimodal learning, agent systems, and personalized AI. I previously worked on LLM memorization, data privacy, and tokenization-related security risks with Prof. Michael R. Lyu at CUHK.
I am currently pursuing the Master of Science in Artificial Intelligence (MSAI) at Nanyang Technological University. Since August 2025, I have been working with Associate Professor Ziwei Liu at MMLab@NTU on papers and research projects, with a current focus on multimodal systems and agentic intelligence.
Before NTU, I received my B.Sc. in Computer Science from The Chinese University of Hong Kong, graduating with a CGPA of 3.785/4.0. My earlier work centered on understanding memorization in large language models, with emphasis on data compressibility, tokenization, and security risks in code LLMs.
Designing systems that can search, perceive, and reason over heterogeneous digital environments, with emphasis on grounded multimodal interaction.
Building and evaluating systems that recover user-level context from long-horizon, file-system-scale behavioral and multimodal traces.
Studying how data properties and tokenization affect memorization, leakage risk, and privacy vulnerabilities in large language models.
The papers below include recent work currently under review and ongoing research projects. At this stage, the four papers listed here are all in the review cycle.
A benchmark for evaluating contextual agents over realistic multimodal personal file systems, with a focus on search, evidence perception, and multi-step reasoning.
Explore the interactive website to experience what the three personal computers look like, and visit the project page for the benchmark overview.
A framework for grounding agent memory and personalization in file-system behavioral traces, spanning data generation, benchmarking, and memory architecture.
This work studies how data compressibility relates to memorization in LLMs and proposes a quantitative perspective on memorization behavior.
An investigation into how BPE tokenization contributes to secret memorization and leakage risks in code LLMs through what we term gibberish bias.
Presented at AINIT 2024. Earlier work on document-level information extraction.
Working on multimodal and agent-oriented papers and research projects while pursuing the M.S. in AI at NTU.
Focused on LLM memorization, entropy-based characterization, dataset inference, and tokenization-related security risks in code LLMs.
Worked on automated data acquisition, knowledge graph modeling, fine-tuning, and graph-based analysis pipelines.
Evaluated memorization difficulty in large language models through entropy, perplexity, and memorization-rate analysis on open-source models.
Studied adversarial attacks on gender recognition systems and analyzed robustness and fairness issues in facial-recognition APIs.
Top 1% in Computer Science.
Top 10% in the Faculty of Engineering, CUHK.
Top 1% in the Faculty of Engineering, CUHK.
If you would like to discuss multimodal learning, agent systems, personalized AI, or LLM memorization and security, feel free to reach out.