Sitemap

A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.

Pages

Posts

Future Blog Post

less than 1 minute read

Published:

This post will show up by default. To disable scheduling of future posts, edit config.yml and set future: false.

Blog Post number 4

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 3

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 2

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 1

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

portfolio

publications

LM-Searcher: Cross-domain Neural Architecture Search with LLMs via Unified Numerical Encoding

Published in EMNLP 2025, 2025

We propose LM-Searcher, a novel approach for cross-domain neural architecture search using LLMs via unified numerical encoding.

Recommended citation: Yuxuan Hu, Jihao Liu, Ke Wang, Jinliang Zhen, Weikang Shi, Manyuan Zhang, Qi Dou, Rui Liu, Aojun Zhou, Hongsheng Li (2025). "LM-Searcher: Cross-domain Neural Architecture Search with LLMs via Unified Numerical Encoding." EMNLP 2025. https://arxiv.org/abs/2509.05657

Pusa v1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized Timestep Adaptation

Published in ICLR 2026, 2026

We present Pusa v1.0, an Image-to-Video model that surpasses Wan-I2V with only $500 training cost using vectorized timestep adaptation.

Recommended citation: Yaofang Liu, Yumeng Ren, Aitor Artola, Yuxuan Hu, Xiaodong Cun, Xiaotong Zhao, Alan Zhao, Raymond H. Chan, Suiyun Zhang, Rui Liu, Dandan Tu, Jean-Michel Morel (2026). "Pusa v1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized Timestep Adaptation." ICLR 2026. https://arxiv.org/abs/2507.16116

MathCanvas: Intrinsic Visual Chain-of-Thought for Multimodal Mathematical Reasoning

Published in The 64th Annual Meeting of the Association for Computational Linguistics (ACL 2026), 2026

We present MathCanvas, the first unified multi-model that enables thinking while drawing auxiliary lines for multimodal mathematical reasoning.

Recommended citation: Weikang Shi, Aldrich Yu, Rongyao Fang, Houxing Ren, Ke Wang, Aojun Zhou, Changyao Tian, Xinyu Fu, Yuxuan Hu, Zimu Lu, Linjiang Huang, Si Liu, Rui Liu†, Hongsheng Li (2026). "MathCanvas: Intrinsic Visual Chain-of-Thought for Multimodal Mathematical Reasoning." ACL 2026. https://mathcanvas.github.io/

Beyond Confidence: Adaptive and Coherent Decoding for Diffusion Language Models

Published in ICML 2026, 2026

We present Beyond Confidence, a novel adaptive and coherent decoding method for diffusion language models.

Recommended citation: Kecheng Chen, Ziru Liu, Xijia Tao, Hui Liu, Xinyu Fu, Suiyun Zhang, Dandan Tu, Lingpeng Kong, Rui Liu, Haoliang Li (2026). "Beyond Confidence: Adaptive and Coherent Decoding for Diffusion Language Models." ICML 2026.

PhoStream: Benchmarking Real-World Streaming for Omnimodal Assistants in Mobile Scenarios

Published in ICML 2026, 2026

We present PhoStream, a benchmark for real-world streaming evaluation of omnimodal assistants in mobile scenarios.

Recommended citation: Xudong Lu, Huankang Guan, Yang Bo, Jinpeng Chen, Xintong Guo, Shuhan Li, Fang Liu, Peiwen Sun, Xueying Li, Wei Zhang, Xue Yang, Rui Liu, Hongsheng Li (2026). "PhoStream: Benchmarking Real-World Streaming for Omnimodal Assistants in Mobile Scenarios." ICML 2026.

SpaceVista: All-Scale Visual Spatial Reasoning from mm to km

Published in ICML 2026, 2026

We present SpaceVista, a unified framework for all-scale visual spatial reasoning from millimeter to kilometer scales.

Recommended citation: Peiwen Sun, Shiqiang Lang, Dongming Wu, Yi Ding, Kaituo Feng, Huadai Liu, Zhen Ye, Rui Liu, Yun-Hui Liu, Jianan Wang, Xiangyu Yue (2026). "SpaceVista: All-Scale Visual Spatial Reasoning from mm to km." ICML 2026.

talks

teaching

Teaching experience 1

Undergraduate course, University 1, Department, 2014

This is a description of a teaching experience. You can use markdown like any other post.

Teaching experience 2

Workshop, University 1, Department, 2015

This is a description of a teaching experience. You can use markdown like any other post.