<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Hao AI Lab @ UCSD</title><link>https://haoailab.com/</link><description>Recent content on Hao AI Lab @ UCSD</description><image><title>Hao AI Lab @ UCSD</title><url>https://haoailab.com/img/HAOAILAB_trimmed.jpg</url><link>https://haoailab.com/img/HAOAILAB_trimmed.jpg</link></image><generator>Hugo -- gohugo.io</generator><language>en-us</language><lastBuildDate>Sun, 15 Mar 2026 12:00:00 -0800</lastBuildDate><atom:link href="https://haoailab.com/index.xml" rel="self" type="application/rss+xml"/><item><title>Into the Dreamverse: Vibe Directing in FastVideo</title><link>https://haoailab.com/blogs/dreamverse/</link><pubDate>Sun, 15 Mar 2026 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/dreamverse/</guid><description>TL;DR: Our new real-time inference stack in FastVideo enables Dreamverse, a prototype for a new interface where users can vibe direct their own “multiverse” of videos.
AI video generation is already good enough to make a convincing clip. But real creative work is not about getting a clip in one shot. It’s about iteration. An idea appears, you test it: keep the subject, change the camera angle, continue the scene, and try again.</description></item><item><title>Create a 5s 1080p Video in 4.5s with FastVideo on a Single GPU</title><link>https://haoailab.com/blogs/fastvideo_realtime_1080p/</link><pubDate>Wed, 11 Mar 2026 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/fastvideo_realtime_1080p/</guid><description>TL;DR: If you work in media generation, you know the frustration: an idea pops into your mind, you type a prompt, maybe provide a reference image, and want to see the result immediately – while the idea is still alive. But existing video generation APIs break the loop. You wait minutes for clips, and each costs enough to make your wallet wince. FastVideo&amp;rsquo;s real-time inference stack fixes this problem.</description></item><item><title>From Physical Commonsense to Scientific Reasoning: Why World Modeling in Video Matters</title><link>https://haoailab.com/blogs/videoscience/</link><pubDate>Thu, 12 Feb 2026 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/videoscience/</guid><description>TL;DR: The golden age of AI video has mastered the &amp;ldquo;look&amp;rdquo; of reality, but it has yet to learn the laws of reality. Without adhering to rigorous scientific principles, even the most photorealistic model remains a high-fidelity hallucination engine rather than a reliable world simulator. To bridge this gap, we introduce VideoScience-Bench: the first benchmark specifically designed to move beyond &amp;ldquo;physical commonsense&amp;rdquo; and evaluate undergraduate-level scientific reasoning in video models.</description></item><item><title>DistCA</title><link>https://haoailab.com/summary/distca/</link><pubDate>Sun, 21 Dec 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/distca/</guid><description>Core Attention Disaggregation for Efficient Long-context Language Model Training</description></item><item><title>CAD: Disaggregating Core Attention for Efficient Long-context Language Model Training</title><link>https://haoailab.com/blogs/distca/</link><pubDate>Wed, 17 Dec 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/distca/</guid><description>TL;DR: Workload imbalance is one of the major problems in training long-context LLM models. Imbalance among data parallel (DP) and pipeline parallel (PP) workers introduces stragglers or bubbles that causes severe slowdown, and the problem becomes more severe as we scale to longer context lengths or more GPUs.
We believe that one of the major reasons for this slowdown is that the core attention, i.e., the $\text{softmax}(QK^T)V$ kernel, colocates with the other linear parts.</description></item><item><title>Fast and Accurate Causal Parallel Decoding using Jacobi Forcing</title><link>https://haoailab.com/blogs/jacobi-forcing/</link><pubDate>Tue, 16 Dec 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/jacobi-forcing/</guid><description>TL;DR: Today’s best LLMs mostly decode autoregressively from left-to-right, which gives great quality but is terribly slow. Diffusion LLM can decode many tokens in parallel thanks to their non-causal, any-order generation, but they must be trained from scratch, or expensively adapted from autoregressive (AR) checkpoints with a mismatched, non-causal diffusion objective; we find this mismatch often hurts quality and breaks many effective KV-cache related serving optimizations. This blog introduces Jacobi Forcing, a new training technique that converts LLMs into native causal parallel decoders.</description></item><item><title>JacobiForcing</title><link>https://haoailab.com/summary/jacobi-forcing/</link><pubDate>Tue, 16 Dec 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/jacobi-forcing/</guid><description>Fast and Accurate Causal Parallel Decoding using Jacobi Forcing</description></item><item><title>AUP: when Accuracy Meets Parallelism in Diffusion Language Models</title><link>https://haoailab.com/blogs/text-diffusion/</link><pubDate>Wed, 10 Dec 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/text-diffusion/</guid><description>TL;DR: Diffusion large language models (dLLMs) promise things that autoregressive LLMs cannot: parallel decoding, error correction, and random-order generation. Over the past year, a wave of papers has pushed this vision, and closed-source systems like Gemini Diffusion and Mercury report impressive throughput numbers. In this blog, we take a step back and ask a simple question: if we look at both speed and accuracy together, are diffusion LLMs actually better decoders than strong autoregressive (AR) models?</description></item><item><title>d3LLM</title><link>https://haoailab.com/summary/d3llm/</link><pubDate>Wed, 10 Dec 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/d3llm/</guid><description>Ultra-Fast Diffusion LLM 🚀</description></item><item><title>CausalWan-MoE Preview: Applying Self-Forcing Distillation To Wan2.2</title><link>https://haoailab.com/blogs/fastvideo_causalwan_preview/</link><pubDate>Tue, 18 Nov 2025 11:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/fastvideo_causalwan_preview/</guid><description>TL;DR: The FastVideo Team is excited to share some of our progress on distilling the Wan2.2-A14B model into an autoregressive architecture, alongside the release of our preview checkpoint for CausalWan2.2-I2V-A14B. In this blog, we’ll first discuss the new MoE architecture behind the open-source SOTA performance of Wan2.2, the differences between bidirectional and autoregressive video models, and then share some of the challenges we encountered when applying Self-Forcing distillation to this architecture.</description></item><item><title>Disaggregated Inference: 18 Months Later</title><link>https://haoailab.com/blogs/distserve-retro/</link><pubDate>Mon, 03 Nov 2025 00:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/distserve-retro/</guid><description>Eighteen months ago, our lab introduced DistServe with a simple bet: split LLM inference into prefill and decode, and scale them independently on separate compute pools. Today, almost every production-grade LLM serving framework – NVIDIA Dynamo, llm-d, Ray Serve LLM SGLang, vLLM, LMCache, MoonCake – runs on disaggregation and demonstrates its power in large-scale, real-world LLM serving workloads, with many more continuing to push its boundaries. Concepts like TTFT (time-to-first-token) and TPOT (time-per-output-token), now standard latency metrics in nearly every serving benchmark, were also popularized through the lens of disaggregation.</description></item><item><title>Scaling Speculative Decoding with Lookahead Reasoning</title><link>https://haoailab.com/blogs/lookaheadreasoning/</link><pubDate>Mon, 22 Sep 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/lookaheadreasoning/</guid><description>TL;DR: We propose Lookahead Reasoning (LR), a technique that significantly accelerates large reasoning models(LRMs) and complements existing speculative decoding methods. Traditional token-level speculative decoding suffers from limited gains because the probability of correctly guessing a long sequence decreases exponentially with length. In contrast, LR operates at the step level, proposing future reasoning steps instead of individual tokens. This is much more effective since a proposed step only needs to be semantically correct, rather than matching exactly word for word.</description></item><item><title>Can RL-based LLM post-training on games generalize to other tasks? (GRL)</title><link>https://haoailab.com/blogs/grl/</link><pubDate>Wed, 27 Aug 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/grl/</guid><description/></item><item><title>A Practical Guideline to Using Lmgame-Bench</title><link>https://haoailab.com/blogs/lmgame-bench-use/</link><pubDate>Thu, 21 Aug 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/lmgame-bench-use/</guid><description/></item><item><title>FastWan: Generating a 5-Second Video in 5 Seconds via Sparse Distillation</title><link>https://haoailab.com/blogs/fastvideo_post_training/</link><pubDate>Mon, 04 Aug 2025 11:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/fastvideo_post_training/</guid><description>TL;DR: We introduce FastWan, a family of video generation models trained via a new recipe we term as “sparse distillation”. Powered by FastVideo, FastWan2.1-1.3B end2end generates a 5-second 480P video in 5 seconds (denoising time 1 second) on a single H200 and 21 seconds (denoising time 2.8 seconds) on a single RTX 4090. FastWan2.2-5B generates a 5-second 720P video in 16 seconds on a single H200. All resources — model weights, training recipe, and dataset — are released under the Apache-2.</description></item><item><title>From Pokémon Red to Standardized Game-as-an-Eval</title><link>https://haoailab.com/blogs/lmgame-bench/</link><pubDate>Fri, 20 Jun 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/lmgame-bench/</guid><description/></item><item><title>FastVideo V1: A Unified Framework for Accelerated Video Generation</title><link>https://haoailab.com/blogs/fastvideo/</link><pubDate>Thu, 24 Apr 2025 11:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/fastvideo/</guid><description>TL;DR: We are announcing FastVideo V1, a unified framework that accelerates video generation. This new version features a clean, consistent API that works across popular video models, making it easier for developers to author new models and incorporate system - or kernel - level optimizations. For example, FastVideo V1 provides 3x speedup for inference while maintaining quality by seamlessly integrating SageAttention and Teacache.
What&amp;rsquo;s New Modern open-source video generation models such as HunyuanVideo and Wan2.</description></item><item><title>ReFoRCE: A Text-to-SQL Agent with Self-Refinement, Format Restriction, and Column Exploration</title><link>https://haoailab.com/blogs/reforce/</link><pubDate>Thu, 10 Apr 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/reforce/</guid><description>TL;DR: We present ReFoRCE, a Text-to-SQL agent that leads the Spider 2.0 leaderboard—the most challenging Text-to-SQL benchmark where even advanced models like GPT-4o score around 10%. ReFoRCE tackles real-world deployment issues such as massive schemas, SQL dialect diversity, and complex queries. It uses table compression to handle long contexts, format restriction for accurate SQL generation, and iterative column exploration for better schema understanding. A self-refinement pipeline with self-consistency and parallel voting further boosts performance, achieving state-of-the-art scores of 31.</description></item><item><title>Fast Video Generation with Sliding Tile Attention</title><link>https://haoailab.com/blogs/sta/</link><pubDate>Tue, 18 Feb 2025 11:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/sta/</guid><description>TL;DR: Video generation with DiTs is painfully slow &amp;ndash; HunyuanVideo takes 16 minutes to generate just a 5-second video on an H100 with FlashAttention-3. Our sliding tile attention (STA) slashes this to 5 minutes with zero quality loss, no extra training required. Specifically, STA accelerates attention alone by 2.8–17x over FlashAttention-2 and 1.6–10x over FlashAttention-3. With STA and other optimizations, our solution boosts end-to-end generation speed by 2.98× compared to the FA3 full attention baseline, without quality loss or the need for training.</description></item><item><title>FastVideo</title><link>https://haoailab.com/summary/sta/</link><pubDate>Tue, 18 Feb 2025 11:00:00 -0800</pubDate><guid>https://haoailab.com/summary/sta/</guid><description>Make Video Generation Faster</description></item><item><title>Dynasor</title><link>https://haoailab.com/summary/dynasor-cot/</link><pubDate>Sun, 16 Feb 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/dynasor-cot/</guid><description>Making Reasoning Models More Token-Efficient</description></item><item><title>Dynasor: More Efficient Chain-of-Thought Through Certainty Probing</title><link>https://haoailab.com/blogs/dynasor-cot/</link><pubDate>Sun, 16 Feb 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/dynasor-cot/</guid><description>TL;DR: We observe reasoning models often exhibit poor token efficiency: they waste many tokens second-guessing themselves. We develop Dynasor-CoT, a certainty-based approach for dynamically allocating inference compute for reasoning models. The intuition is that by probing reasoning models at intermediate steps, we can identify and early terminate problems where they maintain consistently high certainty in their answers. The method is plug-and-play, requiring no model modifications or training, but matches baseline accuracy on benchmarks like AMC23, AIME24, and MATH500 while reducing token consumption by 29% dataset-wide and up to 81% for single problems.</description></item><item><title>GameArena: Evaluating LLM Reasoning through Live Computer Games</title><link>https://haoailab.com/blogs/gamearena/</link><pubDate>Mon, 10 Feb 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/gamearena/</guid><description/></item><item><title>LMGame Bench</title><link>https://haoailab.com/summary/gamearena/</link><pubDate>Mon, 10 Feb 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/gamearena/</guid><description>Evaluating LLM Reasoning through Live Computer Games</description></item><item><title>Efficient LLM Scheduling by Learning to Rank</title><link>https://haoailab.com/blogs/vllm-ltr/</link><pubDate>Mon, 13 Jan 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/vllm-ltr/</guid><description>TL;DR: Traditional Large Language Model (LLM) serving systems rely on first-come-first-serve (FCFS) scheduling. When longer requests block shorter ones in the queue, this creates a cascade of delays that severely impacts overall system latency. LLM inference jobs are particularly challenging to schedule due to their highly unpredictable workload and variable output lengths. We developed a novel learning to rank approach that predicts the relative ranking of output lengths, enabling a more efficient Shortest Job First-like scheduling policy.</description></item><item><title>vLLM-LTR</title><link>https://haoailab.com/summary/vllm-ltr/</link><pubDate>Mon, 13 Jan 2025 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/vllm-ltr/</guid><description>Efficient LLM Scheduling by Learning to Rank</description></item><item><title>MuxServe</title><link>https://haoailab.com/summary/muxserve/</link><pubDate>Mon, 20 May 2024 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/muxserve/</guid><description>Serving Multiple LLMs with Flexible Spatial-Temporal Multiplexing</description></item><item><title>MuxServe: Flexible Spatial-Temporal Multiplexing for Multiple LLM Serving</title><link>https://haoailab.com/blogs/muxserve/</link><pubDate>Mon, 20 May 2024 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/muxserve/</guid><description>TL;DR: Efficiently serving multiple LLMs have emerged as a crucial and time-sensitive demand within the community, especially for LLM endpoint providers. In this blog, we show that the dynamic popularity of LLMs and the unbalanced resource utilization of LLM inference can be leveraged to achieve high GPU utilization and reduce serving cost. We introduce MuxServe, a novel serving system that efficiently serves multiple LLMs with flexible spatial-temporal multiplexing. MuxServe outperforms the spatial partitioning and temporal multiplexing baselines by up to $1.</description></item><item><title>CLLM</title><link>https://haoailab.com/summary/cllm/</link><pubDate>Mon, 06 May 2024 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/cllm/</guid><description>Consistency Large Language Models: A Family of Efficient Parallel Decoders</description></item><item><title>Consistency Large Language Models: A Family of Efficient Parallel Decoders</title><link>https://haoailab.com/blogs/cllm/</link><pubDate>Mon, 06 May 2024 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/cllm/</guid><description>TL;DR: LLMs have been traditionally regarded as sequential decoders, decoding one token after another. In this blog, we show pretrained LLMs can be easily taught to operate as efficient parallel decoders. We introduce Consistency Large Language Models (CLLMs), a new family of parallel decoders capable of reducing inference latency by efficiently decoding an $n$-token sequence per inference step. Our research shows this process &amp;ndash; mimicking human cognitive process of forming complete sentences in mind before articulating word by word &amp;ndash; can be effectively learned by simply finetuning pretrained LLMs.</description></item><item><title>DistServe</title><link>https://haoailab.com/summary/distserve/</link><pubDate>Sun, 17 Mar 2024 12:00:00 -0800</pubDate><guid>https://haoailab.com/summary/distserve/</guid><description>Maximizing Goodput in LLM Serving using Prefill-Decode Disaggregation</description></item><item><title>Throughput is Not All You Need: Maximizing Goodput in LLM Serving using Prefill-Decode Disaggregation</title><link>https://haoailab.com/blogs/distserve/</link><pubDate>Sun, 17 Mar 2024 12:00:00 -0800</pubDate><guid>https://haoailab.com/blogs/distserve/</guid><description>TL;DR: LLM apps today have diverse latency requirements. For example, a chatbot may require a fast initial response (e.g., under 0.2 seconds) but moderate speed in decoding which only needs to match human reading speed, whereas code completion requires a fast end-to-end generation time for real-time code suggestions.
In this blog post, we show existing serving systems that optimize throughput are not optimal under latency criteria. We advocate using goodput, the number of completed requests per second adhering to the Service Level Objectives (SLOs), as an improved measure of LLM serving performance to account for both cost and user satisfaction.</description></item><item><title/><link>https://haoailab.com/contact/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://haoailab.com/contact/</guid><description>contact</description></item><item><title/><link>https://haoailab.com/home/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://haoailab.com/home/</guid><description>home page for Hao Lab @ UCSD</description></item><item><title/><link>https://haoailab.com/math-examples/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://haoailab.com/math-examples/</guid><description>This is an inline (a^=x-b^) equation.
This is an inline $a^*=x-b^*$ equation.
These are block equations:
\[a^*=x-b^*\] \[ a^*=x-b^* \] \[ a^*=x-b^* \] These are block equations using alternate delimiters:
$$a^*=x-b^*$$ $$ a^*=x-b^* $$ $$ a^*=x-b^* $$</description></item><item><title/><link>https://haoailab.com/people/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://haoailab.com/people/</guid><description>people</description></item><item><title/><link>https://haoailab.com/publications/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://haoailab.com/publications/</guid><description>publications</description></item><item><title>Dynasor-CoT Demo</title><link>https://haoailab.com/demo/dynasor-cot/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://haoailab.com/demo/dynasor-cot/</guid><description/></item></channel></rss>