Back to stories
Models

Alibaba Reveals It Built 'Happy Horse,' the Mystery AI Video Model That Topped Global Rankings

Michael Ouroumis2 min read
Alibaba Reveals It Built 'Happy Horse,' the Mystery AI Video Model That Topped Global Rankings

Alibaba Group has officially claimed ownership of HappyHorse-1.0, a video AI model that first appeared anonymously on the Artificial Analysis Video Arena and quickly rose to the top of global leaderboards — surprising the AI community and sending Alibaba shares up over 2% on the news.

The reveal ends weeks of speculation about who was behind the "mystery model" that dethroned ByteDance's Seedance 2 and other well-funded competitors in blind user voting.

From Anonymous Submission to Global #1

HappyHorse-1.0 debuted without any company branding on the Artificial Analysis Video Arena, where users evaluate AI-generated videos side by side without knowing which model produced them. It reached #1 globally in both Text-to-Video (Elo 1333) and Image-to-Video (Elo 1392) categories, an unprecedented achievement for an anonymous submission.

The model was developed by a team formerly operating under Alibaba's Taotian Group Future Life Laboratory, led by Zhang Di — the former Vice President of Kuaishou and technical architect behind Kling AI, one of China's most prominent video generation platforms.

What Makes HappyHorse Different

At 15 billion parameters, HappyHorse-1.0 is notably compact compared to many frontier models, yet it introduces several technical firsts. It is reportedly one of the first open-weight models that natively generates synchronized dialogue, ambient sounds, and effects alongside video — eliminating the need for separate audio pipelines.

Key capabilities include:

Architecture

The model is built on a 40-layer self-attention Transformer that breaks from the popular DiT (Diffusion Transformer) approach. Rather than using cross-attention for text conditioning, HappyHorse places text, image, video, and audio tokens into a single unified sequence, with attention handling all modality fusion natively. The first and last four layers manage modality-specific embedding and decoding, while the middle 32 layers share parameters across all modalities.

Open Source and Commercially Licensed

In a move that distinguishes it from many competitors, Alibaba has announced that HappyHorse-1.0 will be fully open source with complete commercial licensing. Model weights, distilled variants, super-resolution modules, and inference code are expected to be released on GitHub, though as of early April 2026 the weights have not yet been made publicly available. The API is reportedly scheduled to open for access on April 30.

Implications for the AI Video Market

The release intensifies an already crowded AI video generation landscape where OpenAI's Sora has struggled commercially, Google's Veo continues to iterate, and Chinese labs like ByteDance and Kuaishou have been rapidly gaining ground. HappyHorse's combination of top-tier quality, open weights, and efficient inference could reshape pricing expectations across the sector — particularly for startups building on proprietary video APIs that now face a powerful open-source alternative.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

xAI Launches Grok Voice Think Fast 1.0, Tops τ-Voice Bench and Powers Starlink Support
Models

xAI Launches Grok Voice Think Fast 1.0, Tops τ-Voice Bench and Powers Starlink Support

xAI's new voice model scored 67.3% on the τ-voice Bench — well ahead of Gemini 3.1 Flash Live and GPT Realtime — and is now powering Starlink's phone sales and support with a 70% autonomous resolution rate.

2 days ago2 min read
Tencent Drops Hy3 Preview: 295B Open-Source MoE Model Kicks DeepSeek Out of Yuanbao
Models

Tencent Drops Hy3 Preview: 295B Open-Source MoE Model Kicks DeepSeek Out of Yuanbao

Tencent has open-sourced Hy3 Preview, a 295B/21B-activated mixture-of-experts model built in under three months. The Yuanbao chatbot is switching its primary engine from DeepSeek to the new in-house model.

4 days ago2 min read
DeepSeek V4 Preview Lands: 1.6T-Parameter Open Model With 1M Context, Flash Pricing at $0.14/M
Models

DeepSeek V4 Preview Lands: 1.6T-Parameter Open Model With 1M Context, Flash Pricing at $0.14/M

DeepSeek on April 24 released preview versions of V4-Pro and V4-Flash, an open-weight MoE family with a 1M-token context window and pricing that undercuts Western frontier labs.

4 days ago2 min read