Two new anonymous AI models, Horizon Alpha and Horizon Beta, appeared on Open Router and have demonstrated impressive capabilities, especially in UI and design tasks.
Horizon Alpha consistently outperformed Cloud4 Opus, a leading state-of-the-art model, in UI generation tests.
Both Horizon models produced visually superior and tasteful UI mockups even when compared to expensive, top-tier models.
In UI/styling and SVG generation, Horizon outperforms Kimmy K2, Cloud Force Sonnet, and even Cloud4 Opus on subjective and visual quality.
When tested on trivia (like skateboarding history), Horizon provides solid answers, but its accuracy and depth place it slightly below GPT-4's capabilities.
On the custom "Skate Bench" benchmark, Horizon Alpha and Beta score around 20%, between Gro 3 Mini and GPT40.
In programming tool usage and planning, Horizon models eagerly utilize to-do list features and tool integrations early in their workflows.
Horizon models prefer to plan tasks actively; they often start with a to-do list and describe their process, a novel behavior in this context.
Their average generation speed is around 110 tokens per second.
Despite excellent subjective performance, Horizon models do not score highly on standard benchmarks, particularly in math, where they underperform relative to models like Llama 4 Maverick.
Horizon Beta is currently free and yields results surpassing even hand-crafted designs for AI image generation studios.
Users should experiment with Horizon soon, as free access could end abruptly.
The underlying providers are likely collecting user interaction data to refine their models.
The easiest public way to test Horizon’s capabilities is via T3 Chat or other tools supporting Open Router, though some integration workarounds may be required.
The creator is genuinely impressed with Horizon’s quality, recommends quick exploration, and anticipates further industry developments and revelations about these anonymous models.