Back to stories
Industry

Toyota unveils Woven City AI Vision Engine and integrated ANZEN system

Michael Ouroumis3 min read
Toyota unveils Woven City AI Vision Engine and integrated ANZEN system

Toyota and its technology subsidiary Woven by Toyota used a joint announcement on April 22, 2026 to pull back the curtain on a sweeping set of artificial intelligence systems designed to run their experimental Woven City in Susono, Japan as a single coordinated environment. The unveiling moves the long-hyped "living laboratory" decisively into the agentic AI era, with a city-scale foundation model at its core.

A city-scale vision language model

The headline release is the Woven City AI Vision Engine, which Toyota describes as a large-scale AI foundation model that enables the city to understand and respond to real-world conditions in real time. Toyota says the underlying vision language model ranks among the world's leading systems according to the MVBench Leaderboard, a benchmark that measures temporal and spatial understanding in video.

By positioning a vision language model as municipal infrastructure, Toyota is making a bet that physical environments — streets, facilities, mobility fleets — should be perceived and reasoned about by a single multimodal backbone rather than a patchwork of narrow perception models. It is a notably different architecture from the sensor-and-rules smart-city deployments that dominated the last decade.

Integrated ANZEN: behaviour AI meets driving assistance

Alongside the Vision Engine, Toyota introduced what it calls the Integrated ANZEN System. ANZEN fuses Woven City Behavior AI, which interprets human behavioural patterns, with Woven City Drive Sync Assist, which provides driving assistance. Toyota frames the combination as letting "people, mobility technologies and infrastructure operate as a single coordinated system."

That phrasing is important. Rather than treating pedestrian-prediction and driver-assistance as separate stacks — as most automakers do — Toyota is wiring them into one control loop. If a human is about to step off a curb, the same model informing that prediction informs the vehicle that should yield.

Supporting data backbone

The announcement also detailed two infrastructure layers that feed the models: the Woven City Infra Hub, an integrated data platform, and the Woven City Data Fabric, a data management framework. Together they are intended to give the Vision Engine and ANZEN a consistent, real-time picture of the city.

Inventor Garage and a growing roster

The Inventor Garage, a development hub with prototyping spaces and testing areas, has begun operations. Four new companies joined the program as Inventors, bringing the total to 24. Toyota summarises the approach with the Japanese concept of "Kakezan," or multiplication — a bet that cross-industry collaboration, not any single partner, will produce the breakthroughs worth productising.

Chairman Akio Toyoda also contributed to the development of an "Akio Toyoda AI," which Toyota says reflects his leadership approach.

Why it matters

For the broader industry, Toyota's announcement is a test case for two ideas that have been circulating but rarely implemented at scale. First, that the right abstraction for a smart city is a foundation model, not a collection of services. Second, that driver-assistance systems should inherit their understanding of humans from the same model that understands the wider environment.

If Woven City can operationalise those ideas, the Vision Engine and ANZEN may quietly become the clearest demonstration yet of what "physical AI" looks like outside a research paper — with a Japanese automaker, rather than a Silicon Valley lab, holding the template.

Learn AI for Free — FreeAcademy.ai

Take "AI for Business: Practical Implementation" — a free course with certificate to master the skills behind this story.

More in Industry

Meta Will Record Employees' Keystrokes and Screens to Train AI Agents
Industry

Meta Will Record Employees' Keystrokes and Screens to Train AI Agents

Meta is rolling out a Model Capability Initiative that captures US staff mouse movements, keystrokes, and occasional screenshots to train AI agents that can do white-collar work autonomously.

5 min ago3 min read
Alibaba Unveils Qwen Xiaojiuwo, an Ecosystem-Level AI Digital Human, in Pivot to Embodied AI
Industry

Alibaba Unveils Qwen Xiaojiuwo, an Ecosystem-Level AI Digital Human, in Pivot to Embodied AI

Alibaba officially launched Qwen Xiaojiuwo, a digital-human AI assistant that can order food, book rides, and buy tickets through conversation, signaling a strategic push from chat models into embodied, physical-interaction AI.

3 hours ago3 min read
Horizon Robotics Unveils Xingkong, China's First Cockpit-Driving Fusion Chip
Industry

Horizon Robotics Unveils Xingkong, China's First Cockpit-Driving Fusion Chip

Horizon Robotics launched Xingkong (codenamed 'Stellar') today at the Smart EV Development Forum, fusing cockpit and autonomous-driving compute onto a single chip and cutting per-vehicle costs by 1,500-4,000 yuan.

13 hours ago2 min read