AI in Marketing

VOIX Framework Builds AI-Friendly Websites with Two New HTML Elements

Written by Writing Team | Nov 28, 2025 1:00:00 PM

AI agents trying to browse the web have a fundamental problem: Websites were designed for humans with eyes, not machines parsing semantic meaning. The result is brittle, slow, and insecure interactions where agents try to "see" websites like humans do—often failing spectacularly.

Researchers at TU Darmstadt just proposed a radically different approach. Instead of making AI agents smarter at interpreting human-oriented interfaces, make websites declare their capabilities explicitly in machine-readable format.

The VOIX framework introduces two new HTML elements—<tool> and <context>—that allow websites to expose available actions directly to AI agents. The performance difference is staggering: VOIX completes tasks in 0.91 to 14.38 seconds compared to 4.25 seconds to over 21 minutes for traditional vision-based AI browser agents.

In one benchmark, VOIX rotated a green triangle 90 degrees in one second. Perplexity's Comet agent needed ninety seconds for the identical task.

How VOIX Actually Works

The architecture is elegant in its simplicity. Websites use the <tool> element to list available actions by name, parameters, and description. The <context> element provides current application state information.

For a to-do list application, you'd include something like <tool name="add_task"> defining parameters like "title" and "priority," connected to the app's JavaScript logic. When an AI agent wants to add a task, it calls this tool directly rather than visually searching for input fields and submit buttons.

The framework divides responsibilities clearly:

  • The website declares its functions through structured HTML elements
  • A browser agent mediates between the site and AI, handling communication
  • The inference provider decides actions using this structured data rather than screenshot interpretation

This represents a fundamental inversion of current AI browsing approaches. Instead of agents inferring what's possible from visual inspection—a brittle, inefficient, and insecure process—websites explicitly announce their capabilities in formats agents can parse reliably.

The Speed and Security Advantages

The latency improvements stem from eliminating the entire visual interpretation pipeline. Traditional agents must:

  1. Capture screenshots of the webpage
  2. Process images through vision models to identify UI elements
  3. Guess which elements correspond to desired actions
  4. Simulate clicks or text input
  5. Capture new screenshots to verify results
  6. Repeat if the action failed

VOIX agents skip all that. They query available tools, select the appropriate one, pass parameters, and receive immediate confirmation. No vision models. No guessing. No verification loops.

The security benefits are equally significant. Vision-based agents are vulnerable to prompt injection attacks embedded in webpage content—malicious instructions disguised as UI elements that trick agents into unintended actions.

VOIX agents only see explicitly declared tools and context. Attackers can't inject commands through visual UI manipulation because agents aren't interpreting UI visually. They're calling structured functions.

Privacy improves too. The browser agent sends user conversations directly to the LLM provider, keeping the website out of the loop. Agents only access data explicitly released through <context> elements, not the entire rendered page. And because VOIX runs client-side, site owners don't pay for LLM inference costs.

The Hackathon Validation

To test real-world viability, the TU Darmstadt team ran a three-day hackathon with 16 developers. Six teams built different applications using VOIX, most with no prior framework experience.

The System Usability Scale score reached 72.34—above the industry average of 68, indicating strong usability even for developers encountering the framework for the first time.

The applications demonstrated VOIX's flexibility:

  • Graphic design tool: Users clicked objects and gave voice commands like "rotate this by 45 degrees"
  • Fitness planner: Generated full workout plans from prompts like "create a full week high-intensity training plan for my back and shoulders"
  • Soundscape creator: Changed audio environments based on commands like "make it sound like a rainforest"
  • Kanban board: Generated task lists from natural language prompts

These aren't toy demos. They're functional applications where AI agents manipulate application state through structured tool calls rather than simulating human interactions.

The Developer Mindset Shift Required

VOIX solves the AI agent problem but creates a new challenge for web developers: You need to think about your application as an API surface, not just a visual interface.

This means:

  • Defining specific agent actions rather than relying on UI to communicate possibilities
  • Deciding which tools to expose and at what level of granularity
  • Maintaining synchronization between VOIX declarations and actual UI functionality
  • Balancing basic functions (like "add task") with higher-level intents (like "organize my priorities for the week")

For large or legacy codebases, this is non-trivial work. VOIX declarations can fall out of sync with UI implementations. Developers must maintain parallel representations—one for human interaction, one for agent interaction.

The researchers acknowledge this tension but argue it's necessary. "Agents must infer affordances from human-oriented user interfaces, leading to brittle, inefficient, and insecure interactions," they write. Explicit declaration solves those problems at the cost of additional developer overhead.

The Web Standards Battle Ahead

VOIX arrives amid growing recognition that current web architecture doesn't serve AI agents well. Companies like OpenAI and Perplexity are betting on chatbots as the primary web gateway, with AI browsers like Atlas and Comet handling everything from travel booking to online shopping.

But language models struggle with modern website complexity, and prompt injection remains a persistent threat. Making agent-driven browsing practical may require new standards that present information in formats LLMs can reliably parse.

Initiatives like llms.txt (machine-readable website documentation) and the proliferation of Model Context Protocol (MCP) servers suggest the industry recognizes this need. VOIX positions itself as part of that next wave of web standards.

The framework's open nature matters strategically. The researchers built a Chrome extension with chat and voice support that works with any OpenAI-compatible API. It runs with both cloud-based and local LLMs, tested successfully with Qwen3-235B-A22B.

By making the framework freely available and demonstrating cross-platform compatibility, they're attempting to establish a standard before proprietary approaches fragment the ecosystem.

The Chicken-and-Egg Problem

VOIX faces the classic web standards adoption challenge: Websites won't implement new HTML elements until agents support them, and agents won't prioritize unsupported standards until websites adopt them.

Breaking this deadlock likely requires either:

  • Major platform adoption: If Chrome or Safari built VOIX support into browser AI features, developers would have incentive to implement it
  • Killer app demonstration: If a major site showed dramatically better agent experiences through VOIX, others would follow
  • Framework integration: If React, Vue, or Next.js provided VOIX components out-of-the-box, adoption would grow organically

The hackathon results suggest the developer experience is strong enough to support adoption if the ecosystem incentives align. The 72.34 usability score indicates developers can learn and implement VOIX relatively easily.

What This Means for the AI-First Web

If VOIX or similar standards gain traction, we're looking at a fundamental reorientation of web development. Websites wouldn't just be designed for visual consumption—they'd explicitly declare their programmatic capabilities for AI agents.

This has profound implications:

Search engines become action engines

Instead of returning links, AI search could directly execute tasks across websites using declared tools.

User interfaces become optional

For many workflows, users might interact entirely through chat or voice, never seeing traditional UI.

Cross-site workflows become trivial

Agents could chain actions across multiple sites using standardized tool declarations—book flight on one site, reserve hotel on another, schedule transportation on a third, all from one conversation.

Accessibility improves dramatically

Declaring functionality separately from visual presentation makes websites inherently more accessible to assistive technologies.

The Timing Question

The researchers note that language models still struggle with website complexity and that prompt injection remains a persistent threat. These are real barriers. But the trajectory is clear—AI agents interacting with websites is not a question of if, but when and how.

VOIX provides a "how" that's demonstrably faster, more secure, and more reliable than vision-based approaches. Whether it becomes the standard or inspires competing approaches, the core insight stands: Websites designed for human vision alone won't serve the AI-first web adequately.

Developers who start thinking about their applications as both visual interfaces and programmatic tool surfaces will be ahead of the curve. Those who wait for agents to get better at "seeing" websites the way humans do are betting against the direction the entire industry is moving.

The web is about to get a lot more machine-readable. VOIX is one compelling vision for how that happens.

If your organization is planning for AI-agent interactions and needs guidance on preparing web properties for programmatic access while maintaining human-centric design, Winsome Marketing's team can help you architect for both audiences simultaneously.