Company work · Tatsu Works Pte. Ltd. Sensitive metrics and internal deliverables are available upon request in an interview setting.

Design SystemsAI UXTokenizationAgent-Oriented Design

Design System at Scale (AI-First)

Rethinking design systems for an AI-driven future: from designer libraries to machine-readable systems with enforced constraints

Role: Solo exploration (Design + System Thinking)Collaboration: Knowledge-sharing within a 2-person design teamDuration: Ongoing (2026–Present)Tools: Figma, Pencil.dev, Claude, Ant Design (reference)

Problem

Traditional design systems were built for humans, not machines. AI agents struggle to reliably interpret Figma-based systems because component structures aren't machine-readable, spatial relationships are inferred visually, and styling lacks code-aligned definitions.

Outcome

Established an AI-compatible design pipeline where tokens enforce design rules, components are logically reproducible, and AI agents generate UI while maintaining consistency, quality, and brand control.

Trailbound Design System combining CocosCreator prefabs and Figma design system components

What This Is

An ongoing design system migration, transitioning from Figma-only component libraries to a web-based, token-driven system that both humans and AI agents can consume reliably. The focus is on establishing the pipeline and validating the approach.

What This Is Not

  • Not a shipped product redesign
  • Not a replacement for Figma in the design workflow
  • Not a one-time migration. The system evolves with the product

Context

As a company, building for multi-platform products has always been a key objective. At the same time, the rise of AI agents introduced a new challenge: How do we ensure consistency not just across platforms, but also across human-designed and AI-generated interfaces?

My Role

While this initiative is discussed and aligned within a 2-person design team, I drive the exploration and execution independently, covering design, system thinking, and the AI translation pipeline.

Designed For

Non-designers

Anyone who wants to prototype fast while staying within the established art style

UIUX designers

Avoiding human errors in component usage, UX writing, and any standardization that used to be manual and error-prone

AI agents

Consuming components deterministically without creating conflicting duplicates

Problem Signals

Component ambiguity

Structures not always machine-readable

Implicit spatial rules

Padding & alignment inferred visually, not explicitly

Styling mismatch

Layer effects lack code-aligned definitions

AI Agent Failures

  • Inaccurate layout reconstruction
  • Misinterpreted component hierarchy
  • Inconsistent styling rule application
  • Even with Claude + Figma MCP integration

Key Insight

A design system that cannot be reliably interpreted by AI becomes a bottleneck in an AI-first workflow.

Hypothesis

If a design system is

  • Web-based
  • Token-driven
  • Structured like code

Then AI agents can

  • Read components deterministically
  • Reconstruct layouts accurately
  • Generate features within design rules

Foundation

From canvas to system

Instead of relying solely on Figma as the source of truth, I began transitioning the design system into a web-based format using Ant Design as a structural reference. I started with tokens, then encoded design rules directly into the system.

1

Web-Based System

  • Components as documented, inspectable entities
  • Rules explicit rather than implied
  • AI agents consume the system reliably
2

Tokenization First

Tokens form the “grammar” of the system: the most portable, machine-readable layer that both humans and AI understand.

  • Colors
  • Spacing
  • Typography
  • Border radius / elevation
3

Encoded Design Rules

The 4-point grid system (all spacing, sizing, and alignment in multiples of 4) is now enforced by tokens, not by memory.

  • Spacing tokens enforce 4-point increments
  • Components built only using valid values
  • AI agents inherit constraints automatically

What Changed

4-Point Grid System

Previously

  • Applied manually by designers
  • Checked manually within Figma files
  • Human error and inconsistency across teams

Now

  • Spacing tokens enforce 4-point increments
  • Components built only using valid values
  • AI agents inherit constraints automatically

Component Definitions

Previously

  • Visually consistent but structurally ambiguous
  • Layer effects inferred from canvas inspection
  • Variants documented separately from implementation

Now

  • Components are logically reproducible, not just visually consistent
  • Layout rules made explicit in code
  • Variants and states are standardized and machine-readable

Insight

Design systems are evolving from “libraries for designers” to “APIs with enforced constraints for AI agents.” The system no longer suggests good design practices. It enforces them by default.

Pipeline

Bridging design tools and AI workflows

To bridge the gap between traditional design tools and AI workflows, I introduced a translation pipeline, then used it to reconstruct components with tokens, explicit layout rules, and standardized variants.

Figma

Source of existing designs

Pencil.dev

AI-first canvas with structural translation

Claude

Interprets and reconstructs components

Reading from Figma

Getting Claude to accurately read and reconstruct components from Figma was a known challenge. Spatial relationships were inferred, component hierarchy was ambiguous, and styling details were frequently lost in translation.

Reading from Pencil.dev

Getting Claude to read the same components from Pencil.dev was a breeze. The structured output meant AI agents could interpret components deterministically: no guessing, no ambiguity, no reconstruction errors.

Insight

Pencil.dev acts as a “lingua franca” between visual design and AI reasoning. The goal is to make components not just visually consistent, but logically reproducible.

Web-Based Design System: Figma & CocosCreator Integration

This project is built with CocosCreator, not a standard web app. For AI agents to build features reliably, the design system has to mirror Figma's design components and map to CocosCreator's existing prefabs. This way, agents know which scripts and components already exist, and won't create conflicting duplicates.

The web version also shows live component previews with UI motions. Figma can't do this without entering present mode, so having previews here gives designers and non-designers a much clearer picture of how each component actually works.

Reducing Onboarding Friction

One-click MCP setup for AI agents

A design system is only useful if people (and AI agents) can actually start using it. To close the gap between landing on the site and having a working AI workflow, I used Claude to build a one-time MCP setup button placed in the top navigation.

The Problem

New users arriving at the ported design system website had no clear path to connect AI agents with the system. MCP configuration was manual, technical, and easy to get wrong, creating a barrier before anyone could experience the value of an AI-compatible design system.

The Solution

A prominent CTA in the top navigation, visible the moment users land, that handles the entire MCP setup in a single click. Built with Claude, it eliminates the technical setup barrier and lets users go from “first visit” to “AI agent connected” immediately.

Why Top Navigation

Placing the MCP setup CTA in the top navigation was a deliberate design decision. For a design system that claims to be AI-first, the ability to connect an AI agent shouldn't be buried in documentation. It should be the first thing a new user sees.

State-Aware Button

The button detects whether MCP has already been configured. On return visits, it transforms into a checkmark state, reassuring users that setup is complete without requiring them to remember. This prevents confusion and removes the cognitive overhead of wondering “did I already do this?”

First visit

Design system top navigation showing the Setup MCP button for new users

New users see a clear call-to-action to connect their AI agent in one click.

Return visit

Design system top navigation showing MCP Setup Done checkmark for returning users

The site detects the MCP is already configured and shows a checkmark, so users never have to wonder.

Insight

If the design system is built for AI agents, then onboarding AI agents should be as easy as onboarding a human. One button, zero configuration, immediate value. On the next visit, clear confirmation that you're already set up.

Current Status

Where the system stands today

40

Components Ready

Each component carries not just visuals but Cocos Creator metadata: which prefab it maps to and which script is attached. This prevents AI agents from creating conflicting duplicates of components that already exist as scripted prefabs in the engine.

Validated

Pipeline Tested

The system has been tested end-to-end by generating a feedback form popup modal, proving that AI agents can compose working UI from the design system while respecting component rules and layout constraints.

1-click

MCP Setup

AI agent connection time reduced from ~15 minutes of manual configuration to a single button click in the top navigation, with state detection on return visits.

Outcome

A portable, AI-compatible design system approach

What Was Established

AI-compatible design pipeline

Figma → Pencil.dev → Claude. Component reading went from unreliable to deterministic

System-enforced constraints

4-point grid and component rules encoded into tokens, eliminating manual review overhead

Cocos-aware component library

40 components carrying prefab and script metadata to prevent AI-generated duplicates

MCP setup time: ~15 min → 1 click

One-time setup button in top navigation with state-aware return visit detection

Cross-team adoption

Framework shared with 1 additional designer to replicate and extend into website use cases

Adapting Across Contexts

Not all interfaces behave the same. Discord embed game interfaces are highly constrained and interaction-driven, while web/enterprise interfaces follow structured, scalable patterns. Rather than duplicating components, the system ports logic across contexts:

  • Discord-based system used as foundation and proof of concept
  • Principles (tokens, rules, constraints) translated to structured environments
  • Learnings shared to extend into website use cases
  • System logic is portable; components are context-specific

Reflection

This project is not just about rebuilding a design system. It's about redefining how design systems behave across platforms, contexts, and now, creators: both humans and AI.

Even within a small team, the shift is clear: design is moving from creating outputs to defining systems that others (and AI) can build upon. The challenge isn't just making things look right. It's making rules that ensure things are built right, regardless of who or what is building.

What surprised me most was how much clarity came from encoding previously implicit rules. The 4-point grid was always “a best practice,” but the moment it became a system constraint, it stopped being a thing designers had to remember and became something the system guaranteed. That shift in responsibility, from human discipline to system enforcement, changed how I think about design system value.

And sometimes, that shift starts quietly. One system, one rule, one translation layer at a time.

As this work continues, the next milestone is to scale beyond foundational components into more complex UI patterns like forms, navigation systems, and multi-state interactions, and validate how well AI agents can assemble these reliably across both Discord and web-based product surfaces.