INGRAMMAO

Bluegg

Year: 2025–2026Duration: 5 months

Discipline:
Design
Dev
UI/UX
Product Management
Area:
Web
Agents / AI
Prototype

Bluegg is agentic canvas for creatives to think out loud. Watch AI agents work with your ideas through sketches and conversations, shaping them into mind maps, mood boards, storyboards, briefs and more. While most AI tools accelerate execution, Bluegg tackles the upstream problem in every creative project: defining the vision in the first place.

Note: Bluegg was developed independently between 2025 and early 2026, before the release of Google Stitch and Luma Agent.
Bluegg — agentic canvas for creative vision
0:00 / 0:00

Concept Trailer, created by Ingram Mao

The Problem Bluegg Solves

While AI is good at rapid execution, the true bottleneck in modern creative workflows is defining a strong initial vision. Most AI tools push that work into chat interfaces, which flatten the open, associative thinking that early-stage ideation actually needs.

Bluegg bridges this gap by providing an intuitive environment where human ideas, taste and AI capabilities can collaboratively shape creative direction.

0%

of creatives believe Gen AI tools dulls creativity1

0%

of creatives believe AI is making creative work soulless2

% of US adults who say increased use of AI in society will impact people’s ability to Think creatively3

0
worse
0
better
0
Neither
0
Unsure
  1. 1.dezeen.com — AI dulls creativity, design survey
  2. 2.forbes.com — AI is making us less creative, but there's a fix
  3. 3.pewresearch.org — How Americans view AI and its impact on people and society

Bluegg’s Product Thesis

AI should adapt to human creative workflows, not the other way around. Bluegg’s design mirrors the dynamic of a real brainstorming session, integrating voice AI with a spatial canvas—voice for thinking out loud, the canvas for ideas to take shape on. The result is a collaborative partner that amplifies human creativity instead of shallow automation.

Building the Concept Prototype

Product Management
Design
Dev

For the initial prototype, I integrated the Tldraw SDK and LiveKit with Gemini Live, merging voice and visual modalities into a single interface.

Testing this prototype revealed a key product insight: for Bluegg to feel like a collaborator, the AI needed bidirectional awareness of the canvas. In other words, it needs the ability to both read its current state and edit it during real-time conversation.

Initial UX of the Bluegg prototype

Product Positioning

Product Management

Most AI creative tools, even canvas-native ones like Figma Make or Flora, are still fundamentally text-prompt interfaces that assume the user already knows what they want. That assumption breaks down at the earliest, fuzziest stage of a project, which is where creative work needs most help.

I framed Bluegg’s opportunity as an interface problem, not a model problem: the underlying capabilities exist, but the interaction model for upstream ideation hasn’t been designed yet.

Figma Make interface
Flora interface

Building the Agent for Bluegg

Design
Dev

My design for the core architecture separates AI interaction from visual execution. A primary Voice Agent acts as the user’s main point of contact, equipped with tools to continuously read and interpret the canvas. To handle the complexity of visual manipulation, this agent routes all visual commands to a specialized Canvas Editing Agent for stable and precise execution.

Bluegg agent graph — voice agent routing to a canvas editing agent

Simplifying the User Flow

Design
UI/UX

Early iterations leaned on too many controls, which fought the calm, conversational feel I wanted. I pared the interface down to two input modalities: voice for anything you’d naturally say out loud, and a minimal set of buttons to edit the canvas.

Voice and canvas controls live in distinct, non-overlapping zones. Through user testing, the minimalistic interface was something people particularly appreciated.

Final Bluegg interface — voice agent meeting a spatial canvas

Optimizing the Agent Graph for Canvas Editing

Dev

The challenge

User testing showed that synchronous canvas operations blocked the voice agent and broke the user’s creative flow. To achieve low latency and high accuracy, the canvas agent had to be decoupled into asynchronous, highly reliable operations.

Graph iterations

To solve this, I've iterated over multiple agent paradigms to find the best fit for Bluegg:

  1. Version 1 | ReAct Agent Graph
    A single agent quickly became bloated trying to manage the vast array of white-boarding tools.
  2. Version 2 | Prompt Chaining
    Provided deterministic logic, but complex branching led to high latency and unmaintainable code.
  3. Version 3 | Orchestrator Pattern
    Delegating tasks to sub-agents solved the tool bloat, but the central orchestrator was too slow for real-time work with large canvas context (images and JSON).
  4. Final Version | Hybrid Agent Graph
    A chained orchestrator parallelizes goal-setting, style-matching, and context retrieval, then routes execution (diagramming, text generation, image creation) to specialized sub-agents.
Final hybrid agent graph for canvas editing

Building It By Using It

Product Management
Design

Bluegg began as a passion project, born from my background in creative direction. It was the brainstorming partner I’d always wished I had.

I use Bluegg in my own design process and have shared it with peers working across visual design, animation, UI/UX and game design. Their excitement validated the concept, and their ongoing feedback became the driving force that shaped the product.

Specs

Tech Stack

Design: Figma

Agentic Coding Tools: Cursor, Claude Code

Frontend Framework: Next.js 15 (TS + React 19)

Backend Framework: Fastify (TS), python-socketio (Python)

Voice Agent Infra: Hume EVI

Voice LLM: Claude Haiku 4.5 or Gemini 3.0 Flash

Canvas Agent Infra: LangChain, LangGraph, LangSmith

Canvas Agent LLM: Gemini 3.0 Flash, Grok 4, and/or Claude Haiku 4.5

Real-time Communication / Networking: Socket.io

Whiteboard: Tldraw, Tiptap

Database: Supabase

Authentication: Auth0

Deployment: Fly.io

Trailer: Photoshop, After Effects

Credits

Ingram Mao

Ivan Pu

Hunter Kitagawa