Cloudhands.ai

Cloudhands is a multi-modal AI platform designed to simplify how users interact with rapidly expanding AI capabilities. As the number of models and tools continues to grow, users face increasing friction in determining which AI to use for specific tasks and in effectively combining them into workflows.

The product was built to act as an orchestration layer across AI tools—enabling users to generate, edit, and optimize content across formats, including image, video, audio, text, and code, within a unified experience. Rather than navigating fragmented tools, Cloudhands centralizes these capabilities and introduces structured workflows to guide users toward better outcomes.

At its core, Cloudhands explores a shift from isolated AI tools toward coordinated systems—where users can leverage pre-defined workflows, or “agents,” to execute complex tasks more efficiently.

My Role

As Head of Design at Cloudhands, I led the end-to-end product design for a 0→1 multi-modal AI platform, working closely with product, engineering, and company leadership to define both the user experience and overall product direction.

My role extended beyond design into product strategy and process. In a highly ambiguous, early-stage environment, I helped define how the team operated by partnering with a project manager to establish an agile workflow, implement work tracking, and structure the product roadmap in ClickUp. This created alignment across design, product, and engineering, and enabled the team to move from exploration to execution more effectively.

I also played a key role in shaping the product itself—defining how Cloudhands approached AI orchestration, multi-modal workflows, and user interaction models across tools, assistants, and agents. I collaborated directly with stakeholders, including the company’s top investor, to align on vision, priorities, and tradeoffs.

While leading design and contributing to product direction, I remained hands-on—driving research, defining user flows, creating high-fidelity designs, and supporting implementation through beta.

Process

Documentation & Research

Problem

As the AI ecosystem rapidly expands, users face an overwhelming number of tools and models—each optimized for different tasks, such as image generation, writing, coding, or video creation. While individual tools continue to improve, the overall experience has become increasingly fragmented.

Users are often left asking:

  • Which model should I use for this task?

  • How do I move between tools to complete a workflow?

  • How do I get consistent, high-quality results without trial and error?

This fragmentation creates friction at multiple levels. At a basic level, users struggle to select the right tool. At a more advanced level, they lack a clear way to chain together multiple AI capabilities into cohesive workflows.

For Cloudhands, the core challenge was not just building another AI tool—but defining a system that could simplify decision-making, unify capabilities across modalities, and guide users from intent to outcome in a more structured and efficient way.

Predispositions

Before exploring solutions, I established a set of initial hypotheses to guide early product thinking. Given the ambiguity of the space, these predispositions helped frame where to focus and which directions to explore.

  • Users need guidance, not just access
    Existing AI platforms provide powerful tools but place the burden of decision-making on users. I believed the product should actively guide users toward the right tools and workflows, rather than simply exposing capabilities.

  • Workflows matter more than individual tools
    Most real-world use cases require multiple steps—such as generating an image, refining it, and adapting it for different outputs. I assumed that enabling structured, multi-step workflows would be more valuable than optimizing isolated interactions.

  • A unified platform can reduce friction
    Switching between tools creates cognitive and operational overhead. I hypothesized that bringing multiple modalities (image, video, text, audio, code) into a single environment would simplify the user experience.

  • Interaction models need exploration
    It was unclear whether users would best engage with the product through tools, assistants, or more automated “agent”-based workflows. I treated this as a core area of exploration rather than a fixed assumption.

  • Simplicity would be critical for adoption
    While the underlying capabilities are complex, the experience needed to feel approachable. I believed the product should balance power with clarity to avoid overwhelming users.

These predispositions shaped early exploration and were continuously validated, refined, or challenged throughout the design process.

Competitive Analysis & Inspiration

To better understand the evolving AI landscape, I analyzed a range of existing platforms to identify patterns in how users interact with AI tools and to identify gaps in the experience.

Platforms like Poe focused on conversational interactions, allowing users to switch between models within a chat-based interface. While this approach made AI more accessible, it primarily focused on text-based workflows and offered limited support for multimodal or structured use cases.

Other platforms, such as Galaxy.ai, took a tool-based approach—surfacing a wide range of AI capabilities in a single place. While this provided breadth, the experience often lacked cohesion, requiring users to navigate and piece together workflows on their own.

Across both approaches, a consistent gap emerged:

  • Limited support for multi-step workflows

  • Minimal guidance on when and how to use different models

  • Fragmentation between tools, outputs, and user intent

These insights helped frame an opportunity for Cloudhands to move beyond both conversational-only and tool-aggregation models, and instead explore a more structured system—one that could unify multimodal capabilities while guiding users through workflows.

This exploration also informed key product directions, including the introduction of AI recipes (structured workflows), model comparison tools, and the concept of orchestration across different AI capabilities.

What Exists Today

When I joined Cloudhands, the product consisted of a broad set of AI capabilities—including image, video, text, and other generative tools—but lacked a clear structure to guide user interaction.

Functionality was added rapidly, resulting in an experience that felt like a collection of features rather than a cohesive product. There was no strong hierarchy, navigation model, or defined workflows connecting these capabilities. As a result, users were left to interpret how different tools related to one another and how to use them together effectively.

This created several challenges:

  • Lack of direction: Users had no clear starting point or guidance on what to do next

  • Disconnected experiences: Tools existed independently, with little support for multi-step workflows

  • Increased cognitive load: The breadth of functionality made the product feel overwhelming rather than empowering

From a product perspective, this also made it difficult to prioritize features, define a roadmap, or align the team around a clear vision.

This initial state highlighted the need to move from a collection of capabilities to a more intentional, structured system that could guide users and scale effectively.

Concepts

Information Architecture

User Flows

 

Drawings

Concepts

Writing

Usability Testing

Hi-Fidelity Designs + Prototypes

Shared Components

File Layout

Design QA

Key Features

AI Assistant

A conversational interface that allows users to interact with AI in a more natural, guided way. The assistant serves as an entry point to the platform—helping users generate content, refine outputs, and navigate its capabilities without needing to understand the underlying models.

Image Generation

Enables users to create and edit images using AI models. Designed to support both quick generation and iterative refinement, allowing users to experiment, adjust prompts, and evolve outputs within a single workflow.

Video Generation

Provides tools for generating and editing video content using AI. This extends the platform’s capabilities to more complex, multi-step creative workflows, enabling users to move from concept to visual output more efficiently.

Audio Generation

Supports the creation and manipulation of non-speech audio, such as music or sound effects. This complements speech generation and enables a broader range of creative and production workflows.

Speech Generation

Allows users to generate realistic voice outputs from text. This feature supports use cases such as narration, voiceovers, and audio content creation, expanding the platform’s reach into spoken media.

Code Generation

Enables users to generate and refine code through AI. Designed for both technical and non-technical users, this feature helps accelerate development tasks, prototyping, and problem-solving.

User Profile

Provides a centralized space for managing user activity, preferences, and generated content. This supports continuity across sessions and lays the foundation for personalization within the platform.

AI Recipes

Introduces structured, repeatable workflows that guide users through multi-step processes. Recipes reduce the need for trial-and-error by combining multiple AI capabilities into predefined sequences for common use cases.

Compare AI Models (Arena)

Enables users to compare outputs from multiple AI models side-by-side. This helps users evaluate quality, understand differences, and make more informed decisions when selecting models.

Results & Conclusion

Through this work, Cloudhands evolved from a fragmented set of AI capabilities into a more structured, multi-modal platform and was brought to beta release.

The product established a cohesive foundation across key areas, including:

  • AI Assistant for conversational workflows

  • Image, video, speech, audio, and code generation

  • AI recipes to guide structured, repeatable workflows

  • Model management and comparison to support informed decision-making

  • User profiles to support personalization and continuity

This shift from isolated tools to a more unified system enabled users to more effectively navigate and utilize a wide range of AI capabilities within a single experience.

In parallel, I helped establish product and design processes that improved team alignment and execution—introducing agile workflows, implementing work tracking in ClickUp, and structuring the roadmap to support prioritization in a rapidly evolving space better.

While the company underwent layoffs before full-scale launch, the beta release demonstrated a clear product direction. It validated the core opportunity: moving beyond individual AI tools toward guided, multi-step workflows and model orchestration.

Key Takeaways

  • Structuring complexity is critical when working with emerging technologies like AI.

  • Clear workflows and guidance can significantly reduce user friction in multi-tool environments.

  • Early investment in product, process, and alignment enables faster, more effective execution.

  • Designing at the systems level is essential when building multi-modal platforms