Skip to main content

Give ideas more space with Jambot

Amber BravoDirector of Content & Editorial, Figma

Jambot brings the generative power of ChatGPT to FigJam, helping you take your ideas (and meetings) further.

One of the delights of using a Large Language Model (LLM), like OpenAI’s ChatGPT, is its ability to simulate the act of riffing—bouncing ideas back and forth in a way that feels spontaneous, light, and conversational, but with a supercharged bank of knowledge. It’s like that friend who’s funny (if a bit awkward), knows a little something about everything, and never tires of it always being about you. For humans who are uniquely wired to want to explore and push ideas around, AI can also act as a major creative boost, even if the experience of using them can sometimes feel a bit one-sided. A team at Figma wanted to find out what it would take to break the chatbot out of the box and bring the power of GPT into a multiplayer canvas environment. The result is Jambot, a new FigJam widget that lets you ideate, summarize, and riff in FigJam. We sat down with the team to learn what inspired them to make the widget, and why they’re so excited to see ChatGPT go multiplayer.

Jambot brings the generative power of ChatGPT to your FigJam files, letting you ideate, summarize, and play.
Amber B.

Jambot was born out of a hackathon project at Figma. Can you tell me a little bit of the backstory?

Daniel M.

When our internal AI-focused hackathon came up, I pitched the project as “a visual version of ChatGPT.” I love ChatGPT, and I love seeing all kinds of people using it. I think the simple chat-based interface is super powerful to interact with AI, but I don’t think it’s the optimal format for all scenarios. As a simple example, when ChatGPT gives you multiple options, you can dig into one and ask follow-up questions, but when you want to go back to another option, you have to scroll back up, and then start to repeat the same questions. More generally, since the sequence of chat messages is one dimensional, it’s not very natural to branch out to different topics, or tweak things to see how they all connect. I’ve been a heavy user of these tools called Networked Thought—especially Roam Research and Logseq—which basically allow you to create pages that link between each other, so you can connect, organize, and trace ideas. More recently, I also found this tool called Albus, which adds a visual feel to interacting with AI, and so I thought there should be a way to connect these concepts to create a potentially useful alternative to ChatGPT.

I think the simple chat-based interface is super powerful to interact with AI, but I don’t think it’s the optimal format for all scenarios.
Daniel Mejia, Engineer
Sam D.

I was also thinking a lot about LangChain, which is like Python in that it employs an object-oriented, semantic approach, but in a high-level programming language environment. You have to know how to code in order to make any use of it, but I thought it would be really cool to have something that was visually tangible and easy to use.

Aosheng R.

I think we’re stuck in chat boxes. Just like we’re stuck in Zoom right now. It might be too early for LLMs to be overly anthropomorphic; maybe they’re not capable of showing emotion, but right now they sort of pretend to. There are so many missed opportunities when we “talk” to ChatGPT—what identity it has, how contextual it can be. So many GUI (Graphical User Interface) primitives we invented in the past 20 years—from cursors to windows—were designed to make computers effectively more approachable, but it’s like we have to start from the beginning again with AI because, right now, it feels like all we have are more rudimentary, command-line operating system models like DOS.

We have to start from the beginning again with AI because, right now, it feels like all we have are more rudimentary, command-line operating system models like DOS.
Aosheng Ran, Product Designer
Amber B.

How did the team arrive at this visual design direction?

Aosheng R.

The team gradually moved towards the formal expression it has now—an almost game-like quality, where it feels fun and low stakes. There are two node connectors, at either an input or output space, and you can see things flow from left to right, as it connects stickies and processes information—either extracting information from a sticky or running a command—and then outputs something. A lot of interaction designers are obsessed with the history of visual programming and why it did or didn’t become more successful than it is today. To me, it’s exciting to take that kind of visual programming interface and apply it to ChatGPT, it just feels right as a tool.

The Jambot widget embedded in a Figma fileThe Jambot widget embedded in a Figma file
Jambot brings ChatGPT into the FigJam canvas, helping teams organize or expand on ideas.
Amber B.

How did you land on the input and output functions?

Sam D.

Going into the hackathon, the thing that I didn’t want to do is just have some text box, because anyone can write a text box that submits an API request. Having a pre-set list of functions gives users some clear affordances. Having those prompts makes [the capabilities] way more discoverable without having to go down this rabbit hole of, “How do I learn this new language or this new way of interfacing with a computer?”

We had many discussions about exactly what these functions should be, and which ones would be most useful:

1. Ideate: We wanted to be able to have one sticky turn into many stickies.

2. Summarize: Have many stickies turn into one sticky.

3. Play: We wanted to find interesting ways to expand on ideas and summaries.

The thing that I didn’t want to do is just have some text box, because anyone can write a text box that submits an API request. Having a pre-set list of functions gives users some clear affordances.
Sam Dixon, Engineer

The one-to-one behavior, where you convert one sticky into another sticky that’s somehow different, like turning feedback into a song, is really our opportunity to bring an element of play into the tool. We wanted to make Jambot feel Figma-y and whimsical. LLMs like GPT tend to be really good at doing, so it felt natural to incorporate those prompts into Jambot.

A FigJam File shows the Jambot widget in actionA FigJam File shows the Jambot widget in action
Jambot can help quickly ideate, summarize, or playfully expand on content in a FigJam file.
Generative AI and the creative arms race

Ovetta Patrice Sampson, Director of UX Machine Learning, Google, describes this phenomenon as a sort of souped up mad lib and gives us a concise history of how LLMs got started in her Config 2023 talk.

Amber B.

People are going to enjoy using this, just like they enjoy using ChatGPT. It’s playful, but there’s also a real value it can add in the context of work.

Daniel M.

At Figma, our vision is that AI is a tool to supercharge work and collaboration. It’s less about having AI do things for you, and more like AI is helping you to do a better job, whether that’s designing or otherwise. Jambot can be a superior model for certain types of interactions. We’re excited to see what people do with it, and to get feedback that could help inform our direction and features we might want to pursue natively. I want people to say, “Oh, for this task, Jambot is actually much better than a chatbot—it’s more fun; it’s more visual; it’s more interactive.” There’s a lot that we could do to make this an even stronger alternative for use cases that benefit from nonlinearity or visual expression, and since it’s built on top of FigJam, that means Jambot immediately gets visual and multiplayer capabilities that differentiate it from other AI products. This intrinsically makes it well-suited for collaborative work, while also keeping it fun.

At Figma , our vision is that AI is a tool to supercharge our collaborators. It’s less about having AI do things for you, and more like AI is helping you to do a better job, whether that’s designing or otherwise.
Daniel Mejia, Engineer
Sam D.

Jambot lets you play around with an idea or information rather than completing a [clearly defined task]. I think that aligns with the idea of having AI as an assistant rather than an oracle of some sort, where it tells you just one thing, and you have no way of intervening or acting on the output.

Daniel M.

I’m really excited about how Jambot can empower people to fluidly dive into different paths and more effectively leverage AI to learn faster and open up new ideas that can be visualized and tinkered with. Imagine the possibilities when Jambot goes beyond just text inputs and outputs. There’s also the final visual artifact that captures the thinking and learning process which you can later revisit or even share. One additional value that at least for me only emerged after seeing the hackathon widget in action, was how having these functions automatically suggested as next steps really keep people in a flow state with surprising ways to use AI that they wouldn’t otherwise know to type on an empty box (e.g why don’t you turn this into a poem). This was magical to witness.

Get started with Jambot in FigJam.

Jambot was created by Daniel Mejia, Sam Dixon, and Aosheng Ran, along with many other Figma collaborators who helped bring it from demo to widget in just a few, short weeks.

Amber Bravo is the Director of Editorial & Content at Figma. Previously, she's worked as a writer and editor at Google Design, Herman Miller, The FADER, and Dwell Magazine.

In Conversation
Amber BravoDirector of Content & Editorial, Figma
Daniel MejiaEngineer, Figma
Sam DixonEngineer, Figma
Aosheng RanProduct Designer, Figma

Subscribe to Figma’s editorial newsletter

By clicking “Submit” you agree to our TOS and Privacy Policy.

Create and collaborate with Figma

Get started for free