01 Why This Guide Exists 02 How AI Actually Works 03 The Big Three Providers 04 Chat Tools: What They Do Well 05 Why Code Changed Everything 06 From Code to Cowork 07 When to Use What 08 Connectors: How AI Talks to Your Tools 09 The Browser Fallback 10 Skills: Repeatable AI Workflows 11 The Honest State of Things 12 Where This Is All Going The Toolbox — Tools Worth Knowing
Beginner's Guide

Making Sense of AI Tools

A practical, plain-language guide for people who aren't developers but want to understand the landscape — and know when to use what.

By Niels Kristian Schjødt · February 2026 · ~20 min read

Illustration of a person standing in a landscape holding a map, with AI tool icons on the horizon
Section 01

Why This Guide Exists

If you've been trying to keep up with AI tools over the past year, you're probably confused. And honestly? You should be. The landscape is moving faster than anyone can track, new tools launch weekly, and every one of them promises to change everything.

You've heard of ChatGPT. You might use Gemini. Someone on your team is raving about something called "Cowork." There's talk of "agents" and "MCP connectors" and "skills." And you're sitting there thinking: I just want to know which tool to open when I need to get something done.

This guide is for you. Not for developers. Not for AI researchers. For people who do real work every day — sales, operations, marketing, management, finance — and want to use AI effectively without drowning in jargon.

We'll start with the basics (what does AI actually do under the hood?), move through the major tools and how they differ, and end with a practical framework for knowing when to use what. Along the way, I'll explain the buzzwords in plain language so you can navigate this yourself — even when the tools change, which they will.

Section 02

How AI Actually Works (The 60-Second Version)

Let's start with the thing everyone uses but few people actually understand: large language models, or LLMs. These power ChatGPT, Gemini, Claude, and essentially every AI tool you interact with today.

Here's the secret: an LLM is basically an incredibly sophisticated autocomplete. You know how your phone suggests the next word when you're typing a message? An LLM does the same thing — but trained on an enormous amount of text, so its predictions are remarkably good.

When you send a message (called a prompt), the AI doesn't "understand" your question the way a human does. It predicts, one word at a time, what the most likely response should be. The first word, then the next, then the next — until it has a complete answer. It's doing statistics at an incredible scale.

This is both the magic and the limitation. Because it's predicting rather than "knowing," it can sometimes be confidently wrong. But when given good input and the right context, the predictions are so good that the distinction barely matters for most practical work.

💡 What is a "prompt"?

A prompt is simply the message you send to an AI tool. It could be a question ("What are the top competitors in the Danish car market?"), an instruction ("Write a follow-up email to this client"), or a conversation ("Here's my situation — what would you suggest?"). The quality of the prompt directly affects the quality of the response. More context, more specific instructions, and clearer goals all lead to better output.

💡 What is a "token"?

AI models don't process words — they process tokens, which are fragments of text. A token might be a word, part of a word, or even just a punctuation mark. When people talk about a model's "context window" (how much it can consider at once), they're talking about tokens. For practical purposes, think of 1 token as roughly ¾ of a word. A 200,000-token context window can hold roughly a 300-page book.

Infographic explaining how AI predicts the next word, using a phone autocomplete analogy
How AI "thinks" — it's prediction, not knowledge.

Now, an important concept: thinking models. You may have seen options like "Deep Research" in Gemini or "Extended Thinking" in Claude. These aren't fundamentally different AI — they use the same underlying model. The trick is that the AI re-prompts itself internally, breaking a complex question into smaller steps, checking its own reasoning, and refining its answer before showing it to you. Think of it as the AI taking a moment to think before speaking, rather than blurting out the first thing that comes to mind.

This iterative approach — doing things in steps rather than all at once — turns out to be the key idea behind everything that follows in this guide.

Section 03

The Big Three: OpenAI, Google, and Anthropic

The AI landscape has consolidated around three major providers. Others exist, but these are the ones that matter for everyday work:

Provider Tool Personality Where it shines
OpenAI chat.openai.com ChatGPT The Swiss Army knife. First mover, solid at everything — and the best live voice conversation of any AI. General-purpose tasks. Live voice conversations. Large plugin ecosystem. The best "thinking partner" when you talk to it.
Google gemini.google.com Gemini The researcher with a search engine brain. Built by the company that indexes the internet. Up-to-date web research. Image and video understanding. Deep integration with Google Workspace (but no MCP support yet).
Anthropic claude.ai Claude The focused work partner. Quieter brand, deeper on quality and reliability for real work. Complex analysis and writing. Code and technical work. Enterprise workflows via Cowork.

A few things worth knowing about this race:

They all use capable models. Each provider has a large language model that powers their tools, and they're all remarkably good. The lead changes from month to month — one will have the "smartest" model for a few weeks, then another overtakes it. For most practical purposes, the differences at the model level are small enough that they shouldn't drive your choice of tool.

The differences are in the edges. Google Gemini benefits enormously from having Google Search built in. If you need current, real-time information — who just won an election, what a company's latest quarterly results are, what a competitor launched yesterday — Gemini has a structural advantage because it can search the internet's best index. Google has also invested heavily in visual understanding, making Gemini strong with images and video.

Anthropic has focused on being the reliable work partner — less flashy, more focused on getting complex tasks right. This is where Claude Code (for developers) and Claude Cowork (for everyone) came from. OpenAI, meanwhile, is the generalist: a strong product that does most things well, backed by the largest user base and the broadest ecosystem of third-party integrations.

ChatGPT's killer feature: live voice. One area where ChatGPT is clearly ahead is live voice conversations. You can have a real, flowing, spoken conversation with it — and it's remarkably natural. This might sound like a novelty, but it's genuinely powerful in practice. Put in your headphones on your commute, start a live conversation, and tell it: "I'm going to be capturing thoughts. Just help me think through them." Then talk. About the meeting you're walking into. The report you need to write. The strategy you're chewing on. The loose ideas you haven't shaped yet. ChatGPT listens, asks good follow-up questions, and the entire conversation is captured as a written transcript. When you get to your desk, you take that transcript and paste it into Cowork or any other tool to build the actual deliverable — the presentation, the document, the plan. It turns your commute into productive thinking time, and it turns scattered spoken thoughts into structured source material. I use this workflow constantly.

Practical tip: The AI tool you use the most is the one that knows you best. These tools learn from your previous conversations, your preferences, and your patterns. If you've been using ChatGPT for six months, it has context about you that Gemini doesn't. That accumulated context makes a real difference in day-to-day quality. Don't switch tools casually — but do know what each one is best at.
Overview infographic: From simple chat to collaborative agents — the Big Three providers and the shift from chat to agentic workflows
The full landscape at a glance — from chat tools to agentic workflows.
Section 04

Chat Tools: What They Do Well (And Where They Stop)

When most people say "AI," they mean chat. You type a message. AI responds. You refine. It responds again. This back-and-forth is the core interaction model of ChatGPT, Gemini, and Claude.

Chat tools are excellent for:

For all of these, the expected output is relatively simple: a conversation, a short document, a list, a suggestion. You might go back and forth a few times to refine it, but the AI is essentially doing one thing at a time — responding to your latest message.

This is what we call one-shot interaction. You prompt, it responds. Maybe you ping-pong a few times. The output is the conversation itself, or a fairly simple deliverable.

Chat tools have gotten remarkably good at this. They're mature, fast, and reliable. For 80% of what most people need AI for, a chat tool is the right answer.

But there are tasks where chat tools hit a wall.

Google Gemini chat interface showing a simple conversation and attachment options including Google Drive, Photos, and NotebookLM
Google Gemini — a typical chat interface. Simple back-and-forth with options to attach files from Drive, Photos, and more.
Section 05

Why Code Changed Everything (And Why It Matters to You)

To understand why tools like Cowork exist, you need to understand something about code. Don't worry — you don't need to write any. But this insight is the key to understanding the entire AI tools landscape.

When AI writes a text document — an email, a report, a summary — it doesn't have to be perfect. If a word choice is slightly off or a comma is in the wrong place, a human can still read and understand it. The output is forgiving.

Code is the opposite. If a single character is wrong — one misplaced dot, one missing bracket, one typo — nothing works. It doesn't matter if the other 10,000 characters are perfect. One small error and the whole thing breaks.

This is the fundamental problem AI providers had to solve. The same language model that writes you a lovely email also writes code — but code demands a level of precision that a "predict the next word" system can't guarantee in a single pass.

The same AI that casually writes your emails had to learn an entirely different discipline to write reliable code. That discipline — working iteratively, checking its own work, running in small steps — is what powers everything that came after.

The solution? Agents.

Instead of generating thousands of lines of code in one go and hoping it works, AI providers built systems where the AI works in small steps. It writes a piece, checks it, fixes what's wrong, moves to the next piece, tests again. It can run the code, see if there's an error, and correct it — all on its own, without you having to intervene.

This iterative, self-correcting approach is what an "agent" fundamentally is: an AI that doesn't just respond once, but takes actions in a loop — doing, checking, adjusting, continuing — until the job is done.

💡 What is an "agent"?

In AI, an agent is more than a chatbot. It's an AI that can take actions: run code, read files, call external services, check its own output, and work through problems step by step. A chat tool waits for you to type and responds. An agent takes initiative — it can plan a sequence of tasks, execute them, handle errors along the way, and deliver a finished result. Think of it as the difference between asking someone a question and giving someone a project.

Anthropic's Claude Code was the product that cracked this for software development. It combined a powerful language model with this agent-like, iterative approach, and the result was transformative. For the first time, developers could trust AI to produce reliable code — not because the model was perfect, but because the system was designed to catch and correct its own mistakes.

But here's where the story gets interesting for you — even if you've never written a line of code in your life.

Section 06

From Code to Cowork: How a Developer Tool Became Everyone's Power Tool

Something unexpected happened after Claude Code launched. People with technical backgrounds — developers, analysts, engineers — started using it for tasks that had nothing to do with software development.

They realised that a lot of professional work looks a lot like coding from the AI's perspective: it's complex, multi-step, requires precision, and benefits enormously from the agent's ability to work iteratively. Think about what goes into a serious market analysis:

This isn't a quick chat prompt. This is a project. And an agent — one that can work through each step methodically, check its own output, and build up to a finished deliverable — is exactly the right tool for it.

Anthropic noticed this pattern and built Claude Cowork: a product that gives everyone the power of an agent-based workflow, without needing to know code. Under the hood, it uses the same agent technology as Claude Code, but wrapped in a friendly interface with built-in safety guardrails. It runs in a secure environment on your computer, can read and write files, and works through complex tasks step by step.

Cowork is what happens when a tool built for code realises that most complex work has the same structure: multi-step, requires precision, and benefits from an AI that checks its own work.
Claude Cowork interface showing progress tracking, workspace folder, connectors for web search and Chrome, and a skills panel
Claude Cowork — an agent working through a multi-step task with progress tracking, connectors, and skills.

A few important things to understand about Cowork:

It runs code on your behalf. Even though you never see or write code, Cowork is generating and executing code behind the scenes to accomplish your tasks. When it analyses a spreadsheet, it's writing a small program to do it. When it creates a chart, it's running code to render it. This is why it can be so much more precise and reliable than a chat tool for complex tasks — it's not just predicting what a good answer might look like, it's computing it.

It works with your files. You point Cowork at a folder on your computer, and it can read, create, and modify files in that folder. This is fundamentally different from a chat tool, which can only work with what you paste into the conversation.

It's still early. As of February 2026, Cowork is in research preview. It's powerful but rough around the edges — and we'll talk about that honestly in Section 11.

Section 07

When to Use What: A Practical Decision Framework

This is the section you can bookmark. When you're staring at a task and wondering which AI tool to reach for, here's how to think about it:

💬

Quick question or brainstorm

"Who are the main competitors in X?" or "Help me draft this email."

Any chat tool
🔍

Research with current data

"What did Company X announce last week?" or "Find recent reviews of this product."

Gemini
📊

Big, structured analysis

"Analyse these 50,000 rows and produce a report with charts and recommendations."

Cowork
🔄

Repeatable monthly workflow

"Every month, pull our ad data, analyse performance by market, and produce a deck."

Cowork + Skills

The core principle is simple: match the tool to the complexity of the task.

If the expected output is a conversation, a short text, or a simple answer — use a chat tool. They're faster, more mature, and perfectly suited for it. You don't need to spin up Cowork to ask "what's a good subject line for this email."

If the task is multi-step, requires working with files, demands precision with numbers, or produces a complex deliverable — that's where Cowork earns its keep. The overhead of setting it up only makes sense when the task is substantial enough to justify it.

And if the task involves searching for current information on the web, Gemini's integration with Google Search gives it a genuine edge over the others.

Remember: The AI tool you've used the most is the one that knows you best. For everyday chat interactions, your accumulated conversation history, preferences, and patterns make a real difference. Stick with your favourite for the quick stuff — and reach for specialised tools when the task demands it.
Decision flowchart: Pick Your AI Partner — showing when to use Chat, Gemini, Cowork, or Cowork with Skills
The decision map — start with "What do you need?" and follow the path to the right tool.
Section 08

Connectors: How AI Talks to Your Other Tools

So far, everything we've discussed involves you and the AI in a conversation. But what if the AI could reach into your other tools — your Google Drive, your CRM, your project management system — and pull in the data it needs?

That's what connectors do. They're the bridges between AI and the rest of your software ecosystem.

💡 What is MCP?

MCP stands for Model Context Protocol. It's a standard — originally created by Anthropic but now widely adopted — that defines how AI tools connect to external services. Think of it as a universal adapter: just as USB lets any device plug into any computer, MCP lets any AI tool plug into any compatible service.

You'll see connectors called different things depending on where you are: "apps" in ChatGPT, "MCP servers" or "connectors" in Claude. They're all doing the same thing: giving AI access to data and actions in other systems. Gemini doesn't support MCP yet — it has its own built-in "extensions" for Google services (Drive, Gmail, Calendar, Maps) but can't connect to arbitrary third-party tools the way Claude and ChatGPT can.

Learn more about MCP →

Here are some practical examples of what connectors enable:

The key insight is that MCP connectors are becoming a cross-platform standard. The same MCP connector works in Claude Chat, in Cowork, and in ChatGPT. If your company builds an MCP connector for its internal data, that connector can be used across multiple AI tools — not just one. This is a big deal: it means you invest once and benefit everywhere. Google is the notable exception right now — Gemini only connects to Google's own services natively, though this will likely change as MCP adoption grows.

How to: Connect Google Drive to Claude
  1. Open claude.ai and navigate to your settings.
  2. Look for Integrations or Connected Apps.
  3. Find Google Drive and click Connect.
  4. Authorise access with your Google account. You'll choose which folders or files to share.
  5. In your next conversation, Claude can now reference your Drive files. Try: "Summarise the latest version of our Q1 report from Google Drive."
Claude connectors panel showing Drive, Calendar, Gmail, Figma, HubSpot, Linear, Slack, and Claude in Chrome
Claude — connectors for Drive, Figma, HubSpot, Slack, and more.
ChatGPT Apps panel showing GitHub, Research, Linear, and Notion integrations
ChatGPT — apps for GitHub, Linear, Notion, and deep research.
MCP explained — how AI plugs into your work tools. (1 min)
Section 09

The Browser Fallback: When There's No Connector

What happens when AI needs to interact with a tool that doesn't have a connector yet? There's a fallback: browser control.

Several AI tools — including ChatGPT's agent mode, Claude's computer use, and Cowork's built-in browser — can open a web browser and interact with it the way you would. It navigates to a website, reads what's on the screen, clicks buttons, fills in forms.

This sounds impressive, and it is. But it's important to understand why it's a fallback and not the primary way AI should interact with your tools:

Browser control is great for one-off tasks where no connector exists — like filling out a form on a website you rarely use, or checking something on a service that doesn't offer an API. But for anything repeatable or data-heavy, you want a proper connector.

Think of it this way: A connector is like having a direct phone line to someone. Browser control is like sending a person to physically walk to their office, read what's on their screen, and report back. Both get you the information, but one is dramatically faster and more reliable.
Section 10

Skills: Teaching AI to Repeat What You Need

This is where things get really powerful. Imagine you have a monthly task: pull ad spend data, break it down by market, compare against last month, flag anything unusual, and produce a formatted report. Every month, same process.

Without skills, you'd have to explain this from scratch every time. With a skill, you write the instructions once, and the AI follows them reliably every time you invoke it.

A skill is simply a set of instructions written in a text file. It tells the AI exactly how to perform a specific task: what data to pull, what steps to follow, what the output should look like, and what standards to maintain. Think of it as a recipe card that the AI follows precisely.

💡 What is Markdown?

Markdown is a simple way to format text using plain characters. For example, **bold** makes text bold, and # Heading creates a heading. It's the preferred format for AI instructions because it's clean, structured, and easy for both humans and AI to read.

You can write Markdown in any text editor, and many tools support it natively. Google Docs even has a Markdown setting: go to Tools → Preferences → Enable Markdown. You can also paste Markdown content via Edit → Paste from Markdown, or copy your doc as Markdown with Edit → Copy as Markdown.

Markdown basics →

Google Docs Preferences panel showing the Enable Markdown checkbox
Enable Markdown in Google Docs: Tools → Preferences.
Google Docs Edit menu showing Copy as Markdown and Paste from Markdown options
Edit menu with Paste from Markdown and Copy as Markdown.

Skills were originally a concept from Claude Code (for developers), but the idea has been adopted into Cowork and is spreading across the industry. The reason is simple: it works. Once you've written a good skill, the AI follows it consistently without you having to re-explain the task.

Here's what a real skill might look like in practice:

Example: Creating a "Monthly Ad Budget Review" Skill
  1. Create a text file called monthly-ad-review.md in your Cowork workspace folder.
  2. Describe the task in plain language: "Every month, open the ad spend spreadsheet, calculate total spend per market, compare to last month's numbers, flag any market where spend changed by more than 15%, and produce a summary report as an Excel file."
  3. Specify the data sources: "The ad spend data is in ad-spend-2026.xlsx. Last month's report is in reports/."
  4. Define the output format: "The report should have one sheet per market, a summary sheet with charts, and a one-paragraph executive summary."
  5. Save the file. Next time you open Cowork and ask it to "run the monthly ad review," it will find and follow these instructions automatically.

The beauty of skills is that they're just text files. You can read them, edit them, share them with colleagues, and improve them over time. And because they use the MCP-standard format, skills you create for Cowork will likely work with other AI tools as they adopt the same standards.

AI Skills explained — the recipe card concept. (1 min)
Section 11

The Honest State of Things (February 2026)

I wouldn't be honest if I didn't include this section. The tools are genuinely powerful, but they're also genuinely early. Here's what you should know:

Cowork is in research preview. This means Anthropic is shipping updates almost daily. Things that work on Monday might behave differently on Wednesday. Features that work on Mac may not work on Windows yet, and vice versa. If you've tried it and found it frustrating, that's not you — it's the reality of using a product this early in its development.

Connectors are incomplete. As of right now, many connectors are read-only. For example, the Google Drive connector in Claude can read your files, but it can't reliably write back to them. This means you'll sometimes need to bridge the gap manually: download a file from Google Drive, let Cowork edit it locally on your computer, and then upload the result yourself.

The chat tools are more stable. ChatGPT, Gemini, and Claude Chat have had years of refinement. They're mature, reliable products. If you're finding Cowork frustrating, it's perfectly fine to use chat tools for now and come back to Cowork as it matures. You're not falling behind — you're being pragmatic.

Competition will improve everything. Google is very likely working on their own Cowork-equivalent, with native Google Workspace integration. OpenAI will do the same. When those launch, the ecosystem will improve rapidly. If you're deeply invested in Google's tools (Sheets, Docs, Drive), Google's version may serve you better than Cowork does today.

My advice: Start with the chat tools you already know. Get comfortable with them. Learn what good prompts look like. Then, when you have a task that's clearly too complex for a chat window — something multi-step, file-heavy, or repeatable — that's when to reach for Cowork. Don't force it before you need it.
Section 12

Where This Is All Going

If you take one thing from this guide, let it be this: the tools are going to keep getting better, fast. What feels clunky today will feel polished in two months. What requires manual bridging today will be seamless by summer. The underlying models improve every few weeks, and the products built on top of them improve just as quickly.

Here's what I expect to see in the near future:

Your job right now isn't to master every tool. It's to build the muscle. Learn to prompt well. Understand the difference between a quick chat task and a complex agent task. Start experimenting with connectors and skills when you have a real use case. The specifics will change, but the underlying concepts won't.

The best time to start was six months ago. The second best time is today. The tools will only get better — and the people who've been building the muscle will have a compounding advantage over those who waited.

If you want to go deeper, explore the rest of this series: It Wasn't Wrong covers the emotional shift for senior developers, Shifting Gears has practical dos and don'ts, the 75 Questions dives into the bigger picture, the AI Cookbook covers principles and daily tactics, and Don't Just Tell It. Enforce It. is about building automated quality guardrails. For what happens when software becomes nearly free to build, see The Price of Software Just Hit Zero. And for podcasts to keep you sharp, there's my curated list.

Bonus

The Toolbox — Tools Worth Knowing

Beyond the Big Three chat tools, there's a growing ecosystem of specialised AI tools that are genuinely useful right now. This isn't an exhaustive list — it's the tools I personally use and recommend. Click each one to see what it does and how I use it.

📓
Turn any source material into shareable knowledge — podcasts, videos, FAQs, briefings.

NotebookLM is Google's research and knowledge tool, and it's become one of my most-used tools for a specific reason: it's the fastest way to bring people up to speed on a project. You feed it sources — PDFs, Google Docs, YouTube videos, meeting transcripts, website URLs — and it can generate podcasts, videos, infographics, FAQs, study guides, timelines, and more from that material.

Its original killer feature was the podcast generator: upload your sources, and it creates a surprisingly engaging two-host audio discussion of the material. Now it's even interactive — you can join the conversation and ask questions while it records. But the real power is broader than podcasts. It's a knowledge broadcasting machine.

If you use Google Workspace, it's even better — all your Drive documents, Slides, and Sheets are one click away as sources. The key to getting great results: give it as much context as possible. Don't be selective. Throw in the transcripts, the PDFs, the meeting notes, the specs. More context = better output.

How I use it For this very guide, I used NotebookLM to generate the explainer videos and infographics. I uploaded my conversation transcripts and a rough outline of what I wanted, then prompted it to produce visual assets. I also use it to prepare briefings when I need to get a cross-functional team up to speed on a technical project quickly — instead of writing a 10-page doc, I generate a 12-minute podcast they can listen to on their commute.
NotebookLM interface showing sources panel and generated podcast, FAQ, and study guide outputs Try NotebookLM →
🍌
The best AI image generation available right now. Exceptional quality, zero setup.

Gemini's image generation — nicknamed Nano Banana — is, as of early 2026, the best AI image generator available to the public. The quality is remarkable: it handles styles, compositions, text in images, and complex scenes far better than the alternatives. And there's no setup — you just go to gemini.google.com, type a prompt describing what you want, and it generates it.

If you need cover images, illustrations, diagrams, social media visuals, or any kind of imagery for content — this is the tool. The consistency and artistic quality are a clear step above everything else right now.

How I use it Every cover image on this blog was generated with Nano Banana. I wrote detailed prompts describing the style, mood, and composition I wanted — warm editorial illustration in muted earthy tones — and Gemini produced consistent, publication-ready artwork across all five posts. I also use it for quick social media visuals and presentation imagery.
Gemini generating a warm editorial illustration from a text prompt Try Gemini →
🎬
Veo
Google's AI video generator. The best tool for creating video content from text prompts.

Veo is Google's video generation model, and it's currently the leading tool for creating AI-generated video content. If you need product demos, explainer clips, social media video, or visual storytelling — and you want full control over what's generated — Veo gives you that.

The quality and coherence of the video output is genuinely impressive, and it's improving rapidly. For anyone who needs video content but doesn't have a production team or budget, this is a game-changer.

How I use it When I need short video content to explain a concept or demonstrate a workflow, Veo lets me describe the scene and generate it directly. It's particularly useful for product presentations and internal demos where hiring a videographer would be overkill.
Veo interface showing video generation from a text prompt with preview frames Learn about Veo →
💜
Spin up a full website or web app in minutes. Visual editing, instant deployment.

Lovable is a website and web-app builder that lets you go from a text prompt to a live, deployed website in minutes. You describe what you want, it generates the code, and you can visually edit the result — drag things around, change colours, swap content — without touching any code yourself.

For quick landing pages, internal tools, prototypes, or even fairly advanced web services, it's genuinely fast. It handles the domain setup and hosting too, so there's zero infrastructure overhead. You can also do similar things with Claude Cowork (which is what I used to build this blog), but Lovable is purpose-built for it and has the edge on speed and visual editing.

How I use it When I need a standalone landing page or a quick internal tool that doesn't need to live alongside other project content, Lovable is my go-to. It's also great for prototyping ideas with stakeholders — you can have a working version in front of them within an hour instead of waiting for a sprint cycle.
Lovable interface showing a website being built from a text prompt with visual editing controls Try Lovable →
Visual workflow automation. Connect systems, build AI pipelines, orchestrate anything.

n8n is a visual workflow automation tool — think of it as a way to connect different systems and have them talk to each other through a visual drag-and-drop interface. What makes it especially powerful in the AI era is that you can plug AI models directly into these workflows.

This means you can build things like: "When a new support ticket arrives in Zendesk, have Claude analyse it, categorise it, draft a response, and post it to Slack for review." Or: "Every morning, pull the latest sales data, have an AI summarise the trends, and email a brief to the leadership team." These are advanced orchestration workflows, but n8n makes them accessible without deep programming knowledge.

How I use it We use n8n to automate repetitive cross-system tasks — things like syncing data between tools, triggering AI analysis on incoming data, and building notification pipelines. For example, we have a workflow that monitors our key metrics, runs an AI analysis when something changes significantly, and alerts the right team with context already included.
n8n visual workflow editor showing connected nodes for an AI-powered automation pipeline Try n8n →
🤖
Build custom AI agents that can use tools, browse the web, and execute multi-step tasks.

OpenAI's Agent Builder (part of the OpenAI platform) lets you create custom AI agents — purpose-built assistants that go beyond simple chat. These agents can use tools, search the web, read files, execute code, and carry out multi-step procedures autonomously.

Where n8n is about connecting systems visually, the Agent Builder is about creating intelligent actors that can reason through complex tasks. You define the agent's purpose, give it access to the tools it needs, and it handles the rest. It's more technical than the consumer chat tools, but the power is significant for anyone building AI-powered products or internal tools.

How I use it We've experimented with custom agents for tasks like automated code documentation, research synthesis, and data analysis pipelines. The key is defining clear boundaries for what the agent should do and giving it the right tools — then letting it reason through the steps on its own.
OpenAI Agent Builder interface showing agent configuration with tool access and instructions Explore OpenAI Platform →
🧪
Google's all-in-one workshop for building apps, generating images, and prototyping with AI.

Google AI Studio is Google's playground for building with AI. Think of it as something between Cowork and a developer tool — it's more technical than consumer chat, but far more accessible than writing code from scratch. You can build apps, generate images, create functionalities, test prompts, and prototype ideas, all within a single interface.

It gives you direct access to Google's latest models (Gemini, Veo, Imagen) and lets you combine them in creative ways. If Cowork is about working alongside AI on your daily tasks, AI Studio is about building new things with AI as the foundation. It's particularly powerful if you're already in the Google ecosystem.

How I use it I use AI Studio when I want to prototype something quickly — test a prompt idea, build a small utility app, or experiment with combining different AI capabilities (like text + image + code). It's also great for exploring what the latest models can do before committing to building something larger. If you want to understand what's possible with AI right now, this is the sandbox to play in.
Google AI Studio interface showing prompt testing, app building, and model selection Try Google AI Studio →