What Is Cursor's New Feature That Lets AI Agents Film Themselves Coding?

Cursor's new agent computer use feature lets AI agents control their own VMs, film themselves working, and create pull requests. Learn how it works and how to enable it.

Ashley Innocent

Ashley Innocent

25 February 2026

What Is Cursor's New Feature That Lets AI Agents Film Themselves Coding?

TL;DR

Cursor's new agent computer use feature allows AI cloud agents to control their own virtual machines, enabling them to build, test, and verify code autonomously. Agents run in isolated VMs with full development environments, can open browsers and navigate localhost, and produce merge-ready pull requests with videos and logs. Access via Cursor Desktop, Web, Slack, GitHub, or API. Over 30% of Cursor's internal code changes are now created by these autonomous agents.

Introduction

The way developers write code is changing fast. For years, AI coding assistants like Cursor have worked as intelligent autocomplete tools—suggesting the next line, explaining code, or helping refactor existing work. But a major shift is happening.

Cursor announced what they call "the biggest shift in how we build software since the move from Tab autocomplete to working synchronously with agents." Their cloud agents can now control their own virtual machines to build, test, and demonstrate working code.

This isn't just incremental improvement. According to Cursor's announcement, 30-35% of their merged pull requests are now created by autonomous cloud agents.

In this guide, you'll learn what Cursor's agent computer use feature does, how it works, and exactly how to set it up.

💡
For teams building APIs, combining Cursor's autonomous agents with dedicated API tools like Apidog creates a powerful workflow—use Apidog to design, test, and document your APIs while letting agents handle implementation.
button

What Is Cursor Agent Computer Use?

Cursor agent computer use is a feature that lets AI agents operate in their own isolated virtual machines rather than just interacting with your local development environment. When you launch a cloud agent, it spins up a full development environment in the cloud—an isolated VM with everything needed to build, test, and verify software.

The agent recorded itself interacting with desktop applications in its VM.

Why This Matters

Traditional AI coding assistants work by:

Cursor's cloud agents take this further by actually doing the work:

The Technology Behind It

Each cloud agent runs in its own isolated virtual machine. This creates several important benefits:

  1. Isolation: Agents don't compete for local resources
  2. Parallelism: Multiple agents can work simultaneously
  3. Verification: Agents can actually test their changes
  4. Evidence: Each PR includes videos, screenshots, and logs

Cloud Agents vs. Traditional IDE AI Assistance

Understanding the difference between cloud agents and traditional IDE-based AI helps set realistic expectations:

Traditional AI Coding (Cursor Classic Mode):

Cloud Agents (Autonomous):

The key shift is from "AI assists me" to "AI does it for me." This doesn't replace developers—it shifts your role from implementation to review and direction.

Technical Requirements Explained

Understanding what's happening under the hood helps when troubleshooting:

This transparency means you always know what environment your code was built and tested in.

How Cursor Cloud Agents Work

Architecture Overview

When you kick off a Cursor cloud agent, here's what happens:

  1. VM Provisioning: Cursor spins up an isolated virtual machine with a full development environment
  2. Repository Access: The agent clones your repository and any dependencies
  3. Task Execution: The agent reads your requirements, makes changes, and runs tests
  4. Verification: The agent opens browsers, navigates to localhost, clicks through UI to verify everything works
  5. Artifact Generation: The agent records videos, screenshots, and logs showing what was done
  6. PR Creation: The agent creates a pull request with all the evidence

Where You Can Launch Agents

Cursor cloud agents are accessible from multiple platforms:

PlatformHow to Launch
Cursor DesktopSelect "Cloud" in the dropdown under the agent input
Cursor WebVisit cursor.com/agents
SlackUse the @cursor command
GitHubComment @cursor on a PR or issue
APIUse the Cursor API to kick off an agent
LinearUse the @cursor command

Requirements

Before using cloud agents, ensure you have:

Currently, cloud agents support GitHub and GitLab repositories.

Step-by-Step Setup Guide

Step 1: Access the Onboarding

Navigate to cursor.com/onboard to get started. This page walks you through the initial agent configuration and lets you watch the agent set itself up.

Cursor onboard interface

Step 2: Choose Your Launch Platform

You have multiple options:

Option A: Cursor Desktop

  1. Open Cursor IDE
  2. Find the agent input at the bottom
  3. Click the dropdown and select "Cloud"
  4. Describe what you want built

Option B: Cursor Web

  1. Go to cursor.com/agents
  2. Sign in with your account
  3. Select your repository
  4. Describe your task
Cursor Agents on the Web

Option C: GitHub Integration

  1. Navigate to your repository
  2. Open an issue or PR
  3. Comment @cursor build a feature that...
  4. The agent will pick up the task

Step 3: Configure Agent Privileges

When you first use a cloud agent, you'll need to grant:

Step 4: Define Your Task

Be specific about what you want the agent to accomplish. For example:

The more context you provide, the better the agent can deliver.

Step 5: Monitor and Review

While agents work autonomously, you can:

Key Features and Capabilities

Self-Testing and Iteration

Perhaps the most powerful feature: agents can actually verify their work. They can:

Testing UI on Cursor page

The agent spent 45 minutes doing a full walkthrough of cursor docs site. It provided a summary of all the features it tested, including the sidebar, top navigation, search, copy page button, share feedback dialog, table of contents, and theme switching.

Artifact Recording

Every agent run produces rich artifacts:

This makes review fast—you can see exactly what was done without checking out the branch.

Remote Desktop Control

Want to try the changes yourself? You can take over the agent's remote desktop directly. This lets you:

Multi-Platform Access

Cloud agents work wherever you are:

Real-World Use Cases

Building New Features

Describe a feature you need, and the agent builds it end-to-end: scaffolding, implementation, tests, and verification.

Reproducing Bugs

Tell the agent to reproduce a bug, and it can:

Quick Fixes

For small changes, you can delegate entirely to the agent without context switching.

The agent recorded itself to show that it implemented the lint label fix correctly.

UI Testing

Agents can click through interfaces to verify:

API Integration Work

When building API integrations, agents can:

This is where tools like Apidog complement Cursor's agents beautifully—use Apidog to design and test your APIs, then let Cursor agents handle the frontend implementation and integration. For example, you can test localhost APIs with webhook services to verify your API integrations work correctly before deploying.

Apidog Design Interface

Limitations and Considerations

Current Limitations

When to Use vs. Not Use

Best for:

May not be ideal for:

Security Considerations

When granting agents repository access, keep these security best practices in mind:

  1. Use scoped tokens: Instead of granting full repository access, use tokens with minimal required permissions
  2. Review agent actions: Always review PRs thoroughly before merging, especially for production code
  3. Separate environments: Consider using agents only on non-production branches initially
  4. Audit logs: Take advantage of Cursor's artifact recording to audit what the agent did
  5. Start small: Begin with low-risk tasks to build confidence before tackling critical features

Cursor's isolated VM approach is inherently safer than giving agents direct local access, but responsible use still requires vigilance.

Future of Autonomous Coding Agents

Best Practices for Getting the Most Out of Cloud Agents

To maximize productivity with Cursor's autonomous agents, follow these proven practices:

1. Write Clear, Specific Prompts

The quality of agent output directly correlates with prompt clarity. Instead of vague requests like "fix the login bug," be specific: "When a user submits the login form with valid credentials, they should be redirected to /dashboard but instead see a 401 error. The server returns the correct token in the response."

2. Provide Context Early

Give agents relevant context upfront: file paths, relevant code snippets, error messages, or links to related issues. Agents can only work with what they know.

3. Use Iterative Refinement

Don't try to have agents build everything at once. Start with a minimal viable version, review the results, then expand. This produces better outcomes than dumping massive requirements in one prompt.

4. Leverage Multiple Agents in Parallel

Since agents run in isolated VMs, you can kick off multiple agents for different tasks simultaneously—one for a new feature, another for a bug fix, a third for documentation updates.

5. Review Artifacts Thoroughly

The video and screenshot artifacts aren't just nice-to-have—they're your window into what the agent actually did. Watch them to catch issues you might miss in code review.

Integration with Your Existing Workflow

Cursor agents integrate smoothly with common development workflows:

The key insight is that agents augment your workflow rather than replacing it. You still review, test, and approve—but the grunt work happens automatically.

Cursor's announcement represents a broader trend. The AI coding space is heating up:

Cursor's 30-35% autonomous PR rate suggests this model works. Expect more tools to adopt similar approaches as the technology matures.

Cursor Internal merged PRs by cloud agents graph

For API development specifically, combining autonomous agents with dedicated API tools creates a powerful workflow: agents handle implementation and testing, while specialized tools like Apidog manage API design, documentation, and comprehensive testing.

Conclusion

Cursor's agent computer use feature represents a fundamental shift in how software gets built. By letting AI agents operate in their own virtual machines, test their changes, and produce verified, artifact-rich pull requests, Cursor is proving autonomous coding works at scale.

The setup is straightforward: choose your platform, define your task, and let the agent work. With 30%+ of Cursor's own code now written by agents, the technology has proven itself.

Next steps:

Get started with Apidog: Try Apidog free to design, test, and document your APIs while leveraging AI-powered coding assistants like Cursor for implementation.

button

Explore more

Gemini 3.1 pro vs Opus 4.6 vs Gpt 5. 3 Codex: The Ultimate Comparison

Gemini 3.1 pro vs Opus 4.6 vs Gpt 5. 3 Codex: The Ultimate Comparison

Compare Gemini 3.1 Pro, Claude Opus 4.6, and GPT-5.3 Codex across benchmarks, pricing, and features. Data-driven guide to choose the best AI model for coding in 2026.

24 February 2026

What Is Gemini 3.1 Pro? How to Access Google's Most Intelligent AI Model for Complex Reasoning Tasks?

What Is Gemini 3.1 Pro? How to Access Google's Most Intelligent AI Model for Complex Reasoning Tasks?

Learn what Gemini 3.1 Pro is—Google’s 2026 preview model with 1M-token context, state-of-the-art reasoning, and advanced agentic coding. Discover detailed steps to access it via Google AI Studio, Gemini API, Vertex AI, and the Gemini app.

19 February 2026

How Much Does Claude Sonnet 4.6 Really Cost ?

How Much Does Claude Sonnet 4.6 Really Cost ?

Claude Sonnet 4.6 costs $3/MTok input and $15/MTok output, but with prompt caching, Batch API, and the 1M context window you can cut bills by up to 90%. See a complete 2026 price breakdown, real-world cost examples, and formulas to estimate your Claude spend before going live.

18 February 2026

Practice API Design-first in Apidog

Discover an easier way to build and use APIs