- Dec 22, 2025
- Parsed from source:Dec 22, 2025
- Detected by Releasebot:Dec 23, 2025
Hooks for security and platform teams
Cursor unveils hooks to observe, block, and extend the agent loop, enabling enterprise control and security workflows. A partner ecosystem adds MCP governance, code security, dependency checks, agent safety, and secrets management. Enterprise onboarding available.
Earlier this year, we released hooks for organizations to observe, control, and extend Cursor's agent loop using custom scripts. Hooks run before or after defined stages of the agent loop and can observe, block, or modify behavior.
We've seen many of our customers use hooks to connect Cursor to their security tooling, observability platforms, secrets managers, and internal compliance systems.
To make it easier to get started, we're partnering with ecosystem vendors who have built hooks support with Cursor.
Hooks partners
Our partners cover MCP governance, code security, dependency scanning, agent safety, and secrets management.
MCP governance and visibility
MintMCP uses beforeMCPExecution and afterMCPExecution hooks to build a complete inventory of MCP servers, monitor tool usage patterns, and scan responses for sensitive data before it reaches the AI model.
Oasis Security extends their Agentic Access Management platform to Cursor, using hooks to enforce least-privilege policies on AI agent actions and maintain full audit trails across enterprise systems.
Runlayer uses hooks to wrap MCP tools and integrate with their MCP broker, giving organizations centralized control and visibility over all agent-to-tool interactions.
Code security and best practices
Corridor provides real-time feedback to the agent on code implementation and security design decisions as code is being written.
Semgrep automatically scans AI-generated code for vulnerabilities using hooks, giving the agent real-time feedback to regenerate code until security issues are resolved.
Dependency security
Endor Labs uses hooks to intercept package installations and scan for malicious dependencies, preventing supply chain attacks like typosquatting and dependency confusion before they enter your codebase.
Agent security and safety
Snyk integrates Evo Agent Guard with hooks to review agent actions in real-time, detecting and preventing issues like prompt injection and dangerous tool calls.
Secrets management
- 1Password uses hooks to validate that all required environment files from 1Password Environments are properly mounted before shell commands execute, enabling just-in-time secrets access without writing credentials to disk.
To deploy Cursor with enterprise features and priority support, talk to our team.
Original source Report a problem - Dec 22, 2025
- Parsed from source:Dec 22, 2025
- Detected by Releasebot:Dec 23, 2025
2.3
Holiday release brings stability and bug fixes across the core agent and layout controls. Users get four default layouts—agent, editor, zen, browser—with a quick switch shortcut and macOS style navigation for faster workspace changes.
Stability improvements
For this holiday release, we've focused entirely on fixing bugs and improving stability.
This includes the core agent, layout controls, viewing code diffs, and more. We will be slowly rolling these updates out over the week, ensuring there are no regressions during your holiday coding.Layout customization
It's now easier to customize your default layout across workspaces.
Original source Report a problem
We've included four default layouts: agent, editor, zen, and browser. You can use Command (⌘) + Option (⌥) + Tab (⇥) to switch between layouts, or easily jump between different workspaces. Additionally, you can move backwards in this list by including Shift (⇧), similar to macOS. - Dec 18, 2025
- Parsed from source:Dec 18, 2025
- Detected by Releasebot:Dec 23, 2025
Enterprise Insights, Billing Groups, Service Accounts, and Improved Security Controls
Cursor for Enterprise rolls out new features that boost visibility, security, and collaboration. It adds conversation insights to categorize work, share read-only transcripts, and map usage to billing groups. New Linux sandboxing and service accounts enable safer automated workflows across teams.
Conversation insights
Cursor can now analyze the code and context in each agent session to understand the type of work that is being done, including:
- Category: Bug fixes, refactoring, explanation
- Work Type: Maintenance, bug fixing, new features
- Complexity: Difficulty and specificity of prompts
Enterprise customers can also extend these categories across their organization and teams. We protect your privacy by ensuring no PII or sensitive data is collected as part of these insights.
Shared agent transcripts
You can now share agent conversations with your team.
Generate a read-only transcript of any agent conversation to include in your PRs or internal documentation. Transcripts can be forked so others can start new agent conversations from the same context.
Billing groups
Cursor now supports billing groups for fine-grained visibility into where usage occurs.
Map usage and spend to the structure of your organization. Track spend by group, set budget alerts, and keep an eye on outliers. Understand which teams have the highest adoption of Cursor.
Linux sandboxing for agents
Sandboxing for agents supports Linux in addition to macOS.
This allows agents to work effectively within appropriate boundaries. Access is scoped to your workspace and can be configured to block unauthorized network and filesystem access.
Learn more about LLM safety and controls.
Service accounts
Service accounts are non-human accounts (and their API keys) that can configure Cursor, call APIs, and invoke cloud agents.
With service accounts, teams can securely automate Cursor-powered workflows without tying integrations to individual developers' accounts. This makes it easier to manage access, rotate credentials, and keep automations running even as people and roles change.
Service accounts will roll out to Enterprise accounts starting the week of 12/22.
Learn more about Cursor for Enterprise and talk to our team to learn more.
Original source Report a problem - Dec 11, 2025
- Parsed from source:Dec 11, 2025
- Detected by Releasebot:Dec 12, 2025
A visual editor for the Cursor Browser
Cursor Browser introduces a visual editor that blends design and code in one window. Drag and drop layout, live prop editing, and describe-and-prompt workflows let you adjust UI and apply changes instantly.
We're excited to release a visual editor for the Cursor Browser. It brings together your web app, codebase, and powerful visual editing tools, all in the same window. You can drag elements around, inspect components and props directly, and describe changes while pointing and clicking. Now, interfaces in your product are more immediate and intuitive, closing the gap between design and code.
Rearrange with drag-and-drop
The visual editor lets you manipulate a site's layout and structure directly by dragging and dropping rendered elements across the DOM tree.
This unifies visual design with coding. You can swap the order of buttons, rotate sections, and test different grid configurations without ever context-switching. Once the visual design matches what you had in mind, tell the agent to apply it. The agent will locate the relevant components and update the underlying code for you.
Test component states directly
Many modern apps are built in React, where components have properties to control different component states. The visual editor makes it easy to surface these props in the sidebar so you can make changes across different variants of a component.
Adjust properties with visual controls
The visual editor sidebar lets you fine-tune styles with sliders, palettes, and your own color tokens and design system. Every tweak is fully interactive: live color pickers that preview your choices, as well as controls to rearrange grids, flexbox layouts, and typography with precision.
Point and prompt
The visual editor also lets you click on anything in your interface and describe what you want to change. You could click on one element and say, "make this bigger," while on another you prompt, "turn this red," and on a third you type, "swap their order." The agents run in parallel, and within seconds your changes are live.
Up the abstraction hierarchy
Cursor's new visual editor unifies your work across design and code, helping you better articulate what you want so that execution isn't limited by mechanics.
We see a future where agents are even more deeply connected to building apps on the web, and humans express their ideas through interfaces that connect thought to code more directly. These features are a step in that direction.
Read the Browser docs. Learn about all the new features in Cursor 2.2.
Original source Report a problem - Dec 10, 2025
- Parsed from source:Dec 10, 2025
- Detected by Releasebot:Dec 11, 2025
Introducing Debug Mode: Agents with runtime logs
Cursor 2.2 adds Debug Mode a human in the loop debugging workflow that builds hypotheses instruments runtime logs reproduces bugs and verifies fixes with you This turns tough issues into targeted changes and clean ship ready code
Describe the bug
Coding agents are great at lots of things, but some bugs consistently stump them. That's why we're introducing Debug Mode, an entirely new agent loop built around runtime information and human verification.
To build it, we examined the practices of the best debuggers on our team. We rolled their workflows into an agent mode, equipping it with tools to instrument code with runtime logs, prompts that generate multiple hypotheses about what's going wrong, and the ability to call back to you to reproduce the issue and verify fixes.
The result is an interactive process that reliably fixes bugs that were previously beyond the reach of even the smartest models working alone, or could take significant developer time to address.
Reproduce the bug
Describe the bug
To get started, select Debug Mode from the dropdown menu and describe the bug in as much detail as you can.
Instead of immediately trying to generate a fix, the agent reads through your codebase and generates multiple hypotheses about what could be wrong. Some will be ideas you would have thought of on your own, but others will likely be approaches you wouldn't have considered.
The agent then instruments your code with logging statements designed to test these hypotheses. This prepares the agent to receive concrete data about what's actually happening when the bug occurs.
Verify the fix
Reproduce the bug
Next, go to your application and reproduce the bug while the agent collects the runtime logs.
The agent can see exactly what's happening in your code when the bug occurs: variable states, execution paths, timing information. With this data, it can pinpoint the root cause and generate a targeted fix. Often that's a precise two or three line modification instead of the hundreds of lines of speculative code you'd have received with a standard agent interaction.
Verify the fix
At this point, Debug Mode asks you to reproduce the bug one more time with the proposed fix in place. If the bug is gone, you mark it as fixed and the agent removes all the instrumentation, leaving you with a clean, minimal change you can ship.
This human-in-the-loop verification is critical. Sometimes bugs are obvious, but other times they fall into a gray area where the fix might work technically but not feel right. The agent can't make that call on its own. If you don't think the bug is fixed, the agent adds more logging, you reproduce again, and it refines its approach until the problem is actually solved.
This kind of tight back-and-forth is one way we think AI coding works best. The agent handles the tedious work while you make the quick decisions that need human judgment. The result with Debug Mode is that tricky bugs that used to be out of reach are now reliably fixed.
Read the Debug Mode docs. Learn about all the new features in Cursor 2.2.
Original source Report a problem - Dec 10, 2025
- Parsed from source:Dec 10, 2025
- Detected by Releasebot:Dec 11, 2025
- Modified by Releasebot:Dec 12, 2025
2.2
New debugging and design features roll out: Debug Mode adds runtime logs across stacks for root-cause analysis. Browser layout and style editor lets you design in real time with agent-driven updates. Plan Mode adds Mermaid diagrams, parallel agent reviews, and pinned chats for quick reference.
Debug Mode
Debug Mode helps you reproduce and fix the most tricky bugs.
Cursor instruments your app with runtime logs to find the root cause. It works across stacks, languages, and models.
Read more in our announcement.Browser layout and style editor
Design and code simultaneously with a brand new browser sidebar and component tree.
Move elements, update colors, test layouts, and experiment with CSS in real time, then instantly apply changes to your codebase using agent. You can also click on multiple elements and describe changes in text to kick off an agent to make visual changes.
Read more in our announcement.Plan Mode and Parallel Agents
Plan Mode now supports inline Mermaid diagrams, allowing the agent to automatically generate and stream visuals into your plans. You also have more control over how you build them, with the option to send selected to-dos to new agents.
When running multiple agents in parallel, Cursor will now automatically evaluate all runs and give a recommendation for the best solution.
The selected agent will have a comment explaining why it was picked. Judging of the best solution only happens after all parallel agents have finished.In the agent sidebar
In the agent sidebar, pin chats at the top for future reference.
Original source Report a problem - Dec 4, 2025
- Parsed from source:Dec 4, 2025
- Detected by Releasebot:Dec 5, 2025
Improving Cursor’s agent for OpenAI Codex models
Cursor updates its agent harness to support OpenAI Codex frontier model GPT-5.1-Codex-Max with improved tool calls and shell‑oriented workflows. The update preserves reasoning traces, enhances lint checks, and tightens safety so code edits are faster and more reliable.
Building a robust agent harness
Cursor integrates with all frontier AI models for coding.
Each model requires specific instructions and tweaks to our agent harness to improve output quality, prevent laziness, efficiently call tools, and more.
We’ve been partnering with OpenAI to make their models available to developers with Cursor’s agent. This post will cover how we’ve updated our agent harness to support their latest frontier coding model GPT-5.1-Codex-Max.Every model in Cursor’s agent harness has specific instructions and tools made available to optimize that model inside the Cursor environment.
AI labs train new models on a variety of different instructions and tools. In specific domains like coding, models may favor patterns that are more similar to what they’ve already seen in training. When adding new models into Cursor, our job is to integrate familiar instructions and tools alongside Cursor-specific ones, and then tune them based on Cursor Bench, our internal suite of evals.
We measure the quality and robustness of models based on their success rate, ability to call tools, and overall adoption across users. Here are some of the updates we made to our agent harness for Codex.Updating for the latest Codex model
OpenAI’s Codex models are versions of their latest frontier model, trained specifically for agentic coding.
The OpenAI team collaborated closely with us to align the tools and prompts with the Codex CLI harness. Here are some of the changes we’ve made:A more shell-forward approach
OpenAI’s Codex CLI is focused on shell-oriented workflows. As a result, the Codex model receives a limited set of tools during training and learns instead to use the shell to search, read files, and make edits.
If the model is struggling with a difficult edit, it sometimes falls back to writing files using an inline Python script. These scripts are powerful, but tool calling is both safer and a better user experience for edits in Cursor.
To encourage tool calling, we made the names and definitions of tools in Cursor closer to their shell equivalents like rg (ripgrep). We made this change for all models in our harness. We also added instructions like:
If a tool exists for an action, prefer to use the tool
instead of shell commands (e.g. read_file overcat).
Sandboxing in Cursor, which prevents unauthorized file access and network activity without requiring users to manually approve every command, also helps improve security here if the model does still choose to run a shell command.Preambles
Unlike the mainline GPT-5 series of models, the Codex model family currently uses reasoning summaries to communicate user updates as it’s working. These can be in the form of one-line headings or a full message.
For these reasoning summaries, we wanted to strike a balance that would let users follow along with the agent’s progress and identify bad trajectories early, without spamming them to the point that they tune out. We gave the model guidelines to limit reasoning summaries to 1 or 2 sentences, note when discovering new information or initiating a new tactic, and to avoid commenting on its own communication (“I’m explaining to the user…”).
Since Codex models cannot “talk” normally until the end of an agent turn, we removed all language in the prompt related to communicating with the user mid-turn. We found that this improved the performance of the model’s final code output.Reading lints
Cursor makes tools available to all models in our harness for reading linter errors (e.g. ESLint, Biome) and allowing the agent to automatically fix them.
We found that providing Codex with the tool definition alone is not enough to make it inclined to call our read_lints tool. Instead, Codex benefits significantly from clear and literal instructions for when to call it:
After substantive edits, use the read_lints tool to check
recently edited files for linter errors. If you've introduced
any, fix them if you can easily figure out how.Preserving reasoning traces
OpenAI’s reasoning models emit internal reasoning traces between tool calls, which are effectively a “chain of thought” explaining why the model chooses each action. The Responses API is designed to capture and pass along these reasoning items (or encrypted reasoning items in sensitive contexts) so the model can maintain continuity across turns rather than having to reconstruct its plan from scratch.
Codex is especially dependent on this continuity. When reasoning traces are dropped, the model has to infer its previous thought process, which often leads to lost subgoals, degraded planning, misordered tool calls, or repeatedly re-deriving earlier steps. In our Cursor Bench experiments, removing reasoning traces from GPT-5-Codex caused a 30% performance drop. In comparison, OpenAI observed a smaller 3% degradation for GPT-5 on SWE-bench when reasoning traces were omitted.
Given the scale of that impact, we added alerting to ensure that reasoning traces are always preserved and forwarded correctly. This keeps the agent’s internal plan intact and prevent the performance regressions that occur when models are forced to “fill in the blanks” between tool calls.Biasing the model to take action
In Cursor's default agent mode, you want the agent to autonomously read and edit files based on the user request. It can be frustrating when you tab away only to find that the agent was waiting to ask for your permission to proceed.
We’ve been experimenting with more specific instructions to help guide Codex:
Unless the user explicitly asks for a plan or some other intent that
makes it clear that code should not be written, assume the user wants
you to make code changes or run tools to solve the user's problem. In
these cases, it's bad to output your proposed solution in a message, you
should go ahead and actually implement the change. If you encounter
challenges or blockers, you should attempt to resolve them yourself.
In Cloud Agents, our async remote workflow, we make this language even stronger.Message ordering
OpenAI models are trained to respect and prioritize message ordering. For example, the system prompt always takes precedence over user messages and tool results.
While this is helpful, it means we need to tune our harnesses to ensure the Cursor-provided prompt does not include instructions which could accidentally contradict user messages. Otherwise, Codex could get into a state where it does not want to comply with the user request.
For example, at one point we told Codex that it should take care to preserve tokens and not be wasteful. But we noticed that this message was impacting the model’s willingness to perform more ambitious tasks or large-scale explorations. Sometimes it would stop and stubbornly say, I’m not supposed to waste tokens, and I don’t think it’s worth continuing with this task!Looking forward
The pace of model releases is increasing. Our goal is to get the most out of every frontier model inside the Cursor agent harness. There’s more work to be done, and we’ll continue to share improvements we’re making to Cursor.
Original source Report a problem - Nov 21, 2025
- Parsed from source:Nov 21, 2025
- Detected by Releasebot:Sep 29, 2025
- Modified by Releasebot:Dec 23, 2025
2.1
Cursor adds AI code reviews in editor, enhanced Plan Mode with clarifying questions and an interactive UI, plus instant grep across the codebase. Quick plan adjustments with ⌘+F search improve plan quality while Bugbot support and broader integrations roll out to 2.1 users.
Improved Plan Mode, AI Code Review in Editor, and Instant Grep
Improved Plan Mode
When creating plans, Cursor responds with clarifying questions to improve the plan quality. Cursor now shows an interactive UI to easily answer questions.
You can also ⌘+F to search inside generated plans.AI Code Reviews
You can now find and fix bugs directly in Cursor with AI code reviews. It will look at your changes and find issues which you can see in the sidepanel.
This is in addition to Bugbot, which runs on your source control provider like GitHub (including Enterprise Server), GitLab, and more.Instant Grep (Beta)
All grep commands run by the agent are now instant.
Instant grep is supported by all models in Cursor. It is also used when manually searching the codebase from the sidebar, including regexes and matching on word boundaries.
This improvement is slowly rolling out to 2.1 users over the next week.Improvements (11)
Enterprise (3)
Integrations (4)
Original source Report a problem - Nov 6, 2025
- Parsed from source:Nov 6, 2025
- Detected by Releasebot:Nov 18, 2025
Improving agent with semantic search
Cursor's agent now blends semantic search with traditional grep, powered by a custom embedding model, to speed up understanding of large codebases and boost answer accuracy. Benchmarks show improved code retention up to 2.6% and fewer follow ups, delivering a smoother coding experience.
When coding agents receive a prompt, returning the right answer requires building an understanding of the codebase by reading files and searching for relevant information.
One tool Cursor’s agent uses is semantic search, which retrieves segments of code matching natural language queries, such as “where do we handle authentication?”, in addition to the regex-based searching provided by a tool like grep.
To support semantic search, we’ve trained our own embedding model and built indexing pipelines for fast retrieval. While you could rely exclusively on grep and similar command-line tools for search, we've found that semantic search significantly improves agent performance, especially over large codebases:- Achieving on average 12.5% higher accuracy in answering questions (6.5%–23.5% depending on the model).
- Producing code changes that are more likely to be retained in codebases.
- Requiring fewer iterations for users to arrive at a correct solution.
- Increasing accuracy across all models we tested, including all frontier coding models.
Offline evals
We maintain an evaluation dataset, Cursor Context Bench, focused on retrieving information in codebases with known correct answers. This evaluation is run over all of the most-used models in Cursor, including our own Composer.
The comparison looks at performance with two sets of available tools: one that includes semantic search and one that does not. In every configuration, semantic search significantly improves outcomes.Online A/B tests
We also wanted to understand the impact on the end-user experience. We ran an A/B test where both groups used the same model, but one group's agent had access to semantic search while the other relied solely on traditional search tools like grep. We looked at two metrics:- Code Retention: Code written by effective agents is more likely to remain in user codebases. We see agent code retention increases by 0.3% when semantic search is available. This effect increases to 2.6% on large codebases with 1,000 files or more.
- Dissatisfied User Requests: Code written by effective agents requires no follow-ups or corrections. We observed a 2.2% increase in dissatisfied follow-up user requests when semantic search was not available.
The effect size is lower here since the A/B test is on all agent queries and not all requests require search.
Custom retrieval models
One piece that enables these results is our custom embedding model. Our approach uses agent sessions as training data: when an agent works through a task, it performs multiple searches and opens files before finding the right code. By analyzing these traces, we can see in retrospect what should have been retrieved earlier in the conversation.
We provide these traces to an LLM, which ranks what content would have been most helpful at each step. We then train our embedding model to align its similarity scores with these LLM-generated rankings. This creates a feedback loop where the model can learn from how agents actually work through coding tasks, rather than relying on generic code similarity.Conclusion
Original source Report a problem
Semantic search is currently necessary to achieve the best results, especially in large codebases.
Our agent makes heavy use of grep as well as semantic search, and the combination of these two leads to the best outcomes. We’re continuing to test and evaluate all tools we give to the agent harness as models improve. - Oct 31, 2025
- Parsed from source:Oct 31, 2025
- Detected by Releasebot:Nov 18, 2025
Introducing Cursor for Enterprise
Cursor unveils Hooks to observe, control, and extend agent loops, plus Team Rules for consistent development, upgraded analytics, Audit Log visibility, and Sandbox Mode for safer iteration. Enterprise-ready features ship today to boost productivity and governance.
Cursor is used by tens of thousands of enterprises, including Salesforce, NVIDIA, and PwC, to accelerate product velocity and build durable software.
To further support enterprise teams, we’re launching new features that make Cursor’s coding agents more capable, transparent, and secure.Hooks
Hooks allow you observe, control, and extend the agent loop using custom scripts.
{ "version" : 1 , "hooks" : { "beforeSubmitPrompt" : [ { "command" : "./audit.sh" } ], "beforeShellCommand" : [ { "command" : "./audit.sh" }, { "command" : "./allowlist.sh" } ] } }Integrating Cursor Hooks with our engineering intelligence platform
gives us a better sense of how AI is affecting developer productivity throughout the software development lifecycle, and gives engineering leaders the insights they need to make smarter decisions.
Rebecca Fitzhugh
Lead Principal Engineer, AtlassianHooks can
- Add observability: Log agent actions, tool calls, prompts, and completions for future analysis.
- Control the full agent loop: Enforce compliance policies, block unapproved commands, and scrub secrets or proprietary code in real time.
- Extend Cursor with code: Connect external systems, inject context, or trigger automations.
Hooks can be distributed through MDM or with Cursor’s cloud option.
Team Rules
Team Rules bring shared context and best practices to every developer in your organization.
Team Rules make it easy for our agents to stay aligned with how we build at Duolingo. They guide developers toward the right patterns automatically, so we spend less time fixing and more time shipping.
Sarah Deitke
Software Engineer, DuolingoRules can standardize API schemas, enforce conventions, or teach common workflows. Admins can choose to recommend or require rules from the cloud dashboard.
Upgraded Analytics
We’ve rebuilt our analytics to give leaders data about how their teams use AI:
- See daily activity and top users at a glance
- Data on CLI, Background Agent, and Bugbot adoption
- View percentage of AI lines of code on the commit level
- Filter data by Active Directory group
- Exportable replication data per visualization (via API or CSV)
- Data updates every two minutes instead of every 24 hours
The new usage leaderboard has already helped us surface power users and learn how they work. The cleaner dashboard makes the data easier to trust and act on.
Nicolas Arkhipenko
Senior Director of AI Platforms & Developer Productivity, SalesforceThe upgraded dashboard is generally available today, with data from August 27th for all clients 1.7 and later.
Audit Log
Audit Log gives administrators full visibility into every key event on the platform from security changes to rule updates.
It currently tracks 19 event types covering access, asset edits, and configuration updates.
Audit Log data is available in the web dashboard and can be exported as a CSV.Sandbox Mode
Sandbox Mode executes agent terminal commands in a restricted environment to enable faster and safer iteration.
By default, the sandbox blocks network access, limiting file access to the workspace and /tmp/ .
If a command fails due to restrictions, the user can skip it or choose to re-run outside the sandbox.
Enterprise Admins have control over sandbox availability and team-wide git and network access.Working with our team
We want to make Cursor the best place for teams building with AI. Many of these features were inspired by the work we're already doing with many of the most innovative companies in the world, and we’d like to build with yours.
Original source Report a problem