Skip to content

For the complete documentation index, see llms.txt. Markdown versions of all docs pages are available by appending .md to any docs URL.

Page as Markdown

Windsurf

Configure Windsurf AI code editor to use agentgateway

Configure Windsurf, the AI code editor by Codeium, to route requests to your LLM through your agentgateway proxy.

Before you begin

  1. Install the agentgateway binary.
  2. Install Windsurf.

Example agentgateway configuration

cat > /tmp/test-windsurf.yaml << 'EOF'
# yaml-language-server: $schema=https://agentgateway.dev/schema/config
llm:
  port: 3000
  models:
  - name: "*"
    provider: openAI
    params:
      apiKey: "$OPENAI_API_KEY"
EOF

Configure Windsurf

Configure Windsurf to route LLM requests through agentgateway. For more information, review the Windsurf documentation.

  1. Open Windsurf Settings.

    • macOS: Cmd + , or Windsurf > Settings
    • Windows/Linux: Ctrl + , or File > Preferences > Settings
  2. Search for Http: Proxy.

  3. Enter your agentgateway URL.

    http://localhost:3000
  4. Save the settings.

Verify the connection

  1. Open the Windsurf chat panel.
  2. Send a message such as “test”.
  3. Windsurf responds through your agentgateway backend.
Was this page helpful?
Agentgateway assistant

Ask me anything about agentgateway configuration, features, or usage.

Note: AI-generated content might contain errors; please verify and test all returned information.

Tip: one topic per conversation gives the best results. Use the + button in the chat header to start a new conversation.

Switching topics? Starting a new conversation improves accuracy.
↑↓ navigate select esc dismiss

What could be improved?

Your feedback helps us improve assistant answers and identify docs gaps we should fix.

Need more help? Join us on Discord: https://discord.gg/y9efgEmppm

Want to use your own agent? Add the Solo MCP server to query our docs directly. Get started here: https://search.solo.io/.