All posts
2026-04-049 min

Anthropic Bans Third-Party Agents: How to Keep OpenClaw Running Without Interruption

OpenClawAnthropicOllamaLocal LLMsModel SwitchingSelf-Hosting

What Just Happened

If you woke up today and your OpenClaw agent stopped responding, you're not alone.

Anthropic quietly dropped a policy change effective April 4, 2026: Claude subscription accounts can no longer be used through third-party tools that place "outsized strain" on the API. The official framing is about infrastructure load — repeated agentic loops, automated multi-turn calls, 10x the usage of a typical subscriber — but the practical result is clear: OpenClaw on a Claude subscription is now in a grey zone at best, and actively broken for many users.

X is on fire about this right now. The sentiment ranges from "expected this" to "I'm switching everything today."

Here's the good news: OpenClaw was designed exactly for this moment. The gateway abstracts your model provider. Switching costs you 30 minutes, not a week of rewrites.

---

What's Actually Blocked (And What Isn't)

To be precise about what changed:

  • Blocked: Using Claude via a subscription (claude.ai account) through third-party tools and API wrappers
  • Not blocked: Using Claude via the **direct API** with your own API key (pay-as-you-go)
  • Not blocked: Using any other model entirely — Ollama, Gemini, Mistral, Qwen, GPT-4o, etc.
  • If you've been running OpenClaw with your Claude API key (not a subscription), you may be fine for now. But the direction of travel is clear. Many people are using this as a forcing function to go fully local or diversify away from a single provider.

    ---

    Option 1: Switch to Ollama (Fully Local, Free)

    This is the most popular move happening right now. If you have a machine with decent RAM (16GB+), you can run a capable model locally with zero ongoing cost.

    Step 1: Install Ollama

    ```bash

    curl -fsSL https://ollama.ai/install.sh | sh

    ```

    Step 2: Pull a model

    For a good balance of speed and capability on consumer hardware:

    ```bash

    # Good all-rounder, runs on 16GB RAM

    ollama pull qwen2.5:14b

    # Smaller, faster — 8GB RAM is enough

    ollama pull qwen2.5:7b

    # If you have a beefy machine (32GB+)

    ollama pull qwen2.5:32b

    ```

    Qwen 2.5 is the current community favorite for agentic tasks. It follows instructions well, handles code, and plays nicely with tool calls.

    Step 3: Update OpenClaw config

    Open your OpenClaw config (usually `~/.openclaw/config.yaml` or wherever your gateway config lives):

    ```yaml

    model:

    default: ollama/qwen2.5:14b

    baseUrl: http://localhost:11434/v1

    ```

    Restart OpenClaw:

    ```bash

    openclaw gateway restart

    ```

    That's it. Your agent is now running 100% locally, zero API costs, zero Anthropic dependency.

    ---

    Option 2: Keep Claude via Direct API Key

    If you have a Claude API key (not a subscription), you can keep using it — at least for now. The restriction is specifically on subscription accounts being relayed through third-party tools.

    Check your current config:

    ```bash

    grep -i anthropic ~/.openclaw/config.yaml

    grep -i claude ~/.openclaw/config.yaml

    ```

    If you see an API key configured (not OAuth/subscription), you're likely unaffected. Just monitor your API console for any rate limit changes.

    If you want to add a direct API key:

    ```yaml

    model:

    default: anthropic/claude-sonnet-4-5

    apiKey: ${ANTHROPIC_API_KEY}

    ```

    Then set your env var in `.env`:

    ```

    ANTHROPIC_API_KEY=sk-ant-api03-...

    ```

    ---

    Option 3: Google Gemini (Generous Free Tier)

    Gemini 2.0 Flash has a surprisingly good free tier and handles agentic workloads well. If you want to stay cloud-based but reduce Anthropic exposure:

    ```yaml

    model:

    default: google/gemini-2.0-flash

    apiKey: ${GOOGLE_API_KEY}

    ```

    Get a key at [aistudio.google.com](https://aistudio.google.com). The free tier gives you 1 million tokens/day on Flash — more than enough for most personal agent setups.

    ---

    Option 4: OpenRouter (Multi-Provider Fallback)

    OpenRouter is a proxy that gives you access to 50+ models through one API. You load credits once, and you can switch models without touching your config:

    ```yaml

    model:

    default: openrouter/anthropic/claude-sonnet-4-5

    baseUrl: https://openrouter.ai/api/v1

    apiKey: ${OPENROUTER_API_KEY}

    ```

    The benefit: if one provider goes down or gets expensive, you change one line and you're on a different model. No credential rotation, no gateway restart (just a config reload).

    ---

    Running Multiple Models Per Agent

    One underused OpenClaw feature: you can assign different models to different agents. Your main orchestrator can use one model, sub-agents can use cheaper models for execution tasks.

    In your agent config:

    ```yaml

    agents:

    - id: main

    model: ollama/qwen2.5:14b # local, free

    - id: coder

    model: google/gemini-flash # fast, generous free tier

    - id: researcher

    model: openrouter/mistral/mistral-large # good at synthesis

    ```

    This is the architecture that's resilient to any single provider outage or policy change. When Anthropic made this move, people running multi-provider setups didn't even notice.

    ---

    What About Skill Compatibility?

    Skills in OpenClaw are model-agnostic. They're just instructions and tool definitions — they don't know or care which LLM is executing them.

    The one thing to watch: tool calling support. Some smaller local models don't handle tool calls reliably. The safe list for agentic tasks as of April 2026:

    | Model | Tool Calling | Notes |

    |-------|-------------|-------|

    | Qwen 2.5 (any size) | ✅ Excellent | Best local option |

    | Llama 3.3 70B | ✅ Good | Needs 40GB RAM |

    | Gemini 2.0 Flash | ✅ Excellent | Cloud, fast free tier |

    | Mistral Large | ✅ Good | Cloud, good at reasoning |

    | Phi-4 | ⚠️ Inconsistent | Better for simple tasks |

    | Gemma 3 27B | ✅ Good | Solid local alternative |

    ---

    The Bigger Lesson

    This is exactly why self-hosted, provider-agnostic infrastructure matters. OpenClaw gives you the abstraction layer that makes today's situation a 30-minute config change instead of a crisis.

    The people who are most frustrated right now are the ones who went deep on Anthropic-specific patterns — hardcoded Claude prompts, assuming Claude-specific behavior, building on subscription accounts instead of API keys.

    The people who are shrugging are the ones running local models or multi-provider setups. That's the bet the OpenClaw architecture makes, and today it looks like the right one.

    TL;DR: Anthropic blocked third-party agent subscriptions. If you're affected, switch to Ollama + Qwen 2.5 for local-first or Gemini Flash for a generous cloud free tier. Both work out of the box with OpenClaw. Your skills and memory files survive untouched.

    Want to learn more?

    Our playbook contains 18 detailed chapters — available in English and German.

    Get the Playbook