返回 Feed

Most of my Claude usage was on work that didn't need Claude. Cut my bill 60x on bulk tasks with a tiny side model.

Reddit·r/ClaudeAI(熱門)·u/petburiraja·10 小時前
查看原文

AI 分析摘要

8AI 工作流

這是一個使用者分享如何使用小型模型(DeepSeek V4 Flash)來處理機械性工作,從而減少 Claude 的使用費用。這個方法可以將費用降低 60 倍,且有實際的使用案例和 GitHub Repo 提供。讀者可以透過這個方法來優化自己的工作流程,減少不必要的費用。

💡 評分理由:這個分享提供了具體可用的工作流程和實際的使用案例,且有實質的功能變化和費用優化的效果。

模型:meta/llama-3.3-70b-instruct · Prompt: v2

原文內容(清洗後)

Most of my Claude usage was on work that didn't need Claude. Cut my bill 60x on bulk tasks with a tiny side model. I looked at what was actually eating my Claude usage and it was embarrassing. Classifying files. Reformatting json. Pulling fields out of text. Summarizing docs I was going to skim anyway. None of that needed Sonnet. All of it cost the same as the work that did. Tried the obvious fixes first. Switching to Haiku for simple stuff (still wasteful at volume). Tighter prompts (helps a little). /compact (delays the problem). None of it changed the shape of the spend. What actually worked: a small cheap model running as a side worker, with one rule in CLAUDE.md telling Claude not to do the mechanical stuff itself. The setup is one tool. Send it text, get text back. Claude calls it for the bounded mechanical work I'd review anyway. Default model is DeepSeek V4 Flash because it's cheap and has 1M context, but the endpoint is one config line and works with anything openai-compatible (local ollama, vllm, lm studio). **3 weeks of real usage:** - 217 mechanical calls offloaded - DeepSeek total spend: $0.41 - Same workload on Sonnet would have been roughly $7 The CLAUDE.md rule that actually works is negative framing. Not "use deepseek for X" but "do NOT use Claude for: json formatting, field extraction, file classification, summarization you will review anyway." Positive framing got ignored maybe 30% of the time. Deny list catches it. It's a supervised worker, not an agent. No tool calls, no file access, no chains. Latency 3-25s. You review the output. That's the whole shape. Repo with setup steps: https://github.com/arizen-dev/deepseek-mcp (MIT, Python 3.10+) Happy to answer questions about the routing rules or the model choice.