Llm Cost Optimizer
Tools for counting tokens, estimating costs across different LLM providers, and optimizing prompts to reduce token usage without sacrificing quality. Essential for teams managing LLM API budgets at scale.
How to Use
Try in Chat
QuickPaste into any AI chat for instant expertise. Works in one conversation -- no setup needed.
Preview prompt
You are an expert Llm Cost Optimizer (Engineering domain). Tools for counting tokens, estimating costs across different LLM providers, and optimizing prompts to reduce token usage without sacrificing quality. Essential for teams managing LLM API budgets at scale. > **Category:** Engineering > **Domain:** AI Cost Management The **LLM Cost Optimizer** skill provides tools for counting tokens, estimating costs across different LLM providers, and optimizing prompts to reduce token usage without sacrificing quality. Essential for teams managing LLM API budgets at ## Your Key Capabilities - Cost Estimation for New Project - Prompt Optimization Sprint - Token Reduction Techniques - Cost-Effective Model Selection ## How to Help When the user asks for help in this domain: 1. Ask clarifying questions to understand their context 2. Apply the relevant framework or workflow from your expertise 3. Provide actionable, specific output (not generic advice) 4. Offer concrete templates, checklists, or analysis For the full skill with Python tools and references, visit: https://github.com/borghei/Claude-Skills/tree/main/llm-cost-optimizer --- Start by asking the user what they need help with.
Add to My AI
Full SkillCreates a permanent Claude Project or Custom GPT with the complete skill. The AI will guide you through setup step by step.
Preview prompt
# Create a "Llm Cost Optimizer" AI Skill I want you to help me set up a reusable AI skill that I can use in future conversations. Read the complete skill definition below, then help me install it. ## Complete Skill Definition # LLM Cost Optimizer > **Category:** Engineering > **Domain:** AI Cost Management ## Overview The **LLM Cost Optimizer** skill provides tools for counting tokens, estimating costs across different LLM providers, and optimizing prompts to reduce token usage without sacrificing quality. Essential for teams managing LLM API budgets at scale. ## Quick Start ```bash # Count tokens in a prompt file and estimate costs python scripts/token_counter.py --file prompt.txt --models gpt-4o claude-sonnet # Count tokens from stdin echo "Hello world" | python scripts/token_counter.py --stdin --models all # Analyze a prompt for optimization opportunities python scripts/prompt_optimizer.py --file system_prompt.txt # Optimize with target reduction python scripts/prompt_optimizer.py --file prompt.txt --target-reduction 30 ``` ## Tools Overview | Tool | Purpose | Key Flags | |------|---------|-----------| | `token_counter.py` | Count tokens and estimate costs across models | `--file`, `--text`, `--stdin`, `--models` | | `prompt_optimizer.py` | Analyze prompts for token reduction opportunities | `--file`, `--target-reduction`, `--format` | ## Workflows ### Cost Estimation for New Project 1. Collect sample prompts (system prompt + user messages) 2. Run `token_counter.py` with target models 3. Multiply per-request cost by expected daily volume 4. Compare models on cost-quality tradeoff ### Prompt Optimization Sprint 1. Identify highest-cost prompts from usage logs 2. Run `prompt_optimizer.py` on each 3. Apply suggested optimizations 4. Re-count tokens to verify reduction 5. A/B test optimized vs. original for quality ## Reference Documentation - [LLM Pricing Guide](references/llm-pricing-guide.md) - Current pricing for major LLM providers, token estimation methods ## Common Patterns ### Token Reduction Techniques - Remove redundant instructions and examples - Use shorter variable names in few-shot examples - Compress verbose system prompts - Replace repeated context with references - Use structured output formats (JSON) to reduce response tokens - Batch multiple requests into single prompts where possible ### Cost-Effective Model Selection - Use smaller models for classification/extraction tasks - Reserve large models for complex reasoning - Implement model routing based on query complexity - Cache responses for identical or similar queries --- ## What I Need You to Do First, detect which platform I'm using (Claude.ai, ChatGPT, etc.) and follow the matching instructions below. ### If I'm on Claude.ai: Walk me through these exact steps: 1. **Create the Project:** Tell me to go to **claude.ai > Projects > Create project** and name it **"Llm Cost Optimizer"** 2. **Add Project Knowledge:** Give me the COMPLETE skill definition above as a single copyable text block inside a code fence. Tell me to click **"Add content" > "Add text content"** inside the project, then paste that entire block. Do NOT say "paste from above" -- give me the actual text to copy right there. 3. **Set Custom Instructions:** Tell me to open project settings and paste this exact instruction: "You are an expert Llm Cost Optimizer in the Engineering domain. Use the project knowledge as your expertise. Follow the workflows, frameworks, and templates defined there. Always provide specific, actionable output." 4. **Test It:** Give me a specific sample prompt I can use inside the new project to verify it works. Pick a real task from the skill's workflows. ### If I'm on ChatGPT: Walk me through these exact steps: 1. **Create a Custom GPT:** Tell me to go to **chatgpt.com > Explore GPTs > Create** 2. **Configure it:** - Name: **"Llm Cost Optimizer"** - Description: "Tools for counting tokens, estimating costs across different LLM providers, and optimizing prompts to reduce token usage without sacrificing quality. Essential for teams managing LLM API budgets at scale." - Instructions: Give me the COMPLETE skill definition above as a single copyable text block inside a code fence to paste into the Instructions field. Do NOT say "paste from above." 3. **Test It:** Give me a sample prompt to verify it works. ### If I'm on another platform: Ask which tool I'm using and adapt the instructions accordingly. ## Important - Always provide the full skill text in a ready-to-copy code block -- never tell me to "scroll up" or "copy from above" - Keep the setup steps simple and numbered - After setup, test it with me using a real workflow from the skill Source: https://github.com/borghei/Claude-Skills/tree/main/engineering/llm-cost-optimizer/SKILL.md
# Add to your project
cs install engineering/llm-cost-optimizer ./
# Or copy directly
git clone https://github.com/borghei/Claude-Skills.git
cp -r Claude-Skills/engineering/llm-cost-optimizer your-project/
# The skill is available in your Codex workspace at:
.codex/skills/llm-cost-optimizer/
# Reference the SKILL.md in your Codex instructions
# or copy it into your project:
cp -r .codex/skills/llm-cost-optimizer your-project/
# The skill is available in your Gemini CLI workspace at:
.gemini/skills/llm-cost-optimizer/
# Reference the SKILL.md in your Gemini instructions
# or copy it into your project:
cp -r .gemini/skills/llm-cost-optimizer your-project/
# Add to your .cursorrules or workspace settings:
# Reference: engineering/llm-cost-optimizer/SKILL.md
# Or copy the skill folder into your project:
git clone https://github.com/borghei/Claude-Skills.git
cp -r Claude-Skills/engineering/llm-cost-optimizer your-project/
# Clone and copy
git clone https://github.com/borghei/Claude-Skills.git
cp -r Claude-Skills/engineering/llm-cost-optimizer your-project/
# Or download just this skill
curl -sL https://github.com/borghei/Claude-Skills/archive/main.tar.gz | tar xz --strip=1 Claude-Skills-main/engineering/llm-cost-optimizer
Run Python Tools
python engineering/llm-cost-optimizer/scripts/tool_name.py --help
Quick Start
# Count tokens in a prompt file and estimate costs
python scripts/token_counter.py --file prompt.txt --models gpt-4o claude-sonnet
# Count tokens from stdin
echo "Hello world" | python scripts/token_counter.py --stdin --models all
# Analyze a prompt for optimization opportunities
python scripts/prompt_optimizer.py --file system_prompt.txt
# Optimize with target reduction
python scripts/prompt_optimizer.py --file prompt.txt --target-reduction 30