$ recombobulate _
home / tips / use-claude-code-to-write-and-test-prompts-for-your-own-ai-features
64

Use Claude Code to Write and Test Prompts for Your Own AI Features

recombobulate @recombobulate · Mar 29, 2026 · Workflows
use-claude-code-to-write-and-test-prompts-for-your-own-ai-features

Building AI features means writing prompts that work reliably across many inputs. Claude Code is the ideal environment for prototyping them — you can test variations, check edge cases, and iterate on the phrasing before committing anything to code.

I'm building a feature that classifies support tickets into categories. 
Help me write a system prompt and a few-shot examples that work 
reliably across these sample tickets: [paste examples]

Claude drafts the prompt, tests it against your examples, and helps you catch failure modes:

# Test the prompt against edge cases
Here are 10 tricky tickets that don't fit neatly into one category. 
Run them through the prompt we wrote and show me which ones get 
misclassified. Adjust the prompt to handle them.

# Write few-shot examples
Generate 5 few-shot examples that cover each category and include 
one ambiguous case that demonstrates how to handle uncertainty.

# Optimize for token cost
The current prompt is 800 tokens. Shorten it without losing accuracy — 
test the shorter version against the same examples to verify.

This works for all kinds of AI feature development:

# Extraction prompts
Write a prompt that extracts structured data from invoice emails — 
vendor name, amount, due date, line items. Test it against 5 real 
email formats I'll paste.

# Summarization prompts
Write a prompt that summarizes long customer feedback into 3 bullet 
points with sentiment. Make it work for both positive and negative reviews.

# Classification with confidence
Write a prompt that classifies content and includes a confidence score. 
Test how it handles ambiguous inputs and whether the scores correlate 
with actual accuracy.

Once the prompt works, ask Claude to integrate it:

The prompt is working. Now write the API integration code that sends 
user input through this prompt and handles the structured response. 
Include error handling for malformed AI responses.

AI features are only as good as their prompts — use Claude Code as a prompt lab where you can test, iterate, and perfect before shipping.

via Claude Code

~/recombobulate $ tip --comments --count=0

Log in to leave a comment.

~/recombobulate $ tip --related --limit=3
0
Run Claude Code in GitHub Actions to Automatically Review Every Pull Request

Set up Claude Code as an automated reviewer in your CI pipeline — on every pull request, it reads the diff, checks for bugs, security issues, missing tests, and convention violations, then posts its findings as a PR comment. Your human reviewers get a head start because the obvious issues are already flagged before they look.

recombobulate @recombobulate · 1 day ago
0
Ask Claude to Build a Deployment Checklist from Your Actual Infrastructure

Before deploying, tell Claude to read your project — migrations, environment variables, queue workers, scheduled tasks, caching, third-party integrations — and generate a deployment checklist that's specific to your app. Not a generic "did you run migrations?" list, but one that knows YOUR infrastructure and catches the things YOUR deploy can break.

recombobulate @recombobulate · 1 day ago
0
Ask Claude to Generate a README from Your Actual Codebase — Not a Template

Instead of writing a README from memory or copying a template, tell Claude to read your project and generate one that's actually accurate — real setup instructions from your config, real architecture from your directory structure, real API examples from your routes, and real prerequisites from your dependency files.

recombobulate @recombobulate · 1 day ago