- Split monolithic index.ts (2370 lines) into modular structure: - src/handlers/ for route handlers - src/utils.ts for shared utilities - src/config.ts for configuration - src/types.ts for TypeScript definitions - Add DB workload multiplier for smarter database resource calculation: - Heavy (analytics, logs): 0.3x multiplier - Medium-heavy (e-commerce, transactional): 0.5x - Medium (API, SaaS): 0.7x - Light (blog, portfolio): 1.0x - Fix tech_specs with realistic vcpu_per_users values (150+ technologies) - Fix "blog" matching "log" regex bug - Update documentation to reflect new architecture Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
5.6 KiB
5.6 KiB
CLAUDE.md
This file provides guidance to Claude Code (claude.ai/code) when working with code in this repository.
Project Overview
Cloudflare Worker-based AI server recommendation service. Uses OpenAI GPT-4o-mini (via AI Gateway), D1 database, KV cache, and VPS benchmark data to recommend cost-effective servers based on natural language requirements.
Production URL: https://server-recommend.kappa-d8e.workers.dev
Commands
# Development
npm run dev # Start local development server (wrangler dev)
npm run deploy # Deploy to Cloudflare Workers
npm run typecheck # TypeScript type checking
# Database operations (D1)
npx wrangler d1 execute cloud-instances-db --file=schema.sql # Apply schema
npx wrangler d1 execute cloud-instances-db --file=seed.sql # Seed data
npx wrangler d1 execute cloud-instances-db --file=fix-tech-specs.sql # Update tech specs
npx wrangler d1 execute cloud-instances-db --command="SELECT ..." # Ad-hoc queries
# View logs
npx wrangler tail
Architecture
src/
├── index.ts # Main router, CORS, request handling
├── config.ts # Configuration constants
├── types.ts # TypeScript type definitions
├── utils.ts # Utilities (bandwidth, response, AI, benchmarks, candidates, techSpecs)
└── handlers/
├── health.ts # GET /api/health
├── servers.ts # GET /api/servers - List servers with filtering
└── recommend.ts # POST /api/recommend - AI-powered recommendations
Key Data Flow
- User sends request (
tech_stack,expected_users,use_case,region_preference) - Tech specs calculation with DB workload multiplier based on use_case
- Candidate filtering with flexible region matching
- VPS benchmarks retrieval (Geekbench 6), prioritizing same provider
- AI analysis returns 3 tiers: Budget, Balanced, Premium
- Results cached in KV (5 min TTL, empty results not cached)
D1 Database Tables (cloud-instances-db)
providers- Cloud providers (50+)instance_types- Server specificationspricing- Regional pricingregions- Geographic regionstech_specs- Resource requirements per technology (vcpu_per_users, min_memory_mb)vps_benchmarks- Geekbench 6 benchmark data (269 records)benchmark_results/benchmark_types/processors- Phoronix benchmark data
Key Implementation Details
DB Workload Multiplier (recommend.ts)
Database resource requirements vary by workload type, not just user count:
| Workload Type | Multiplier | Example Use Cases |
|---|---|---|
| Heavy | 0.3x | analytics, log server, reporting, dashboard |
| Medium-Heavy | 0.5x | e-commerce, ERP, CRM, community forum |
| Medium | 0.7x | API, SaaS, app backend |
| Light | 1.0x | blog, portfolio, documentation, wiki |
Example: PostgreSQL (vcpu_per_users: 200) with 1000 users
- Analytics dashboard: 1000 / (200 × 0.3) = 17 vCPU
- E-commerce: 1000 / (200 × 0.5) = 10 vCPU
- Personal blog: 1000 / (200 × 1.0) = 5 vCPU
Bandwidth Estimation (bandwidth.ts)
Estimates monthly bandwidth based on use_case patterns:
| Pattern | Page Size | Pages/Day | Active Ratio |
|---|---|---|---|
| E-commerce | 2.5MB | 20 | 40% |
| Streaming | 50MB | 5 | 20% |
| Analytics | 0.7MB | 30 | 50% |
| Blog/Content | 1.5MB | 4 | 30% |
Heavy bandwidth (>1TB/month) prefers Linode for included bandwidth.
Flexible Region Matching (candidates.ts)
Region matching supports multiple input formats:
LOWER(r.region_code) = ? OR
LOWER(r.region_code) LIKE ? OR
LOWER(r.region_name) LIKE ? OR
LOWER(r.country_code) = ?
Valid inputs: "korea", "KR", "seoul", "ap-northeast-2", "icn"
AI Prompt Strategy (ai.ts)
- Uses OpenAI GPT-4o-mini via Cloudflare AI Gateway (bypasses regional restrictions)
- Server list format:
[server_id=XXXX] Provider Name...for accurate ID extraction - Scoring: Cost efficiency (40%) + Capacity fit (30%) + Scalability (30%)
- Capacity response in Korean for Korean users
Bindings (wrangler.toml)
[[kv_namespaces]]
binding = "CACHE"
id = "c68cdb477022424cbe4594f491390c8a"
[[d1_databases]]
binding = "DB"
database_name = "cloud-instances-db"
database_id = "bbcb472d-b25e-4e48-b6ea-112f9fffb4a8"
[vars]
OPENAI_API_KEY = "sk-..." # Set via wrangler secret
Testing
# Health check
curl https://server-recommend.kappa-d8e.workers.dev/api/health
# Recommendation (e-commerce)
curl -X POST https://server-recommend.kappa-d8e.workers.dev/api/recommend \
-H "Content-Type: application/json" \
-d '{
"tech_stack": ["php", "mysql"],
"expected_users": 1000,
"use_case": "e-commerce shopping mall",
"region_preference": ["korea"]
}'
# Recommendation (analytics - heavier DB workload)
curl -X POST https://server-recommend.kappa-d8e.workers.dev/api/recommend \
-H "Content-Type: application/json" \
-d '{
"tech_stack": ["postgresql"],
"expected_users": 500,
"use_case": "analytics dashboard",
"region_preference": ["japan"]
}'
Recent Changes
- Modular architecture: Split from single 2370-line file into 7 modules
- DB workload multiplier: Database resource calculation based on use_case
- KV caching: 5-minute cache with smart invalidation (empty results not cached)
- OpenAI integration: GPT-4o-mini via AI Gateway for better recommendations
- Bandwidth estimation: Automatic bandwidth category detection for provider filtering
- Tech specs update: Realistic vcpu_per_users values for 150+ technologies