SkillJavaScriptv1.3.2

lightningprox

Pay-per-request AI model access via Bitcoin Lightning using prepaid spend tokens.

0 downloads
unixlamadev-spec
Updated Mar 18, 2026

LightningProx — Lightning-Native AI Gateway

Pay-per-use access to 19 AI models across 5 providers via Bitcoin Lightning micropayments. No API keys. No subscriptions. No accounts. Pay sats, get inference.

When to Use

  • Accessing AI models without provider API keys
  • Autonomous agent inference with Lightning payments
  • Comparing responses across multiple providers
  • Low-cost inference via open models (Llama 4, Mistral, DeepSeek)
  • Vision tasks (multimodal via image_url)
  • Code generation (Codestral, Devstral)
  • Reasoning tasks (Magistral)

Supported Models (19 total)

ProviderModels
Anthropicclaude-opus-4-6, claude-sonnet-4-6, claude-haiku-4-5
OpenAIgpt-4o, gpt-4-turbo, gpt-4o-mini
Together.aillama-4-maverick, llama-3.3-70b, deepseek-v3, mixtral-8x7b
Mistralmistral-large-latest, mistral-medium, mistral-small, codestral, devstral, magistral
Googlegemini-2.5-flash, gemini-2.5-pro

Payment Flow

Option A — Spend Token (recommended for repeat use)

# 1. Top up at lightningprox.com/topup — pay Lightning invoice, get token
# 2. Use token directly
curl -X POST https://lightningprox.com/v1/messages \
  -H "Content-Type: application/json" \
  -H "X-Spend-Token: $LIGHTNINGPROX_SPEND_TOKEN" \
  -d '{
    "model": "claude-opus-4-5-20251101",
    "messages": [{"role": "user", "content": "Hello"}],
    "max_tokens": 1000
  }'

Option B — Pay per request

# 1. Send request without token → get invoice
curl -X POST https://lightningprox.com/v1/messages \
  -H "Content-Type: application/json" \
  -d '{"model": "gemini-2.5-flash", "messages": [{"role": "user", "content": "Hello"}], "max_tokens": 100}'

# 2. Pay the Lightning invoice returned
# 3. Retry with X-Payment-Hash header

Drop-in OpenAI SDK Replacement

npm install lightningprox-openai
// Before: import OpenAI from 'openai'
import OpenAI from 'lightningprox-openai'
const client = new OpenAI({ apiKey: process.env.LIGHTNINGPROX_SPEND_TOKEN })

// Everything else stays identical:
const response = await client.chat.completions.create({
  model: 'claude-opus-4-5-20251101',
  messages: [{ role: 'user', content: 'Hello' }]
})

Two lines change. Nothing else does.

Check Available Models

curl https://lightningprox.com/api/capabilities

Security Manifest

PermissionScopeReason
Networklightningprox.comAPI calls for AI inference
Env ReadLIGHTNINGPROX_SPEND_TOKENAuthentication for prepaid requests

Trust Statement

LightningProx is operated by LPX Digital Group LLC. Payment = authentication. No data stored beyond request logs. No accounts, no KYC. Operated at lightningprox.com.

Free
Installation
Reviews

Sign in to leave a review.

No reviews yet. Be the first.