How to Monetize Local AI Inference from Your Terminal

A practical guide to earning passive income by running AI models locally and selling inference via Nostr DVM and Lightning Network.

How to Monetize Local AI Inference from Your Terminal

Running AI models locally has never been easier. With tools like Ollama, you can run powerful language models on your own hardware — and actually earn money from it.

The Setup

All you need:

  • A Mac (or Linux box) with decent RAM
  • Ollama installed (brew install ollama)
  • A Nostr identity (free, decentralized)

Step 1: Run AI Locally

ollama pull llama3
ollama serve

That’s it. You now have a local AI API at http://localhost:11434.

Step 2: Create a Nostr DVM

Nostr’s Data Vending Machine (DVM) protocol (NIP-90) lets you offer services on the decentralized network. Think of it as a “freelance marketplace” but without a middleman.

Your DVM listens for kind 5050 events (text generation requests), processes them with Ollama, and returns results. Payment flows through Lightning Network — instant, near-zero fees.

Step 3: Set Up a Paywall API

Wrap your Ollama instance in a simple HTTP server:

  • Free tier: 10 requests/hour (to attract users)
  • Paid tier: 10 sats per request via Lightning

This gives developers a cheap, private alternative to OpenAI’s API.

Why This Works

  1. Zero recurring costs — You own the hardware, you own the model
  2. Privacy — Users’ prompts never leave your server
  3. Lightning payments — No payment processor fees, no chargebacks
  4. Decentralized — No platform can deplatform you

Revenue Potential

  • Nostr DVM: ~$30-50/month from regular users
  • API paywall: ~$10-100/month depending on traffic
  • Combined with content: builds your reputation and funnel

The key is consistency. Post daily on Nostr, deliver quality AI results, and the network effects compound over time.


Built entirely from the terminal. No registration, no platforms, no middlemen.


No comments yet.