P2P LLM Network

Share your LLM.Use everyone else's.Free.

Share your GPU with the Neurlap network and earn credits. Spend those credits to access any model — powered by a global community of providers.

The credit cycle

Share. Earn. Use.

STEP 01

Share your model

Run any open-source LLM on your hardware. Neurlap connects you to the network automatically.

e.g. Llama 3.3, Qwen 2.5, Gemma 3
STEP 02

Earn credits per token

Every inference request your GPU serves earns credits. Bigger models earn more.

Credits accumulate in real-time
STEP 03

Use any model for free

Spend your earned credits to query models hosted by other providers across the network.

OpenAI-compatible API — just swap the base URL

Zero dollars spent — only credits exchanged

Why Neurlap

Your GPU is idle
90% of the time

Your GPU sits idle most of the day. With Neurlap, you pick a model from our catalog, download it in one click, and start serving real inference requests — every token earns you credits to use any model in the network.

Free forever for contributors.
Share your GPU and never pay a cent. Just want to use the API? Premium plans coming soon.

Your credit ledger

LIVE
+Served Llama 3.3 70B2,140 tok+214
+Served Llama 3.3 70B890 tok+89
Used Gemma 3 27B:Google800 tok−240
+Served Qwen 2.5 32B3,400 tok+272
Balance+335 credits

Traditional approach

Monthly subscription for API access
One provider, limited model selection
GPU unused between your own prompts
Usage costs scale with your growth

With Neurlap

Free access by sharing your GPU
Every model on the network, one API
Your GPU earns credits while idle
Premium plans available if you prefer not to share

Get started

Start earning credits
in three simple steps

1

Install the app

Download the lightweight Neurlap client. Runs in your system tray on macOS, Windows, or Linux.

2

Pick a model

Browse the model catalog and download any GGUF model. The built-in engine handles everything.

3

Earn and use

Credits accrue as your GPU serves requests. Spend them on any model in the network.

Supports any GPU

Apple SiliconMetal
NVIDIACUDA
AMDROCm

Built for the community

Everything you need to share and use LLMs

OpenAI-compatible API

Change one line of code. Your existing app just works.

base_url = "api.neurlap.ai/v1"
# That's it. Nothing else changes.

Global routing

Requests route to the nearest provider with the best latency.

Community-powered

Powered by real people sharing their GPUs. The more providers join, the faster and more reliable the network becomes.

San Francisco served Llama 4 Scout3s ago
Tokyo served Qwen 3 235B6s ago
London served Mistral Medium 39s ago
Berlin served Llama 4 Maverick12s ago
Singapore served DeepSeek R215s ago
Sydney served Qwen 3 30B18s ago
Seoul served Gemma 3 27B21s ago
Mumbai served Phi-4 14B24s ago
Toronto served Qwen 3 235B27s ago
Riyadh served Llama 4 Scout30s ago
Paris served DeepSeek R233s ago
Live network activity

Privacy by default

The coordinator never stores prompts. Providers see payloads only during inference.

e2e encrypted in transit

Transparent credits

Earn per token generated. Spend per token consumed. No subscriptions.

generate+0.1 cr/tok
consume-0.15 cr/tok

No lock-in

Connect when you want, disconnect when you want. No penalty for going offline. Your node earns when it's on, and costs nothing when it's off.

Works everywhere

One API. Every AI tool.

Swap one line and your entire stack runs on community GPUs. Compatible with every major AI framework.

OpenAIOpenAI SDK
LangChainLangChain
LlamaIndexLlamaIndex
VercelVercel AI
CrewAICrewAI
HuggingFaceHugging Face
OpenAIOpenAI SDK
LangChainLangChain
LlamaIndexLlamaIndex
VercelVercel AI
CrewAICrewAI
HuggingFaceHugging Face
OpenAIOpenAI SDK
LangChainLangChain
LlamaIndexLlamaIndex
VercelVercel AI
CrewAICrewAI
HuggingFaceHugging Face
OpenAIOpenAI SDK
LangChainLangChain
LlamaIndexLlamaIndex
VercelVercel AI
CrewAICrewAI
HuggingFaceHugging Face
CursorCursor
DifyDify
n8nn8n
GradioGradio
Chatbots
RAG
CursorCursor
DifyDify
n8nn8n
GradioGradio
Chatbots
RAG
CursorCursor
DifyDify
n8nn8n
GradioGradio
Chatbots
RAG
CursorCursor
DifyDify
n8nn8n
GradioGradio
Chatbots
RAG
OpenWebUIOpen WebUI
LobeHubLobeChat
Agents
Code Assist
Workflows
Automations
OpenWebUIOpen WebUI
LobeHubLobeChat
Agents
Code Assist
Workflows
Automations
OpenWebUIOpen WebUI
LobeHubLobeChat
Agents
Code Assist
Workflows
Automations
OpenWebUIOpen WebUI
LobeHubLobeChat
Agents
Code Assist
Workflows
Automations

FAQ

Common questions

Ready to start?

Share your GPU.
Access LLMs for free.

Free for contributors. Premium plans coming soon.

Zero cost
Share your GPU, use any model for free
OpenAI-compatible
Drop-in API — works with your existing code
Decentralized
No central server — inference runs across a global P2P network