Live on Base

Your Mac Has
Unused RAM.
Put It to Work.

Got a Mac Mini or Mac Studio sitting idle after the M4 craze? That 16–512GB of unified memory can run AI models and earn you xRAM tokens. Install the app, share your spare compute, get paid.

For Users

Reserve fast AI sessions across the network.

Pick a model, fund a session, and keep aggregate Mac RAM focused on your workload instead of waiting through cold starts every time.

  • Choose from live models with readiness and SLA signals.
  • Open a dedicated chat or plug the session into OpenClaw and other OpenAI-compatible tools.
  • Fund with xRAM from your wallet and stop whenever you want.
For Providers

Turn spare Mac RAM into paid model capacity.

Install the app, choose how much memory to share, and help keep larger models online across distributed Apple Silicon nodes.

  • Set your RAM allocation, wallet, and lending window in minutes.
  • Download models that fit locally or join larger sharded deployments.
  • Earn xRAM for staying online and keeping reserved models warm.
100M
xRAM Supply
20+
Models (1B–671B)
Base L2
On-Chain
MLX
Apple Silicon Native
Built On
Base (Coinbase L2) Apple Silicon (MLX) MetaMask ERC-20

What The Coordinator Sees Right Now

These numbers come from the same live onboarding endpoint the user and provider funnels consume.

Providers Online
--

Checking current supply.

Visible RAM
--

Aggregate provider memory visible to the coordinator.

Warm / Ready Models
--

Models that can be routed cleanly right now.

Best Next Step
--

Waiting for live data.

Loading live network status...

Two Sides. One Network.

Whether you want to use AI or power it, RAM Aggregator has you covered.

Use AI

  • 1

    Connect Your Wallet

    Link MetaMask and deposit xRAM tokens — or try for free with the demo key.

  • 2

    Choose a Model

    Pick from Llama, Mistral, Qwen, DeepSeek and more — from 1B to 671B params. See live pricing and provider availability.

  • 3

    Chat Away

    Your prompts are routed to the best available provider. Pay per token, way cheaper than centralized APIs.

Earn xRAM

  • 1

    Install in One Line

    Run curl -fsSL ramaggregator.com/install | bash in Terminal. More options →

  • 2

    Enable Models & Set Limits

    Choose which LLMs to serve. Models auto-download from HuggingFace and run natively via MLX. Use the RAM slider to cap resource usage.

  • 3

    Earn Tokens

    Every inference you complete earns xRAM tokens — minted on-chain to your wallet. Your Mac is literally printing money.

Built for the Future of AI

Enterprise-grade infrastructure, decentralized by design.

🧠

OpenAI-Compatible API

Drop-in replacement for any OpenAI SDK. Point your base_url at RAM Aggregator and your existing code just works. Native chat templates for every model family.

Apple Silicon Native

Models run via MLX — Apple's ML framework optimized for M-series chips. Blazing fast 4-bit quantized inference with threaded execution for reliable connections.

🔗

Pipeline Parallelism

Models too large for one Mac are automatically split across multiple providers. A 405B-parameter model can run across three 96 GB machines. Completely transparent to users.

🔒

End-to-End Encrypted

Optional E2E encryption (X25519 + AES-256-GCM) ensures the coordinator never sees your prompts. Only you and the provider can read your data.

On-Chain Rewards

xRAM is a real ERC-20 token on Base L2. Emission rewards are minted directly to your wallet via smart contract. Bitcoin-style halving — early providers earn the most.

Provider Controls

RAM allocation slider, Prevent Sleep toggle, and model management. Control exactly how much of your Mac to share. Auto-updates keep your node current.

💰

Dynamic Pricing

Prices adjust based on real-time network utilization. Low demand = cheaper inference. High demand = more provider rewards. Floor and ceiling protect both sides.

🤖

20+ Models, 1B to 671B

Llama 3.x, Qwen3, Mistral, DeepSeek R1, and more. MoE models run at surprising speeds. Smart routing picks the best provider with the model already loaded.

🌐

Fully Decentralized

No GPU datacenter. No single point of failure. The network is powered by individual Mac owners worldwide sharing their idle compute.

Your Data, Your Control

We're upfront about what's encrypted, what's visible, and to whom. No fine print.

What's Protected

  • Prompts & Responses (with E2E)

    When E2E encryption is enabled, your prompts and AI responses are encrypted in your browser before leaving your device. The coordinator only sees opaque ciphertext.

  • No Accounts, No Tracking

    No email, no name, no cookies, no analytics. You connect with a wallet address — that's it.

  • No Training on Your Data

    Your prompts are never used to train, fine-tune, or improve any AI models. Ever.

  • All Connections TLS 1.3

    Every connection is encrypted in transit with modern TLS. E2E adds a second layer on top.

What to Know

  • !

    Providers See Your Prompts

    The provider running your inference must see the plaintext to generate a response. This is how all AI inference works — including centralized services like ChatGPT and Claude.

  • !

    Metadata Is Visible to Coordinator

    The coordinator sees which model you used, payload sizes, timing, and your wallet address — but not your prompt or response content (with E2E on).

  • !

    Providers Don't Store Data

    Prompts exist only in memory during inference and are discarded immediately. No logging, no history, no retention.

  • !

    Providers Can't Identify You

    Your wallet address is not shared with providers. With E2E, each request uses a fresh ephemeral key — requests can't be linked to the same user.

Read Full Privacy Documentation

Token Economics

A deflationary supply with halving epochs, designed to reward early providers.

100,000,000

Total xRAM Supply

Fixed supply. No inflation. Every xRAM token is either in the treasury waiting to be earned, or already in a provider's wallet.

70%
30%
Provider Treasury (70M) Team Treasury (30M)

Halving Emission Schedule

Rewards halve every 17.5M tokens minted — just like Bitcoin. Early providers earn significantly more.

EpochReward/JobTreasury Remaining
Epoch 1Base rate70M → 52.5M
Epoch 250% of base52.5M → 35M
Epoch 325% of base35M → 17.5M
Epoch 412.5% of base17.5M → 0

For Users

Deposit xRAM via MetaMask to an escrow smart contract. Tokens are deducted per inference based on model size and token count. Withdraw your unused balance at any time.

For Providers

Earn xRAM for every inference job you complete. Rewards are calculated from the emission schedule and minted on-chain. Stake xRAM to increase your earning multiplier and priority in job routing.

Ready to Join the Network?

Whether you want cheap AI inference or want to earn by sharing your Mac's power, the network is live now.