Stacks·12 min read·May 3, 2026

Higgsfield, but Free: Open Generative AI Setup Guide

One repo, four studios, 200+ models. No subscription. You bring API keys and pay per call. Full setup and the cost trap most people miss.

One repo that bundles image generation, video generation, lip sync, and cinematic camera controls into a single app. MIT licensed, 10.9K stars on GitHub, built by Anil Matcha. No subscription. You bring your own API keys and pay per call. Runs as a desktop app (Mac, Windows, Linux) or a web interface.

Repo: github.com/Anil-matcha/Open-Generative-AI

The Problem Nobody Talks About

Higgsfield charges $100/month for a wrapper around the same AI models you can access directly. You're paying for the interface, not the AI. Most of the models underneath (Flux, Kling, Veo, Sora) are available through API providers at per-call pricing. If you generate 20 images a month, the subscription is a bad deal. If you generate 2,000, it might make sense. But you never get to see that math because the pricing is a flat monthly fee.

Open Generative AI strips the wrapper off. Same models, same quality, same workflow. The difference is you see exactly what each generation costs because you're paying the API directly.

Why This Matters

200+ models in one UI. 50+ for text-to-image, 55+ for image-to-image, 40+ text-to-video, 60+ image-to-video, 9 lip sync models. Switch between Flux, Midjourney, Kling, Sora, and Veo without leaving the app.

Cinema Studio is the headline feature. Pick your camera (8K Digital, 70mm Film, S35 Studio), choose from 11 lens types, set focal length from 8mm ultra-wide to 85mm portrait, dial in aperture from f/1.4 to f/11. The app builds the cinematic prompt for you. This is what Higgsfield was charging for.

Local inference if you want zero API cost. The desktop app bundles sd.cpp for running Stable Diffusion models on your machine. Apple Silicon gets Metal GPU acceleration. You can also connect a Wan2GP server on a CUDA machine for Flux, Wan 2.2, and Hunyuan Video locally.

Workflow Studio for chaining operations. Node-based visual pipeline builder. Chain image gen into video gen into lip sync in one flow.

Step 1: Install the Desktop App

Download the installer for your platform from the GitHub releases page.

Mac (Apple Silicon): Download the .dmg arm64 file. After installing, you may need to bypass Gatekeeper:

xattr -cr "/Applications/Open Generative AI.app"

Then right-click the app and select Open.

Windows: Download and run the x64 .exe installer.

Linux: AppImage or .deb package. On Ubuntu 24.04+, if the AppImage fails:

sudo sysctl -w kernel.apparmor_restrict_unprivileged_userns=0

Step 2: Get Your API Key

Open Generative AI routes cloud model requests through an API gateway called Muapi. You need a key from their site to use the cloud models.

Create an account, generate a key, and the app prompts you for it on first launch. You only pay for what you generate. No monthly minimum.

This is a preview. The full guide continues inside.

The complete version includes everything above plus:

Plus 12 other full guides on agent builds, MCP setups, and Claude workflows. All free inside.

  • Step 3: Pick a studio and generate (all five studios explained)
  • Step 4: Set up local inference with sd.cpp and Wan2GP
  • Step 5: Run from source (self-host the web version)
  • The cost math: per-call vs Higgsfield subscription breakdown
  • Honest limitations (Muapi dependency, local inference gaps, RAM requirements)
Join My Skool (Free)