It's a local OpenAI-compatible API + desktop app that drives various web LLM chat UI (via Playwright), so you can use it from SillyTavern and other clients without paying for the official API. Slightly cursed yet surprisingly effective.
What is this? · Quick start · Client setup · Documentation · Releases · Issues · Discord Server ·
INTENSERPNEXT2PREVIEW.mp4
If you're here because you want to use an LLM in SillyTavern without wiring up the paid official API: Welcome to the club! IntenseRP Next v2 drives the official web apps (like DeepSeek, Z.AI, Moonshot, QwenLM) in a real browser, and re-exposes them as an OpenAI-compatible endpoint.
Unlike the official API, this is usually free (DeepSeek / GLM / Kimi are free to use with limits, and paid plans aren't added yet) and it gives you access to the full web UI experience (including reasoning toggles, search, file uploads, and more). Not without tradeoffs, of course - see below.
- Download a release (see Releases) and run it (or run from source)
- Click Start and log in when the browser opens
- Point your SillyTavern client at
http://127.0.0.1:7777/v1(default) and pickdeepseek-*/glm-*/moonshot-*/qwen-*mode IDs
And it's done! It should Just Work™️.
IntenseRP Next v2 (sometimes shortened to "IRP Next v2") is a local bridge between:
- an OpenAI-style client (like SillyTavern), and
- a provider web app (currently: DeepSeek, GLM Chat, Moonshot, QwenLM)
Under the hood it:
- Starts a local FastAPI server (OpenAI-compatible routes under
/v1) - Launches a real Chromium session (Patchright/Playwright)
- Logs in (manual or auto-login)
- Intercepts the provider's streaming network responses
- Re-emits them as OpenAI-style SSE deltas for your client
In normal human terms: free acces to LLMs via their official web apps but turned into an API you can use in SillyTavern and other OpenAI-compatible clients.
Of course, there are also official paid APIs, but not everyone can pay for them, so this is kind of a free alternative. 🙂
If you read this far, you probably have a use case in mind! But here's the objective truth:
It would work well for you if you:
- want free-ish access to provider web models via the official web apps
- prefer a clicky desktop app over a pile of scripts
- are OK with the occasional wait or hiccup (web apps change)
Not the best fit if you:
- need high throughput / parallel requests (this uses one live browser session)
- want to run headless on a server
- want something that never breaks (that's perhaps the biggest caveat)
Note
- Provider web apps change. When they do, a driver can break until it's updated.
- IntenseRP currently processes one request at a time (requests are queued). This is on purpose (single live browser session).
- This project is not affiliated with DeepSeek, ZhipuAI, SillyTavern, or any provider.
v2 is a full rewrite based on lessons learned from the original IntenseRP API (by Omega-Slender) and my own IntenseRP Next v1. The focus is less on a pile of features and more on making it sane to maintain and hard to break.
It's a more modular codebase with a Playwright-first approach (network interception, no scraping), a better UI (PySide6), and a cleaner settings model, plus built-in update and migration flows.
If you want to compare, have a look:
| Area | IntenseRP API / Next v1 | IntenseRP Next v2 |
|---|---|---|
| Backend | Python (Flask) | Python (FastAPI) |
| UI | customtkinter | PySide6 (Qt) |
| Automation | Selenium-based | Playwright (Patchright) |
| Scraping | HTML parsing (plus workarounds for NI) | Native Network interception |
Tip
First launch can take a bit - v2 will verify/download its browser components.
Windows (recommended)
- Download the latest
intenserp-next-v2-win32-x64.zipfrom Releases - Extract it anywhere
- Open the
intense-rp-nextfolder and runintenserp-next-v2.exe - Click Start and wait for the browser to open
Linux
- Download the latest
intenserp-next-v2-linux-x64.tar.gzfrom Releases - Extract and run:
tar -xzf intenserp-next-v2-linux-x64.tar.gz
cd intense-rp-next
chmod +x intenserp-next-v2
./intenserp-next-v2If it complains about missing libraries, you may need Qt6 deps installed on your system. The best way is to install the qt6-base package via your package manager, but if it doesn't stop you can just install the missing libs manually.
From source (for devs)
Requirements: Python 3.12+ (3.13 recommended)
git clone https://github.com/LyubomirT/intense-rp-next.git
cd intense-rp-next
python -m venv venv
source venv/bin/activate # Linux/Mac
# or: venv\\Scripts\\activate # Windows
pip install -r requirements.txt
python main.pyOnce the app says Running (Port 7777):
| Setting | Value |
|---|---|
| Endpoint | http://127.0.0.1:7777/v1 |
| API | OpenAI-compatible chat or text completions |
| API key | Leave blank (unless you enabled API keys) |
| Model | deepseek-* / glm-* / moonshot-* |
Available model IDs (depends on provider):
- DeepSeek:
deepseek-auto(uses your IntenseRP settings)deepseek-chat(forces DeepThink off)deepseek-reasoner(forces DeepThink on, Send DeepThink follows your setting)
- GLM Chat:
glm-auto(uses your IntenseRP settings)glm-chat(forces Deep Think off)glm-reasoner(forces Deep Think on, Send Deep Think follows your setting)
- Moonshot:
moonshot-auto(uses your IntenseRP settings)moonshot-chat(forces Thinking off, Send Thinking off)moonshot-reasoner(forces Thinking on, Send Thinking follows your setting)
Note: these IDs are behavior presets (modes). GLM has separate real model selection in Settings. Moonshot moonshot-* IDs are still behavior presets, not a separate backend model selector.
If you change the port in Settings, update the endpoint to match (example: http://127.0.0.1:YOUR_PORT/v1).
- Browser takes forever on first run: it may be downloading/verifying Chromium. Let it cook, then try again.
- Client cannot connect: confirm the app says Running, and the endpoint matches your port (
http://127.0.0.1:7777/v1by default). - 401 Unauthorized: you probably enabled API keys in Settings. Either disable them or add a key in your client.
- Login loops / stuck sign-in: try disabling Persistent Sessions, or clear the profile in Settings (it wipes saved cookies).
- Slow responses: requests are queued (one at a time), and DeepThink can add extra time.
Tip: enable the console and/or logfiles before reporting issues. Logs help a lot when diagnosing!
There are a few highlights I think are worth calling out. Most have been in v1 as well, but v2 has them all better and in a cleaner way.
- 🖥️ A desktop UI that starts/stops everything for you (and doesn't require terminal work)
- 🔌 An OpenAI-compatible API under
/v1for SillyTavern and other OpenAI-compatibles - 🧩 A formatting pipeline: templates, divider, injection, name detection
- 🧠 Provider behavior toggles: DeepSeek, GLM Chat, and Moonshot behavior controls
- 🔐 Optional LAN mode and API keys
- 🪵 Built-in extensive logging: console window, log files, console dump
- ♻️ Built-in v1 migrator + built-in update flow (when running packaged builds)
Current:
- DeepSeek (usable; in "verification" stage)
- GLM Chat (usable; beta-like, Search supported)
- Moonshot (usable; first integration stage)
More detail lives in docs/ (best viewed as the docs site - see below).
There is a full docs site with screenshots and details if you want to dig a bit deeper:
Check out the docs site here.
Local preview (Zensical):
python3 -m pip install -r docs/requirements.txt
zensical serveIf IntenseRP Next v2 is useful to you, and you wish to help, thank you!! The easiest support is a star and a quick issue report / feature request when something is missing or broken.
If you want quicker back-and-forth or community troubleshooting, you can also join the IntenseRP Next Discord server.
If you want to help financially as well (optional, but appreciated), see: Support the Project in the docs.
- IntenseRP is designed for local or LAN use. Do not expose it to the public internet unless you know what you're doing.
- If you enable Available on LAN, consider enabling API Keys too.
- Your config directory contains sensitive data (credentials, API keys, session cookies). Treat it like a password vault.
Bug reports, suggestions, and PRs are welcome!! 💖
Just note a few things:
- This is still a fast-moving codebase. A PR can become outdated quickly.
- Provider behavior changes are inevitable (web UIs are a moving target).
- I move this in a very "me" way due to how fast things change, meaning not every idea will align with my vision even if it's objectively good.
If you're not sure where to start, open an issue first - it saves everyone time.
| LyubomirT | Omega-Slender | Deaquay | Targren | fushigipururin | Vova12344weq |
| Project Maintainer | Original Creator | Contributor to OG | Feedback & Proposals, Code | Code and Concept Contributor | Early Testing, Bug Reports, Suggestions |
Full list: https://github.com/LyubomirT/intense-rp-next/graphs/contributors
IntenseRP Next v2 is licensed under the MIT License. See the LICENSE file for details.
Note
Original IntenseRP API by Omega-Slender is also MIT-licensed, but previously was a CC BY-NC-SA 4.0 project. This v2 rewrite is a new codebase and is not a derivative work, so the license has been switched to MIT for simplicity. I'm not affiliated with Omega-Slender, even if I'm the official successor to their project (starting from v1).
- FastAPI, Pydantic, Uvicorn
- PySide6 (Qt)
- Playwright + Patchright
- Feather Icons / Lucide Icons
- SillyTavern (client ecosystem)
- IntenseRP API (Omega-Slender) - original inspiration
- Me (LyubomirT) - for doing all the work :D
- RossAscends (for STMP)
- Developers of Zensical (docs generator)
