Maniac: Your best model in one click.
Maniac is an enterprise AI platform that makes it easy to replace existing LLM API calls with fine-tuned, task-specific models. Drop Maniac in with one line of code to:
Capture and structure production LLM traffic
Automatically fine-tune and evaluate Small Language Models (SLMs) on your tasks
Replace over-generalized LLM calls with higher performance, lower latency models built for just what you need
Focus engineering time where it matters most: building and refining high-quality model evaluations—not managing infrastructure, hyperparameters, or bespoke fine-tuning pipelines
All with virtually no changes to your existing codebase.
Getting started
Sign up for Maniac
Head over to https://app.maniac.ai/auth/register
Dropping Maniac into your Codebase
Install the library
Initialize client
Create a container
Containers log inference and automatically build datasets for fine-tuning and evaluation. initial_model sets the model used in that container until a Maniac model is deployed.
Run inference in a container
Running inference will auto-generate inference logs. Data can also be manually uploaded.
Note: We recommend defining the system prompt at the container level. All inference requests executed through that container will automatically inherit this system prompt. If a request’s
messagesarray includes its own system prompt, it will override the container-level system prompt for that request only.
Optimizing your model
The inference logs in your container now serve as training data for a new SLM—fully yours, lower latency, cheaper, and optimized specifically for your task.
Launch Optimization
Once you've defined an eval, the Optimization dashboard lets you configure and run post-training pipelines using techniques such as SFT, GRPO, and GEPA.
Each stage of the pipeline is modular, allowing you to select base models, the evaluation to optimize against, adjust hyperparameters, swap classifier heads, and experiment with different training strategies.
Deploy.
Optimized models can be be deployed into a container from the Models tab. Once deployed, you can chat with your generated models, and inference requests are now routed through the Maniac model instead of the initial_model.

Need help?
📧 Email us at [email protected]
We'll get back to you within a day.
Last updated

