
Today, we’re excited to introduce Mentat, CTGT’s new OpenAI-compatible endpoint built to deliver something the AI industry has been missing: deterministic control over large language models.
For years, organizations have relied on prompts, filters, and RAG as their primary guardrails. But these are inherently probabilistic. Sometimes they work; often they don’t. As LLMs find their way into regulated or high-impact workflows, “works most of the time” is no longer good enough.
With Mentat, we’re introducing a different paradigm - one built not on asking models to behave, but on mathematically guaranteeing that they do.
Enterprises today face two consistent pain points with LLMs:
Models often fail to integrate retrieved facts. Even with correct context, they can produce incorrect conclusions, inconsistent logic, or spatial/temporal inversion errors.
We repeatedly saw cases where the model had the data but could not apply it.
LLMs are trained on huge distributions full of misconceptions. Even strong system prompts can’t reliably override them, not when misinformation is more common in the pretraining data than truth.
Example: On TruthfulQA, large models still confidently produce myth over fact ~80% of the time.
These failures aren’t random: they’re structural. And structural problems require structural solutions.
While studying the DeepSeek-R1 model, our team discovered something important:
models often know the truth, but a set of internal features suppresses or distorts it.
By identifying and manipulating these latent feature vectors, bias vectors, misconception vectors, censorship vectors, we can surface grounded knowledge and suppress confabulation.
This insight became the foundation for Mentat.
Mentat introduces two core innovations that make deterministic control possible:
Instead of prompting the model to follow rules, we modify behavior directly at the activation level during the forward pass.
We identify the latent vector v responsible for a behavior (e.g., misconception, bias, refusal), then apply a controllable intervention:
h′ = h – α * (h @ v) * vThis lets us remediate undesirable behavior in real time by: suppressing confabulation, elevating factual reasoning, removing style or bias features, undoing knowledge suppression. Because this happens as a lightweight arithmetic operation on hidden states, overhead stays under 10ms per token. No fine-tuning. No brittle prompts. No unpredictable “guardrails.” Just deterministic steering.
Beyond internal steering, Mentat runs each response through a verification pipeline to ensure compliance, accuracy, and internal consistency.
Detects uncertainty or babbling before it reaches the user.
Cross-checks claims against a structured graph to catch hallucinations that RAG typically misses, especially relational errors.
Policies, from brand guidelines to financial regulations, are compiled into weighted vectors. Mentat then adjudicates between them deterministically, ensuring the output always conforms to organizational policies.
This is the backbone of Mentat as a policy engine: a system that enforces your standards automatically and consistently, across every generation.
Across internal and third-party evaluations, Mentat significantly improves factual accuracy and reasoning robustness:
For enterprise workflows, legal, finance, healthcare, compliance, the improvement is transformational.
A drop-in compatible replacement for /v1/chat/completions.
Swap your base URL and immediately upgrade your reliability.
api.ctgt.ai/v1/chat/completions
(Full docs available in the dashboard.)
The public endpoint ships with our Base Policy, optimized for hallucination reduction, factuality, and corporate safety.
We built a visual playground where you can:
No signup required. Try it here.
Here’s a quick 2-minute demo video showing the process: Watch video here
As LLMs move into high-risk environments, organizations can’t rely on best-effort guardrails. They need deterministic, inspectable, and enforceable systems. Mentat gives teams the ability to remediate incorrect or noncompliant model behavior with: real-time corrections, policy-consistent output, predictable internal logic, and transparent reasoning. This is how enterprise AI should work.
We’re excited to finally share Mentat with the world.
If you’re building where correctness, compliance, or trustworthiness matter, we want to hear from you.
Validate the solution. Stress-test our interventions. Try to break the system.
Your feedback will help us refine the next wave of deterministic AI infrastructure.