Siddharth Ramakrishnan

Writing

Specialized Intelligence

Part of the AI Systems topic guide.

There is a common 2026 assumption about models that goes something like this: frontier labs are so far ahead that fine-tuning small open models is mostly a toy exercise. That is directionally true for broad tasks. It is much less true for narrow, repeated workflows where the interface is fixed and the labels match the job.

I ran a small experiment around C/C++ numeric vulnerability triage and got a result I wanted to record here: a distilled Qwen2.5-Coder-7B model is able to beat GPT-5.2 on a frozen real-world benchmark.

The Setup

The task was deliberately narrow. Given a code snippet, the model had to decide whether it contained a numeric vulnerability, classify the subtype, and return structured fields for vulnerable, subtype, location, and reason. The benchmark focused only on CWE-190 and CWE-191: integer overflow and integer underflow in C/C++.

The eval set was a frozen 140-example PrimeVul slice with 20 real positives and 120 negatives. Because the set is negative-heavy, raw accuracy is not very interesting here. The more useful metric is balanced binary accuracy: positive recall plus negative accuracy, divided by two.

The Result

The base Qwen2.5-Coder-7B model was extremely conservative. It reached 63.8% balanced accuracy and only caught 30% of the true positives. The best student, a distilled run warm-started on Juliet and then trained on PrimeVul task outputs, reached 73.8% balanced accuracy with 95% positive recall. On the same eval, GPT-5.2 came in at 70.8% balanced accuracy with 85% positive recall.

PrimeVul Numeric-Triage Benchmark
Model Balanced Accuracy Positive Recall Negative Accuracy
Qwen2.5-Coder-7B base 63.8% 30.0% 97.5%
Qwen + Juliet stage 1 50.0% 0.0% 100.0%
Qwen + PrimeVul distilled 70.0% 85.0% 55.0%
GPT-5.2 70.8% 85.0% 56.7%
Qwen + Juliet -> PrimeVul distilled 73.8% 95.0% 52.5%

Eval set: 140 PrimeVul examples, including 20 positives and 120 negatives.

That is not a massive win, but it is a real one. On this narrow workflow, a cheap specialist beat the frontier baseline.

Why It Matters

The interesting lesson is not that small models are secretly smarter. It is that specialization still matters. Public matched data, a tightly scoped objective, structured outputs, and frontier-model distillation were enough to move a small open model from "barely catches the positives" to "competitive with and slightly ahead of the frontier model we distilled from."

The best model mostly won by becoming more aggressive. It found more of the real vulnerabilities, but it also produced more false positives. So this is not a story about a uniformly better reasoner. It is a story about getting the tradeoff right for a particular operating point.

I also think this pattern is underappreciated because people collapse everything into one question: which model is smartest? In practice the more useful question is often: what is the cheapest model that can be tuned to do this one repeated job well enough to matter? For that question, specialized intelligence can still win.

The Right Takeaway

I would not generalize this too far. This is not evidence that 7B models have better general reasoning than frontier models. It is not evidence that you can solve broad vulnerability research with a weekend fine-tune. It is distillation on a narrow task, with some remaining overlap risks that the experiment notes discuss explicitly.

But that narrower claim is still important. If the workflow is repeated, the scope is well-defined, and the outputs can be constrained, then there is still real room for task-specific models that run at a fraction of the cost of the general systems they borrow from.

The code, benchmark outputs, and experiment notes are on GitHub.