Challenged RLC-AI performance claims before NVidia/Humain use

February 24, 2026 at 4:09 PMstrategyhigh

Situation

Peter personally interrogated the RLC-AI 9-10% performance advantage claim by going directly to Damen Knight (engineer who ran benchmarks) and Max Spevack. Discovered gains largely disappear when benchmarking code is properly optimized (uses torch.compile, etc.). Then asked Damen to evaluate whether Brian's marketing write-up is accurate or misleading: 'makes it sound awesome instead of pointless for data center deployments.'

Reasoning

RLC Pro AI GA targets Feb 25 and the board meeting is today -- performance claims will be part of the narrative to both audiences. Google explicitly wants RLC-AI benchmarks, so if claims don't hold up CIQ's credibility with a major customer is at stake. Peter values accuracy over hype -- the 9-10% number only applies to unoptimized code and disappears on real workloads. By going straight to the engineer who did the work, he bypassed telephone-game distortion. The framing ('makes it sound awesome instead of pointless') shows he's trying to make messaging defensible, not kill it.

Additional Context

Damen's benchmarks ran on A30 GPUs in his home lab. Gains come from unoptimized code paths; once torch.compile and vllm cuda graphs are used, RLC-AI is roughly equivalent to Ubuntu. Hardware matters significantly -- results vary across GPU types. Google roadmap meeting happened day before, where Google requested RLC-AI benchmarks. Board meeting happening today (Feb 24).

Observed Evidence

Peter initiated the investigation: 'I want to get a better understanding of the performance testing.' Damen confirmed: 'Later, when we optimised our benchmarking code, most of those gains disappeared.' Peter acknowledged: 'Yup. And lots of places where Ubuntu outperforms us.' Then asked Damen to audit Brian's write-up. Separately told Bjorn that Google wants RLC-AI benchmarks.

Confidence Breakdown

30/35
Evidence
10/30
Pattern
17/20
Source
15/15
Corroboration

Reasoning Depth Analysis

Org Signal:CTO personally verifying technical claims signals accuracy matters more than marketing convenience
Who Affected:Brian (marketing author), Bjorn (product messaging), Google (wants benchmarks), board (hears claims today)
Precedent:Sets standard that performance claims must survive CTO-level scrutiny before going external
Consequences:If claims are overstated to Google or board and later debunked, credibility damage would be severe
Timing:Day before RLC Pro AI GA target, day of board meeting, immediately after Google asked for benchmarks

Source

reflection

AI Confidence

72%

Related Context

💬
DM with Max & Damen Knight

slack

I want to get a better understanding of the performance testing... where that 9-10% number comes from, what circumstances it applies to

💬
DM with Max & Damen Knight

slack

So the last bullet point, when we optimized our benchmarking code... we were equivalent, yes?

💬
DM with Max & Damen Knight

slack

Now that I understand your results I think his write up says the same thing, just makes it sound awesome instead of pointless for data center deployments.

💬
DM with Bjorn Hovland

slack

And they want RLC-AI benchmarks.

Outcome

No outcome recorded yet.

Decision ID: 549863b3-5361-433f-bb21-1d4e2429fb2e