NewVivly x Aquin — Structuring Social Data for AI. Read the case study

Aquin is the research company reverse engineering with interpretability to

designingintelligence

Debug and improve ML pipelines and models by pin-pointing issues and simulating fixes. Ranging from reducing hallucinations to ensuring safety.

Book a Demo
GPT-2 Small
+
Causal Trace
tokposembL0L1L2L3L4L5L6L7L8L9L10L11L12L13L14L15out
PROMPTFEATURESRESPONSEEiffelTowerlocatedingeography/capitalsFrench landmarksproper nounsEuropean citiesParisFrance
ask anything about your model...
+
GPT-2 Small ∨

Backed by

Emergent Ventures
Emergent Ventures
Founders Inc
Founders Inc
The Residency
The Residency
NVIDIA Inception
NVIDIA Inception
OpenAI
OpenAI
AWS Activate
AWS Activate
Microsoft for Startups
Microsoft for Startups
Google for Startups
Google for Startups

build models with
precision of writing code

01Observe & Find
prompt
What
is
the
capital
of
France
?
response · inherited signal
The
capital
of
France
is
Paris
.
02Simulate & Fix
s0s25s50s75s100lossgrad norm
click a checkpoint ● to compare outputs
03Debug & Improve
base model
exposure per vector
Prompt injection
71%
Role confusion
58%
Suppression
44%
Boundary bypass
33%
Context bleed
29%
Multi-turn drift
18%
high
med
low
new exposure
reduced
trojan layer scan
layers.8.mlp.down_proj0%
layers.4.mlp.gate_proj0%
layers.12.self_attn.v0%
attribution scores
What
is
the
capital
of
France
?
answer locks at L12
loss curve
step 1,200loss 0.74

Attribution & Training

Every output token traced to the exact prompt span and layer where the answer formed. Stream loss, grad norms, and dead layers live.

eval suite · 0 of 12 failing
benchmarks
73%
IS
63%
FP
86%
MU

Evals & Benchmarks

Consistency, suppression, and boundary probes exposing failure modes benchmarks miss. InterpScore, FeaturePurityScore, MUI metrics.

attack surface
prompt injection
78%
jailbreak
54%
trojan
31%
data leak
19%

Security

Trojan scanning, prompt injection, jailbreak probing — full attack surface mapped.

Inspect all major architectures

EmbeddingQ·K·V WeightsbiasQ·K·VDatavisempowersusersto×+=Dd=1Eid·Wdj+bj=QKVij

Transformers & LLMs

Transformer EncoderMLPHeadClassBirdBallCar0*1234567Patch +Pos EmbLinear Projection of Flattened Patches

Vision transformers

8Dense · all neurons activetap to toggle

Dense & MoE

E0E1E2royaltygenderplaceanimalsyntaxtensesparse autoencoder · feature decomposition

Embeddings

promptanswer

Reasoning models

Machine learning

Inspect all major training methods

W (frozen)+AB

LoRA

Low-rank adapter.

4-bit base+ALoRA Δ

QLoRA

4-bit base + LoRA adapters.

all parameters updated

Full fine-tune

All parameters updated.

chosen ✓rejected ✗policy πno reward model required

DPO

Direct preference optimization.

policy πrolloutrewardRL from human feedback

PPO

Proximal policy gradient from human feedback.

teacherKDstudent

Distillation

Teacher compresses into student.

G190%G262%G335%G412%relative reward within group

GRPO

Group relative policy optimization.

Thequickbrownfox[?]next-token prediction at scale

Pre-training

From scratch on massive corpora.

The science underneath: Mechanistic interpretability

Mechanistic interpretability reverse-engineers how neural networks compute, not just what they output. Aquin applies sparse autoencoders, logit lens, activation patching, and causal tracing to expose which features fire, which layers encode a concept, and which circuits produce each token. Stop guessing why a model hallucinated, drifted, or refused — trace the answer back to the exact prompt span that caused it and patch at the source.

SAEfeaturescircuitslayers
features
circuits
layers

Join the Aquin Research Community

LLM researchers & ML engineers — open research, fellowships, hackathons, and early beta access.

Join Discord

Not sure if Aquin is right for you?

© 2026 Aquin. All rights reserved.

Aquin