A transparent, single source of truth for our performance claims, measurement methodology, and beta metrics.
0.80s
~1.8s
98.5%
127
45,892
99.9%
* All figures exclude synthetic tests and internal traffic. See "Measurement Rules" below.
Time from query received → top-k returned
Why: determines perceived speed
Upload → searchable (median/p95)
Why: measures time-to-value for new data
4xx/5xx % per endpoint
Why: reliability indicator
% of queries returning 0 candidates
Lower is better
% of searches that add ≥1 candidate to shortlist
Proxy for relevance/utility
Pearson r between interviewer and AI scores
Per skill dimension
Scripted tasks with 15 recruiters on identical 5k-resume corpus:
Find a senior Python + AWS candidate from corpus
Paste JD and shortlist top 5 matches
Interview (15-min segment) and produce scorecard
ATS keyword search + manual scanning
12 min avg
Natural language + JD match + Copilot
3.5 min avg
Reported metric: time-to-first-shortlist (median)
Analytics are tenant-isolated with no cross-customer model training
PII redaction for transcripts with configurable TTL policies
Opt-out available for analytics at tenant level
Important: We do not train shared foundation models on customer data. Your resumes and interview data are never used to improve models for other customers.
2025-08-08
Published Metrics page; added quick-stats; defined p95/p50 rules
2025-08-01
Beta launch; initial metrics tracking enabled
2025-07-15
Added AI↔Human score correlation metrics
Production logs, vector DB metrics, and OpenAI usage reports (aggregated in our warehouse).
Latency is skewed; p95 reflects worst-case user experience that 95% of users experience.
No. We don't train shared foundation models on customer data. Your data stays yours.
Experience the speed and accuracy we measure. Join our beta to help shape these metrics.