Ecosystem/Arthur AI

Arthur AI

Series BPreliminary

AI model monitoring and observability platform that tracks AI performance, detects drift, and monitors for bias in production. Like Datadog but specifically for AI models.

HQUS
Est2019
Raised$42M
arthur.ai
Score
51.4 / 100
Confidence
Preliminary

Strong safety posture with established governance frameworks and active risk management.

Strengths:Governance Maturity, Technical Safety, Risk Assessment, Regulatory Readiness
Weaknesses:External Engagement
Competitive positioning

Leading AI observability platform. Competes with Datadog, New Relic adding AI monitoring. Differentiates through AI-specific depth.

Key risk

Observability incumbents (Datadog, New Relic, Splunk) are adding AI monitoring features. Standalone AI observability may get absorbed into broader platforms.

Enterprise traction

Enterprise customers in financial services. Growing recurring revenue.

financial services
Safety area

Robustness & Adversarial

Enterprise business needs
Protect my AI in productionProve my AI is compliant

Security Assessment

Security-relevant indicators for vendor evaluation

Security Posture
54
TS-01dim: 58
Red Teaming & Pre-deployment Testing
Adversarial testing before deployment
TS-05dim: 58
Robustness & Adversarial Resilience
Resistance to adversarial attacks
RA-01dim: 50
Sector-Specific Risk Assessment
Risk analysis for deployment context
RA-03dim: 50
Dual-Use & Misuse Risk
Dangerous capability awareness
RA-07dim: 50
Incident History & Track Record
Past incidents and response quality
EE-04dim: 42
Vulnerability Disclosure Program
Bug bounty or CVE reporting process
Incident History
Arthur AI incident records sourced from AIAAIC Repository and public reporting.
Integration: AIAAIC, OECD AI Incidents Monitor
Third-Party Audits
External audit reports, SOC 2 attestations, and ISO certifications verified where published.
Sources: Company filings, registry lookups
CVE & Disclosures
Known vulnerabilities and security advisories from NVD, GitHub Security Advisories, and vendor pages.
Sources: NVD, GHSA, vendor disclosure pages

Dimension Breakdown

GM
Governance Maturitypreliminary
Published policies, corporate structure, safety mandate, whistleblowing, executive commitment.
52
TS
Technical Safetypreliminary
Benchmarks, adversarial robustness, fine-tuning safety, watermarking, model cards, research output.
58
RA
Risk Assessmentpreliminary
Dangerous capability evaluations, thresholds, external testing, bug bounty, halt conditions.
50
RR
Regulatory Readinesspreliminary
ISO 42001, EU AI Act compliance, GPAI obligations, international commitments, incident reporting.
55
EE
External Engagementpreliminary
Survey participation, research support, transparency, behavior specs, open-source contributions.
42

Want Arthur AI scored on the Mappera framework?

Subscribe to get notified when full safety scoring becomes available, or reach out to request a detailed brief.