Open Source

Building AI in the Open for Accountability

We believe transparency is a prerequisite for trustworthy AI. Our open-source research code, evaluation tools, and datasets are published to enable scrutiny, reproducibility, and independent review.

These resources support AI governance, safety evaluation, and risk assessment, not autonomous deployment.

Open source code collaboration

Deployment Boundary Notice

The open-source tools and datasets listed on this page are provided for research, evaluation, and governance support.

They are not production systems, do not constitute deployment approval, and are not substitutes for independent AI risk audits or regulatory review.

Featured Repositories

Reference implementations and evaluation frameworks used to support AI safety analysis and governance review

Reference inference framework designed to demonstrate safety controls, including guardrails, output filtering, and decision logging.

This project is intended to illustrate defensible inference patterns and support audit discussions, not to serve as a turnkey deployment system.

Safety evaluation Governance design Audit preparation
Python PyTorch CUDA
3.2K 456 forks Updated 2 days ago

Interactive decision-support tool for analyzing cost, quality, and latency tradeoffs in AI system design.

Used to support documented decision-making during system planning and governance review.

Architecture evaluation Risk tradeoff analysis Governance documentation
TypeScript React D3.js
892 134 forks Updated 1 week ago

Benchmarking toolkit for evaluating bias and disparate impact across protected attributes and use cases.

Designed to support pre-deployment bias evaluation and documentation of known limitations.

Fairness review Regulatory documentation Risk assessment
Python PyTorch scikit-learn
1.8K 289 forks Updated 5 days ago

Interpretability toolkit providing multiple explanation methods (e.g., SHAP, LIME, attention visualization) through a unified interface.

Used to support explainability analysis and audit readiness, not post-hoc justification.

Interpretability review Audit evidence Model documentation
Python PyTorch TensorFlow
2.4K 367 forks Updated 3 days ago

Document integrity evaluation toolkit combining cryptographic hashing and ML-based tamper detection.

Designed to support forensic review and compliance workflows, with traceable decision signals suitable for audit and legal contexts.

Forensic analysis Compliance evaluation Evidence generation
Rust Python bindings WebAssembly
1.1K 156 forks Updated 1 day ago

Reference implementations for uncertainty estimation and calibration analysis, including conformal prediction, ensembles, and Bayesian methods.

Used to evaluate whether model confidence aligns with observed reliability.

Uncertainty evaluation Calibration analysis Deployment readiness review
Python PyTorch JAX
967 112 forks Updated 4 days ago
Software development and code review

Open Datasets

Datasets published to support evaluation, benchmarking, and reproducible research

TERA-BIAS-21

Bias evaluation dataset for NLP tasks across multiple protected categories.

250,000 annotated samples License: CC BY 4.0 Use: Research & evaluation only Version: v2.1

DocTamper-Bench

Dataset for evaluating document integrity and tamper detection methods.

100,000 documents (pristine and tampered) License: Apache 2.0 Use: Evaluation & benchmarking Version: v1.3

MedSafe-Eval

Evaluation dataset for healthcare AI safety analysis with expert physician annotations.

15,000 annotated cases License: Research use only Use: Safety evaluation / Governance review Version: v1.0

Uncertainty-Calib

Multi-domain benchmark for uncertainty calibration and confidence assessment.

500,000 samples License: MIT Use: Evaluation & methodological research Version: v2.0
Data analysis and benchmarking

How to Contribute

Join a community focused on accountable AI

1

Find an Issue

Browse repositories for governance, safety, and evaluation tasks labeled "good first issue" or "help wanted."

2

Fork & Develop

Implement changes following documented contribution and review standards.

3

Submit a Pull Request

Submit a PR with a clear description of scope and rationale. Reviews are conducted with a focus on correctness, documentation, and reproducibility.

How This Work Is Used

These open-source resources inform and support:

Independent AI risk audits
Pre-deployment governance reviews
Uncertainty and bias evaluations
Audit-ready documentation

Open source enables scrutiny.

Independent audits establish accountability.

Build With Us

Whether you are a researcher, engineer, or student, you are welcome to contribute to work focused on defensible, auditable AI systems.