PLANNED INITIATIVE

Project Watchtower

Watchtower is a planned open platform for AI verification. The goal is to provide independent, reproducible benchmarks for AI system behavior, focusing on safety and non-determinism.

The Problem

Current AI evaluation is often opaque, relying on self-reported metrics from model developers. There is a lack of independent, standardized testing for critical failure modes in real-world deployment.

Intent

We aim to build a suite of public tests that anyone can run. Watchtower will prioritize transparency in methodology and data availability.

Note: This project is currently in the research and planning phase. No release dates are available.