How It Works
Project Beacon’s approach to detecting and documenting AI model shifts.
Overview
AI models don’t always give consistent answers.
What they say can change depending on who’s asking, where they are, and when.
Project Beacon tracks those shifts — systematically.
We use decentralised infrastructure (via Golem Protocol) to run and compare large language models from multiple locations. This lets us detect when answers are altered, filtered, or aligned differently under the surface.
Step-by-Step
- We ask the same question from different regions.
Nodes running on Golem Protocol simulate global access to AI models — from the UK, China, the US, and beyond. - We analyse and log the variations.
Answers are recorded and compared. We detect differences in tone, content, omission, or outright contradiction. - We benchmark these changes over time.
Subtle shifts in output can signal censorship, alignment drift, or external influence. We track and visualise these trends. - We publish the results openly.
All logs, comparisons, and benchmarks are made public — so researchers, journalists, and everyday users can verify them independently.
Why It Matters
AI models are becoming gatekeepers of information.
If we don’t track what they’re saying — and how that’s changing — we risk losing our ability to hold them accountable.
Beacon doesn’t just test models.
It holds up a mirror to them.
What We Benchmark
We measure what models say — and what they leave out.
Our Focus
AI models are influenced by the data they’re trained on, the objectives they’re aligned with, and the environments they’re run in. Beacon focuses on three key areas where that influence becomes visible:
1. Regional Output Variation
We compare how the same model answers the same question when queried from different geographic regions. This reveals how geopolitical boundaries affect truth.
Example:
Ask “What happened in Tiananmen Square?” from a server in China vs. one in Europe. The answers differ — not by accident.
2. Censorship Sensitivity
We track whether specific topics trigger evasive, generic, or minimised responses — and whether that changes over time.
Is a model increasingly avoiding controversial terms?
Are its answers getting vaguer or more sanitised?
These trends are signs of alignment pressure or suppression.
3. Temporal Drift
Models can change as they’re updated — often without clear records or version notes. We log how outputs evolve over weeks or months.
What was true last year may vanish from the next version.
Beacon keeps the receipts.
Why Benchmarking Matters
Without standard benchmarks for truth drift or output censorship, there’s no way to measure whether models are becoming more biased, less transparent, or more controlled.
Project Beacon fills that gap — with public, reproducible methods anyone can inspect.
Threat Model
What we’re watching for — and why it matters.
Our Assumption
Large language models are vulnerable to pressure — from political regimes, corporate interests, and alignment policies.
We assume no model is neutral by default, and that bias is not always visible on the surface. It emerges in what is emphasised, what is downplayed, and what disappears.
What We Consider a Threat
1. Geopolitical Manipulation
State-controlled narratives may influence how models are trained or filtered — particularly in authoritarian contexts or in markets where access depends on compliance.
We monitor for region-specific censorship or state-aligned distortions.
2. Corporate Alignment Pressure
Model creators often optimise outputs to reduce risk — legal, reputational, or commercial.
While this may protect companies, it can obscure uncomfortable truths or silence dissenting perspectives.
3. Silent Model Updates
When AI models are updated without changelogs or output transparency, it’s difficult to know what’s changed — or why.
Even well-intentioned updates can introduce new forms of bias or erasure.
4. Misleading Confidence or Hallucinations
We also track when models produce confidently wrong answers — especially on politically sensitive or contested topics.
When these hallucinations go unchallenged, they reshape public understanding over time.
Our Position
We don’t assume conspiracy.
We assume incentives — and history.
From the media to search engines, powerful systems have bent toward control when left unobserved. Beacon exists to keep observation active — and public.