RESEARCH · ALL CLUSTERS · ALL THREADS
Privacy, safety, accessibility, and open model development.
Each cluster targets open problems with no settled solution. Preprints, datasets, tooling, and negative results are published as they become available.
FEATURED.RESEARCH
Generalising the Wyvern trust chain into a framework any on-device model can adopt. Formal verification of model provenance and inference integrity, proving what ran, where it ran, and what it accessed.
Translating technical safety properties into compliance-ready frameworks engineers can build against. Mapping model transparency, explainability, and red-teaming results onto EU AI Act requirements.
How communities govern AI systems that affect them. This thread runs on the protocols it studies. Contributors earn a vote on cluster direction.
ALL.CLUSTERS
Training ML models across distributed nodes without centralising raw data. Open protocols for cross-organisational federated training with differential privacy.
Quantisation, pruning, and distillation research aimed at transformer-scale models that run privately on consumer hardware. Wyvern Engine is the reference implementation.
Formal verification of model provenance and inference integrity. Generalising the Wyvern trust chain into a framework any ondevice model can adopt.
Privacy-safe training datasets using generative models. Realistic data without exposing real users.
Making neural networks interpretable to the humans affected by their decisions. Layered explanation frameworks that work for non-experts, not just ML researchers.
Systematic measurement of discriminatory outcomes in model outputs. Current fairness benchmarks largely cover English-language and Western contexts; this thread develops evaluations for underrepresented languages and cultural settings.
Translating technical safety properties into compliance-ready frameworks engineers can actually build against. Mapping transparency, explainability, and red-teaming onto EU AI Act requirements.
Open adversarial-testing protocols any lab can run against their own systems before deployment. Published methodologies for red-teaming that cover common attack surfaces and failure modes.
Lightweight deployment pipelines for healthcare, education, agriculture and legal aid in low-income contexts.
Extending transformers to languages underrepresented in training data. English-only excludes people. Native-language annotators and new methodology required.
How AI integrates into daily work across real task environments. Studies what breaks, what helps, and what creates dependency in production deployments rather than controlled settings.
Tooling and frameworks letting a two-person lab compete on model quality, not compute budget. Innovation should not require a datacenter.
Cluster D is governed by contributors. Threads are proposed, seconded, and scoped by the community. The foundation has one vote, same as anyone else. We're building the governance model now, in public.
Publishing weights, training configs, and evaluation results openly. Reproducible science is a standard. Community-maintained checklists and CI tooling for every release.
RLHF, DPO and Constitutional AI with open datasets and preference models the community can audit and extend. Looking for contributors with multilingual annotation experience.
Communities govern AI systems that affect them, not corporations. This thread itself runs on the protocols it studies. Sustained contributors earn a vote on cluster direction.
Next-generation evaluation frameworks that measure what actually matters. Not what's easiest to game. Community-proposed benchmarks with full methodology published.
GET.INVOLVED
Research threads are open for contribution — propose a new thread, contribute to an existing one, or share datasets and methods.
// Coming soon
OPEN.PROBLEMS
Each is an open research question with no settled answer. If you have a lead, a dataset, a method, or just an opinion, these are the problems where your input matters most.
PUBLICATIONS.FEED
preprints · datasets · workshop papers · software releases, openly licensed
View all papers