Some recent projects I have been involved in center on the idea of public model specifications with citations. As many of you know, Frontier AI Labs are already training AI systems using “constitutions” or “model specifications”—plain language descriptions of intended behavior. The problem is that these are not truly verifiable. If a chatbot outputs something that seems to violate its model spec, you cannot really get an explanation. If you do ask, it might hallucinate a response, but it does not reveal how it actually works internally.