d_model is a fundamental AI research lab. We're focused on automating interpretability and, more broadly, the science of deep learning — building the tools that let humans understand and steer the systems they train.
Other vendors build hackable environments for specific tools. We work on models themselves — making them more understandable[1][2], more cooperative[3], and more aligned.
We think alignment is a real and unsolved problem, but an empirical one. The path forward looks like serious experimental work, layers of inoculation, and automating the research itself[4] — a marriage of science and philosophy.
If you want to work on the parts of the problem that matter the most, talk to us.