
OpenAI aims to ensure AI benefits humanity, prioritizing safety, diversity, and widespread benefits.
OpenAI is seeking a Research Engineer/Scientist to join their Interpretability team. The team studies internal representations of deep learning models and is interested in applying their understanding to ensure the safety of powerful AI systems. The role involves developing and publishing research on techniques for understanding representations of deep networks, engineering infrastructure for studying model internals at scale, and collaborating across teams to work on projects.