Generative AI has undoubtedly been the most talked about technological innovation in 2023, and perhaps in the last decade. Anthropic debuted its public facing Claude chatbot in July and has been serving customers like Slack, Zoom and Notion for almost a year via its Claude API. But prior to that, the company spent over two years researching AI safety to make Claude as helpful, honest and harmless as possible.
In this talk, Ben Mann, co-founder and product engineering lead at Anthropic will outline his top lessons learned building Claude including techniques for aligning an AI system with human values and bootstrapping scalable oversight of model outputs.
Benjamin Mann is a co-founder and member of the technical staff at Anthropic, an AI safety startup based in San Francisco. He was previously a member of the technical staff at OpenAI, where he worked on infrastructure, efficiency, and safety for GPT-3. Before that, Mann was a senior software engineer at Google, where he helped build Google's carpooling service Waze Carpool. He has also worked at research organizations like the Machine Intelligence Research Institute and startups focusing on AI and automation. He studied computer science at Columbia University. His goal is to develop AI systems that are helpful, harmless, and honest.