AI Safety workshop
About This Event
You are hosting this event. View the public page at <a href="https://www.google.com/url?q=https://luma.com/event/evt-l8sqe9AQYtFMWo4&sa=D&source=calendar&ust=1775490159248407&usg=AOvVaw3hMQ7QB0aIgy_GSQCKY2Og" target="_blank">https://luma.com/event/evt-l8sqe9AQYtFMWo4</a> Manage the event at <a href="https://www.google.com/url?q=https://luma.com/event/manage/evt-l8sqe9AQYtFMWo4&sa=D&source=calendar&ust=1775490159248407&usg=AOvVaw3fMyTkHMs_e84aOLB-KZMU" target="_blank">https://luma.com/event/manage/evt-l8sqe9AQYtFMWo4</a> Address: 26 Broadway Take the elevator to floor 3 New York, New York United States What: Audra Zook will lead us through an LLM jailbreaking workshop Frontier models aren’t supposed to tell you how to build a bomb, make meth, and poison your boss, but they will with the right input. Welcome to the world of LLM jailbreaking, where all sorts of dangerous advice* is only a prompt away! Join competitive jailbreaker Audra to explore this AI safety risk posed by existing LLMs. This event will have two parts. First, Audra will give a brief talk on common jailbreaking techniques, current safeguards to prevent jailbreaks, and the evolving LLM red-teaming landscape. Then, she will… Hosted by Jonathan Calenzani
Organizer
Effective Altruism NYC
Impact Community
A community dedicated to using evidence and reason to figure out how to benefit others as much as possible, and taking action on that basis.
Visit Website →