
Hack your way into LLMs
Dive deep into the intriguing world of mechanistic interpretability at EPFL! π With SAIL hosting an engaging hackathon, you'll have the chance to explore a curated adversarial dataset with five innovative coding models. Your mission? Understand what's happening internally when these models face potentially harmful requests.
Two Tracks to Explore:
π Probes as Monitors: Classify internal activations to detect harmful compliance before the model generates potentially risky outputs.
π― Attribution: Identify which specific inputs push a model toward supporting an attacker.
Whatβs Included:
- Claude Code Credits
- GPU Instances
- Mentorship
- Food throughout the weekend
This is your chance to achieve a publishable outcome in just 48 hours! π
Prizes Awaiting:
π° CHF 1,000
π° CHF 500
π° CHF 300
Bring your PyTorch fluency and curious mind focused on model internals β that's all you need to participate!
Important:
π Register by May 6th
π
Event Date: May 9-10
π₯ Team Size: Up to 5 members
π Register Here
Join us for a thrilling weekend of innovation and discovery!
NOTE: We cannot guarantee the accuracy of the information we provide about this event. Visit the event's website to verify details such as date, opening hours, prices and location.




