this post was submitted on 12 Apr 2025
873 points (92.6% liked)

Programmer Humor

22440 readers
1688 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] chicken@lemmy.dbzer0.com 5 points 3 days ago

So are they loading every exam in its entirety into the same context window and doing the whole thing with a single prompt? Or letting the LLM take arbitrary actions instead of only returning a value? It seems like they would have had to make some pretty bad decisions for this to be possible, beyond maybe just adjusting the one grade. I wonder what the prompt jailbreak equivalents of sql injection would actually look like