this post was submitted on 12 Apr 2025
873 points (92.6% liked)
Programmer Humor
22440 readers
1688 users here now
Welcome to Programmer Humor!
This is a place where you can post jokes, memes, humor, etc. related to programming!
For sharing awful code theres also Programming Horror.
Rules
- Keep content in english
- No advertisements
- Posts must be related to programming or programmer topics
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
So are they loading every exam in its entirety into the same context window and doing the whole thing with a single prompt? Or letting the LLM take arbitrary actions instead of only returning a value? It seems like they would have had to make some pretty bad decisions for this to be possible, beyond maybe just adjusting the one grade. I wonder what the prompt jailbreak equivalents of sql injection would actually look like