Socratic Prompting
I’ve noticed that some prompts produce answers that are not obviously wrong, but still feel a little too smooth. The model lands somewhere plausible, gives a tidy paragraph or two, and moves on. If the question is simple, that’s often fine. But for anything with hidden assumptions, tradeoffs, or location-specific details, I’ve started to suspect …