As I think about how to design assessments that are not readily, answered by large language models, I find myself coming back to something I’ve been interested in for a while: rewarding students for understanding and honestly conveying their degree of certainty in their answers.
This is something that ChatGPT and the likes are notoriously bad at in their current instantiations.
In its simplest form, this might mean taking more points off for wrong answers than for questions left blank.
@mavori @ct_bergstrom As @DonnaG mentioned, I'd be concerned about how confidence differences might play out in a wagering scheme.
And, if these are true/false tests, with @mavori 's scheme someone always betting certain would have an advantage over someone always betting uncertain if the expected number of correct answers in the test > 50% - which is usually the case.