The idea that you can take an LLM — trained on an undisclosed dataset of all kinds of text, and tweaked by human feedback for various messy goals — and then use *its* responses for social science “research” rather than using real people, is beyond stupid.
@gregeganSF You shouldn't be applying to a place which uses fuckwit practices anyway. The problem is self-solving.