Yeah I get it but look - you say "hey Intern give me this data" and they come back with "here is a banana and a receipt for my Uber ride to buy bananas also I brought coffee for everyone but you please validate my parking" and you go "well, wait. Hang on. How did bananas and coffee come into the discussion, also this wasn't part of the scope and I recognize that's my fault for not giving you a proper brief" and the intern looks at you askance, already viewing you in terms of tik tok memes BUT when you ask how bananas entered the chat, you'll at least get some reasoning. AI gonna tell you bananas are data. And if you say try again you might get bananas, you might get oranges, you might get data. And if you use that data, you might find out later on that it's actually bananas. And if you query it enough you might come across some form of "this data is bananas" which somehow got you into the "desktop fruit" dataset but there's nothing you can do to get fruit out of the data because what you're doing? Is querying a black box that belongs to someone else and they don't really know where the Markov turned left, either and even if they did, there's nothing they can do except hand-tag "do not give bananas in response to data" which is great unless you're a grocery wholesaler. The reliability of interns is an easily assessed quality. You can back-test any errors. You can tune your input to maximize your output. If you don't like the output you can try your luck and run it again? But any answer you get, you're going to have to try again a different way anyway since "factual correctness" is simply not a vector within the GPT space.