I can’t remember where I read it but someone said “LLM’s provide three types of answer: so vague as to be useless, directly plagiarized from a source and reworded, or flat out wrong but confidently stated as the truth.” I’m probably butchering the quote, but that was the gist of it.
Hold on let me have chat gpt rephrase that for you.
I’m not exactly sure of the source, but there was a statement suggesting that language models offer three kinds of responses: ones that are too general to be of any value, those that essentially mimic existing content in a slightly altered form, and assertions that are completely incorrect yet presented with unwavering certainty. I might be paraphrasing inaccurately, but that was the essence.
Are all responses so non-committal? “I’m not exactly sure”, “I might be paraphrasing inaccurately”.
I hope this sort of phrasing doesn’t make its way into common usage by students and early careers people. Learning to run away from liability at every opportunity is not going to help them.
This is just chat gpt rephrasing the comment above me. Don’t worry though, when chat gpt is wrong it’s quite confident sounding and even cites sources that don’t exist but look quite convincing!
To me the answers are useful enough and I appreciate that it understands vague questions. When I don’t know enough about a topic to know what terms to punch into a search engine, I can use ChatGPT as a first step and go from there.
I can’t remember where I read it but someone said “LLM’s provide three types of answer: so vague as to be useless, directly plagiarized from a source and reworded, or flat out wrong but confidently stated as the truth.” I’m probably butchering the quote, but that was the gist of it.
Hold on let me have chat gpt rephrase that for you.
Are all responses so non-committal? “I’m not exactly sure”, “I might be paraphrasing inaccurately”.
I hope this sort of phrasing doesn’t make its way into common usage by students and early careers people. Learning to run away from liability at every opportunity is not going to help them.
This is just chat gpt rephrasing the comment above me. Don’t worry though, when chat gpt is wrong it’s quite confident sounding and even cites sources that don’t exist but look quite convincing!
So the same as answers on Reddit then
To me the answers are useful enough and I appreciate that it understands vague questions. When I don’t know enough about a topic to know what terms to punch into a search engine, I can use ChatGPT as a first step and go from there.