Study Finds That 52 Percent of ChatGPT Answers to Programming Questions Are Wrong
44 comments
@darkling@mstdn.social @cstross@wandering.shop It's the wrong question. The correct question is, "What proportion of answers to programming questions given by programmers who understand the language and the question are wrong." @unlucio @darkling @cstross If I spend longer than it would have taken me to do it myself helping ChatGPT through a problem, I've wasted my time @darkling @cstross An LLM is an even worse version of some asshole that weighs in on *everything* and asserts wrong answers just as confidently as right answers @ezeno@mastodon.uno @cstross@wandering.shop I once knew someone who managed to achieve a grade of 18% on a five-options multiple-choice exam... @cstross@wandering.shop Color me soooooooooo shocked. I worked for over 30 years as a geophysicist in the oil business interpreting seismic data. Companies always tried to sell us artificial intelligence software to allow the computer do the interpretation for us going back as far as the 90s. Up until my retirement about 7 years ago I found that geophysicists would spend more time correcting the "interpretation" the machine did than it would have taken to do the interpretation themselves. I do not trust this "artificial" intelligence. @cstross Every piece of sample code ever provided to me by a project manager or non-technical co-worker that used functions that didn’t exist, had obvious syntax errors, or did things I considered insane…. Turned out to be from an LLM. Their attempt to show me how easy it was to write the necessary code turned into a lesson in why programmers should just be left alone to do the necessary thing. @JustinDerrick@mstdn.ca @cstross@wandering.shop this is a sufficiently well-known problem that there is now an established class of software attacks that is based upon predicting fictitious library names likely to be generated by ChatGPT or other LLMs, then publishing libraries under those names containing malicious code. I am thinking about all the people I see on Hacker News raving about how EFFICIENT talking to AIs is making them and giggling. I am also hoping I never have to deal with any system they were involved in building... Honestly, isn't it surprising it's that low? I'd've thought it would bork a lot more questions than that. Real Programmers don't use languages that computers can *output*. Systems that don't require a keyboard with at least 15 extra non-USB keys, and a specialized foot-pedal connected to the GPIO pins, are mere children's toys used by web developers and JavaScript vapers. And Real Programmers don't interpret figures like 52/100 in anything other than octal. (Heh! It has been a while since someone set up a Real Programmers joke.) @cstross that's the fun part. companies are going to go deep on ai coding only to absolutely fuck themselves over. the code ai can generate is often remedial code that i would never run on a production server. i've never seen it write code that isn't shit. companies think engineers are expensive, they're about to fuck around and find out. @cstross In my (admittedly very limited) experience, I'm not even getting internally consistent answers, such that variables change name and other errors. But the value, such that it is, has been in getting suggestions for new ways of solving something, which I then can do something with using actual reference documentation. @cstross ...one way to be rid of AI garbage... @cstross …which is why I don’t use it for that purpose. Of course, I’ve asked ChatGPT for nonessential programming, like making a Quine in 6502 assembly, and it succeeded in doing that. But for normal work, I don’t dare touch it because if it makes a mistake, then I will have no idea how to fix it. |
@cstross ... but how many human answers to programming questions are wrong?
(OK, probably not 52%, but I bet you it's higher than you first thought...)