You must log in or register to comment.
Wow, the text generator that doesn’t actually understand what it’s “writing” is making mistakes? Who could have seen that coming?
I once asked one to write a basic 50-line Python program (just to flesh things out), and it made so many basic errors that any first-year CS student could catch. Nobody should trust LLMs with anything related to security, FFS.
My experience with ChatGPT goes like this:
- Write me a block of code that makes x thing
- Certainly, here’s your code
- Me: This is wrong.
- You’re right, this is the correct version
- Me: This is wrong again.
- You’re right, this is the correct version
- Me: Wrong again, you piece of junk.
- I’m sorry, this is the correct version.
- (even more useless code) … and so on.