this post was submitted on 18 Jun 2023
104 points (100.0% liked)
Technology
37705 readers
99 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
LLM generated text can also be easily detected provided you can figure out which model it came from and the weights within it. For people training models, this won't be hard to do.
I agree with the take that getting better and better datasets for training is going to get easier over time, rather than harder. The story of AlphaZero is a good example of this too - the best chess AI quickly trounced any AI trained on human games simply by playing against itself. To me, that suggests that training on LLM output will lead to even better results, since you can generate so much more of it.
The chess engine's training is anchored by the win/lose outcome of the game. LLM training is anchored by what humans like to read and write. This means that a human needs to somehow be in the loop.
I think OpenAI's own chatGPT detector had double digit false negative and positive rates. I expect as diversity of LLMs proliferates, it will become increasingly harder to detect.