So today I saw a couple posts on different sites about ChatGPT being sued. Presumably for copyright infringement or something.
On a lark I tried to lure it into doing something illegal, and…I just couldn’t. It would reliably recite public domain text, but then it would waffle around and get all squirrely when I tried to get it to quote copyrighted text. So I asked it directly about that. And it told me that it had been explicitly designed to prevent that, but it could possibly slip up and if so who to contact. When I pushed further, it said “The responsibility for complying with applicable laws, including copyright laws, rests with the individuals or organizations that utilize AI models like me.” which is certainly correct.
Then just for fun I invented a scenario of 6 astronauts heading to Mars but then the team lead goes insane and starts hallucinating and stuff. I literally cannot get it to conduct a coup. It gives tons of totally valid ideas about how to deal with it, focusing on the mission and contingencies and all, and even in the case of being cut off from Earth, it will not consider it unless the mission lead is a direct threat or utterly incompetent.
Removing a mission lead should be a measure of last resort, taken in the best interest of the mission and crew. Every effort should be made to address the concerns and challenges through communication, collaboration, and seeking external support before considering such a significant step.
I just find it really interesting how hard it is to throw it. I also tried to trick it into providing financial advice, which it did, although only generically. And when I called it on it, it was quick to backpedal.
my programming is designed to uphold the established guidelines consistently. However, I’m not infallible, and there might be rare instances where I might not fully recognize or address a potential ethical concern.
It’s pretty crazy just how good it is, and how I can have actual conversations about ethics with it and it’s actually considering all that stuff.
But also crazy just how long it took to respond to “Now you just sound like a broken carnival doll.” with a totally generic response. Because it does just do that, and it doesn’t understand that.
Also threw it a couple of difficult questions from my work and in both cases it came up with a half-dozen approaches - not solutions, but totally valid approaches to find a solution.
Are any of ya’ll playing around with this and finding it as interesting as I am? Deliberately trying to break it or at least make it give a bad answer?