Even claiming such a thing is it’s basically posting a huge target on your own back. Regardless of how long it might have taken for those models to be hacked, that timeline is now much shorter and certainly guaranteed.
Technology
Which posts fit here?
Anything that is at least tangentially connected to the technology, social media platforms, informational technologies and tech policy.
Post guidelines
[Opinion] prefix
Opinion (op-ed) articles must use [Opinion] prefix before the title.
Rules
1. English only
Title and associated content has to be in English.
2. Use original link
Post URL should be the original link to the article (even if paywalled) and archived copies left in the body. It allows avoiding duplicate posts when cross-posting.
3. Respectful communication
All communication has to be respectful of differing opinions, viewpoints, and experiences.
4. Inclusivity
Everyone is welcome here regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, education, socio-economic status, nationality, personal appearance, race, caste, color, religion, or sexual identity and orientation.
5. Ad hominem attacks
Any kind of personal attacks are expressly forbidden. If you can't argue your position without attacking a person's character, you already lost the argument.
6. Off-topic tangents
Stay on topic. Keep it relevant.
7. Instance rules may apply
If something is not covered by community rules, but are against lemmy.zip instance rules, they will be enforced.
Companion communities
!globalnews@lemmy.zip
!interestingshare@lemmy.zip
Icon attribution | Banner attribution
If someone is interested in moderating this community, message @brikox@lemmy.zip.
They want people to try. It’s independent bug testing that costs only as much as publishing an article on a website and incrementing a version number
"AI" has a massive inability (or is purposefully deceptive) to distinguish the difference between bugs, which can be fixed, and fundamental aspects of the technology that disqualify it from various applications.
I think the more likely story is that they know this can be done, know about this particular jailbreak person, can replicate their work (because they didn't so anything they hadn't done with previous models in the first place), and are straight up lying and betting the people that matter to their next investment round (scam continuation) won't catch wind.
You're giving these grifters way too much credit.
That's not really compelling because people would try regardless
They have a 500k bounty for jailbreaks.
They have open beta programs for that while also not having to tell hilarious and bold face lies that end up embarrassing them.
I mean, it's fundamental to LLM technology that they listen to user inputs. Those inputs are probablistic in terms of their effects on outputs... So you're always going to be able to manipulate the outputs, which is kind of the premise of the technology.
It will always be prone to that sort of jailbreak. Feed it vocab, it outputs vocab. Feed it permissive vocab, it outputs permissive vocab.
Ok? Either openai knows that and lies about their capabilities, or they don't know it and are incompetent. That's the real story here.
It also could be they're both liars and incompetent.
I think the answer is that they are incompetent but also that they are lying about their capabilities. Why else have they rushed everything so much and promised so much?
They don't really care about the fallout, they are just here to make big promises and large amounts of money on their shiny new tech.