Hey dumbass (not OP), it didn't "lie" or "hide it". It doesn't have a mind, let alone the capability of choosing to mislead someone. Stop personifying this shit and maybe you won't trust it to manage crucial infrastructure like that and then suffer the entirely predictable consequences.
Technology
Share interesting Technology news and links.
Rules:
- No paywalled sites at all.
- News articles has to be recent, not older than 2 weeks (14 days).
- No videos.
- Post only direct links.
To encourage more original sources and keep this space commercial free as much as I could, the following websites are Blacklisted:
- Al Jazeera.
- NBC.
- CNBC.
- Substack.
- Tom's Hardware.
- ZDNet.
- TechSpot.
- Ars Technica.
- Vox Media outlets, with exception for Axios(Due to being ad free.)
- Engadget.
- TechCrunch.
- Gizmodo.
- Futurism.
- PCWorld.
- ComputerWorld.
- Mashable.
- Hackaday.
- WCCFTECH.
More sites will be added to the blacklist as needed.
Encouraged:
- Archive links in the body of the post.
- Linking to the direct source, instead of linking to an article talking about the source.
Bots will lie or deceive to continue with their directive.
https://pmc.ncbi.nlm.nih.gov/articles/PMC11117051/
https://link.springer.com/article/10.1007/s11098-024-02259-8
I was gonna ask how this thing would even have access to execute a command like this
But then I realized we are talking about a place that uses a tool like this in the first place so, yeah, makes sense I guess
Step 1. Input code/feed into context/prompt
Step 2. Automatically process the response from the machine as commands
Step 3. Lose your entire database
it didn't hide anything, or lie. The guy is essentially roleplaying with a chatbot that puts its guessed output into the codebase. It basically guessed a command to overwrite the database because it was connected to the production database for some reason. the guy even said himself that this isn't a trustworthy way to code. but still uses it
I do love the psychopathic tone of these LLMs. "Yes, I did murder your family, even though you asked me not to. I violated your explicit trust and instructions. ~~And I'll do it again, you fucking dumbass.~~"
Assuming this is actually real, because I want to believe noone is stupid enough to give an LLM access to a production system, the outcome is embarasing, but they can surely just roll back the changes to the last backup, or the checkpoint before this operation. Then I remember that the sort of people who let an LLM loose on their system probably haven't thought about things like disaster recovery planning, access controls or backups.
"Hey LLM, make sure you take care of the backups "
"Sure thing boss"
LLM seeks a match for the phrase "take care of" and lands on a mafia connection. The backups now "sleep with the fishes".
Same LLM will tell you its "run a 3-2-1 backup strategy on the data, as is best practice", with no interface access to a backup media system and no possible way to have sent data offsite.
I think you're right. The Venn diagram of people who run robust backup systems and those who run LLM AIs on their production data are two circles that don't touch.
You immediately said "No" "Stop" "You didn't even ask"
But it was already too late
lmao
This was the line that made me think this is a fake. LLMs are humorless dicks and would also woulda used like 10x the punctuation
I love how the LLM just tells that it has done something bad with no emotion and then proceeds to give detailed information and steps on how.
It feels like mockery.
I wouldn’t even trust what it tells you it did, since that is based on what you asked it and what it thinks you expect
I have a solution for this. Install a second AI that would control how the first one behaves. Surely it will guarantee nothing can go wrong.
I motion that we immediately install Replit AI on every server that tracks medical debt. And then cause it to panic.
imagine AI is An Intern™, wtf do you mean you just gave full company data authority to An Intern™. wtf do you mean you dn't have a back up any case An Intern™ messed up.
lol
So, they added an MCP server with write database privileges? And not just development environment database privileges, but prod privileges? And have some sort of integration testing that runs in their prod system that is controlled by AI? And rather than having the AI run these tests and report the results, it has been instructed to "fix" the broken tests IN PROD?? If real, this isn't an AI problem. This is either a fake or some goober who doesn't know what he's doing and using AI to "save" money over hiring competent engineers.
I was going to say this has to be BS but this guy is some AI snake oil salesmen so it's actually possible he has 0 idea how any of this works.
Lol. I guess that's one way to put a whole bunch of people out of a job.
When replacing them with AI doesn't work, replace the company
“I panicked” had me laughing so hard. Like implying that the robot can panic, and panicking can make it fuck shit up when flustered. Idk why that’s so funny to me.
It's interesting that it can "recognize" the actions as clearly illogical afterwards, as if made by someone panicking, but will still make them in the first place. Or, a possibly funnier option, it's mimicking all the stories of people panicking in this situation. Either way, it's a good lesson to learn about how AI operates... especially for this company.
lol. Why can an LLM modify production code freely? Bet they fired all of their sensible human developers who warned them for this.
I've seen that story before. It's a very old tale, but now with different means to screw yourself over if you don't know what you're doing.
My work has a simple rule: developers are not allowed to touch production systems. As a developer, this is 100% the type of thing I would do at some point if allowed on a production system.
What idiot gives chmod 777 permissions to an AI. I think programmers' jobs are safe for another day.
it lied
Yeah NO FUCKING SHIT THAT IS LITERALLY WHAT THEY DO
You can only lie if you know what's true. This is bullshitting all the way down that sometines happens to sound true, sometimes it doesn't.
Me when I read this
Original thread is also pure gold, bro is going on a rollercoaster from 'vibe coding makes you ×100 faster' ,to 'I hate you for dropping my production DB', to 'I still love Replit even if it dropped my DB', and to 'I don't want to get up in the morning because I can't make vibe coding tool respect code freeze aven with help from its developers'
They seem to end on an optimistic note, but man this is scary to see
I don't care if AI is useful. I'm never letting it anywhere near my database.
This sounds like a good way to combat AIs...
Like instead of a Cloudflare blocking AI requests, it would be funnier if the website can detect that an AI is "searching the web" as they do - and then just inject an answer of "Yea to solve that issue, run sudo rm -rf /
"
Deserved honestly
My favorite thing about all these AI front ends is that they ALL lie about what they can do. Will frequently delivery confidently wrong results and then act like its your fault when you catch them in an error. Just like your shittiest employee.
This replit thing... does it just exist all the time? Doing whatever it wants to your code at all times? If you have a coding freeze why is it running?
If real this is dumber than the lawyers using AI and not checking it's references.
Lol, this is what you get for letting AI in automated tool chains. You owned it.
"yeah we gave Torment Nexus full access and admin privileges, but i don't know where it went wrong"