How is an AI agent any different than any other software just because it does inference with a LLM? If I order something from their website and I get overcharged due to a bug, are they also not responsible? It’s not like agents can’t be tested or like guardrails can’t be put into place.
I know as a software engineer, I’m responsible for the code in any PR that has my name on it, regardless of what tools I may have used to generate the code, including AI. Are their dev teams not responsible for making sure their shit works?