No way, who saw this coming?
Isolate this shit. Test this shit. Log this shit.
I don't know a good solution where there needs to be zero trust of these types of things, as this requires full trust for it to do what it needs.
I read that Google’s Gemini team came up with a system where the LLM has to write simplified Python‐like code to describe the operations it wants to perform. It has to describe what it is going to do, including what information is being pulled from where and who it is being sent to. Before any operations are performed they are vetted by standard, human configured security software.
LLMs can be useful tools but they require guard rails.
Remove the pod bay guard rails.