Posted by gregpr07 7 hours ago
Essentially it’s just: remove .py files an execute del os.environ[“SESSION_TOKEN“]? This doesn’t really sound very secure, there are a number of ways to bypass both of these.
It’s just security through obscurity
I'd be pretty keen to actually hear more about the Unikraft setup and other deeper details about the agent sandboxes regarding the tradeoffs and optimizations made. All the components are there but has someone open-sourced a more plug-and-play setup like this?
https://github.com/unikraft/cli
Feedback is very much appreciated, we're listening! :)
The problem is not what the LLM shouldn't have access to, it's what it does have access to.
The usefulness of LLMs is severely limited while they lack the ability to separate instructions and data, or as Yann LeCun said, predict the consequences of their actions.
Of all the problems in agent security, sandboxing solves the easiest problem.