
Secure execution of code generated by Large Language Models
Learn how to securely execute LLM-generated code. Explore attack vectors, threat modeling, and AWS-based mitigation strategies for code generated by Large Language Models.
You work as a lead data scientist and ML Architect for the fitness and health tracker app 'AnyFitness' and want to allow your users to interrogate their fitness and health data. Your CTO is delighted, but health data is very sensitive so you need to really be sure that the privacy and security of the handled data is not compromised.
- A theoretical introduction to common attack vectors against LLMs, and how to mitigate them
- A practical guideline on how to approach security risks when dealing with LLMs, by using the threat modeling technique
- A case study on how to securely create an execution environment for LLM generated code through AWS Lambda functions
- A summary about how AWS services can help you in this journey
- LLM02: Insecure Output Handling
OWASP Definition: Insecure Output Handling refers specifically to insufficient validation, sanitization, and handling of the outputs generated by large language models before they are passed downstream to other components and systems. Since LLM-generated content can be controlled by prompt input, this behavior is similar to providing users indirect access to additional functionality.
Why is it relevant? This is the most directly relevant risk when executing LLM-generated code, because the code is generated by an LLM and is thus not fully under our control. - LLM06: Sensitive Information Disclosure
OWASP Definition: LLM applications have the potential to reveal sensitive information, proprietary algorithms, or other confidential details through their output. This can result in unauthorized access to sensitive data, intellectual property, privacy violations, and other security breaches.
Why is it relevant? The dynamically generated code could potentially contain or access sensitive information if not properly controlled. - LLM08: Excessive Agency
OWASP Definition: An LLM-based system is often granted a degree of agency by its developer – the ability to interface with other systems and undertake actions in response to a prompt. The decision over which functions to invoke may also be delegated to an LLM ‘agent’ to dynamically determine based on input prompt or LLM output. Excessive Agency is the vulnerability that enables damaging actions to be performed in response to unexpected/ambiguous outputs from an LLM (regardless of what is causing the LLM to malfunction; be it hallucination/confabulation, direct/indirect prompt injection, malicious plugin, poorly-engineered benign prompts, or just a poorly-performing model).
Why is it relevant? Giving an LLM the ability to generate and execute code grants it a high degree of agency that needs to be carefully controlled. - LLM09: Overreliance
OWASP Definition: Overreliance can occur when an LLM produces erroneous information and provides it in an authoritative manner. While LLMs can produce creative and informative content, they can also generate content that is factually incorrect, inappropriate or unsafe. This is referred to as hallucination or confabulation. When people or systems trust this information without oversight or confirmation it can result in a security breach, misinformation, miscommunication, legal issues, and reputational damage.
Why is it relevant? Blindly trusting and executing LLM-generated code without proper oversight / controls can lead to security vulnerabilities.
- Robust Access Controls
Implement strict access controls to limit who can interact with the LLM and what kind of inputs they can provide. Make sure to follow a least-privilege principle throughout the whole pipeline, especially when additional elements such as vector stores for RAG are involved. - Data Classification and Handling
Perform a risk analysis and classification of any involved data. This includes training data, supporting data such as knowledge bases (for RAG), etc. Consider removing or sanitizing sensitive data from your workflows. - Input Validation and Sanitization
Use input validation to filter out obvious malicious inputs and input sanitization to neutralize potentially harmful elements within user inputs. - Output Validation and Sanitization
Validate and sanitize the data that was produced by the LLM, before further processing it. - Output Monitoring
Employ real-time monitoring of the LLM's outputs to detect and respond to suspicious activities promptly. Consider logging prompts and responses to allow anomaly detection. - Verify: Is LLM-generated data needed, or is the data deterministic / known?
Only use LLM-generated data when you need it. If an information can be taken from a deterministic / verified source, use it from there. An example can be a user ID: If you know a user information in your code, use the proven source instead of a generated one. - Regular Security Updates
Continuously install patches, update security protocols and defenses to keep up with evolving techniques and tactics. - Awareness and Training
Educate developers and users about the potential risks and best practices for interacting with LLMs securely.
- What are we working on?
- What can go wrong?
- What are we going to do about it?
- Did we do a good job?
- The user uploads a .csv file, or the data is supplied by a backend system for the user interaction. You can imagine the .csv file as mental placeholder for any data provisioning you might want to implement for your app.
- The LLM is supplied with the user query as well as some information about the data it should write code for.
- Code and data are supplied to the Code interpreter (we place the python symbol as a placeholder for other programming languages).
- User uploads a CSV containing medical data to a cloud storage
- User navigates to web application
- User sends prompt on client side, which is forwarded to logic by API
- Logic downloads the CSV containing medical data from cloud storage, to extract some basic information (e.g. number of rows, name of columns, etc.)
- Logic sends CSV information and user prompt to the LLM, which generates and returns python code
- Logic sends the generated code and pointer to the CSV to the execution environment
- Execution environment downloads the CSV containing medical data from the cloud storage
- The execution environment executes the python code and sends back the response to the logic
- The execution environment writes to the execution environment logs
- The logic environment writes to the logic environment logs
- Logic sends back the final response to the user
A [threat source] with [pre-requisites],can [threat action],which leads to [threat impact],negatively impacting [goal] of [impacted assets].
- (a) A threat actor with access to the public facing application, can inject malicious prompts which overwrite existing system prompts, which leads to the LLM generating malicious code and the execution environment executing it, for example to mine crypto currencies or execute lateral movement, impacting the isolation of the execution environment.
- (b) A threat actor can trick the execution environment into loading a CSV from a different user, and thus gaining access to sensitive data.
- (c) The adversary gains continuous access (persistence), waiting for the next interaction to send out information.
- (d) A threat actor can bypass authentication and authorization, and thus use the solution as an anonymous user.
- (e) A threat actor can trick the execution environment into writing malicious log data in the execution environment, which impacts the log data management software.
- Red: Not acceptable, must apply mitigations (which either reduce the probability or impact)
- Orange / Yellow: Can be accepted or mitigated
- Green: Are accepted by default
- (1) Ephemeral and isolated execution environment:
- Addresses: a, b, c, e
- Description: We want to have an ephemeral execution environment that is torn down after it fulfills its duty. Also, we need to ensure that the adversary is not able to stand up long lasting tasks (e.g. mining crypto currencies on our system). We therefore need to limit the execution time to what we would expect as the time it takes our model to generate the code and the answer. An ephemeral infrastructure allows us to "accept that the environment breaks during execution".
- (2) No outgoing network connections:
- Addresses: a, c
- Description: To contain the reach of the code, we prohibit any outgoing network requests except for the ones needed to download a CSV file.
- (3) Isolate language and libraries in the execution environment:
- Addresses: a, c
- Description: We design and build the execution environment in a "final" setup, i.e. prevent installing additional libraries. Also, we enforce usage of the programming language that we define.
- (4) Limit files that can be downloaded from the cloud storage:
- Addresses: b
- Description: We ensure that only legitimate files can be loaded by the execution environment. Since we know the user and their files in the "Logic" component, we can e.g. create a temporary download URL for a single file to the execution environment.
- (5) Use Guardrails:
- Addresses: a, b, c, d, e
- Description: An effective measure and first layer of defense, is setting up guardrails for incoming prompts as well as outgoing generated code.
- (6) Sanitize logging of the execution environment:
- Addresses: e
- Description: We sanitize all logs written by the execution environment.
It not only gives us the possibility to use an ephemeral infrastructure, it also provides workload isolation of execution environments through Firecracker.
Firecracker is virtual machine monitor (VMM) that uses the Linux Kernel-based Virtual Machine (KVM) to create and manage microVMs, which provide enhanced security and workload isolation over traditional VMs, while enabling the speed and resource efficiency of containers.
At the time of this writing in preview, we could also change the architecture entirely to use the "Code interpretation" feature of Agents for Amazon Bedrock.
For both cases, to prevent that a threat actor download additional libraries, outgoing network access for the Lambda is blocked (see 4.2).
In combination with a bucket policy, which only allows access to the bucket to the before-mentioned Gateway endpoint, we can additionally lock down our S3 bucket. You can find more information about this pattern in the AWS documentation.
- Special thanks to Massimiliano Angelino who created the first "talk to your CSV" implementation on Amazon Bedrock that inspired this blog, and to Luca Perrozzi for review and valuable feedback.
Any opinions in this post are those of the individual author and may not reflect the opinions of AWS.