Skip to the content.

Link for the Room

📚 Study Notes

LLM Output Risks

Improper Output Handling (LLM05)

 


âť“ What vulnerability refers to situations where a system blindly trusts whatever the LLM generates and uses it without verification, filtering, or sanitisation?Improper output handling

Sensitive Information Disclosure (LLM02)

LLM Sensitive Info Risks

How leaks happen:

Common misconceptions:

 

[!CAUTION] Leaked model info (API keys, URLs, prompts) lets attackers hack or steal data without touching the system.

 

Attack Cases

This section basically explains how Large Language Models (LLMs) can become security vulnerabilities when their output is trusted and used unsafely. The core idea is that attackers don’t need to inject malicious code directly — they can trick the model into generating it for them, and the system executes or renders it without realizing the danger.

Main attack scenarios explained

1. Model-generated HTML/JavaScript (XSS via LLMs)

Impact:

 

[!CAUTION] The attack vector isn’t the input field, it’s the model’s output.

 

2. Model-generated commands or queries (automation abuse)

Examples:

Why this is dangerous:

[!CAUTION] LLM outputs can be risky. They might run bad code or reveal secrets if used without checking. Always treat what the model gives you as untrusted.

 


âť“ What is the content of flag.txt? THM{***_*******_*********_**_****}