RLM: Reasoning Language Modelโa model (like o1 or DeepSeek-R1) that generates intermediate 'reasoning tokens' (scratchpad) before the final answer to improve performance
Reasoning Tokens: Intermediate output tokens generated by an RLM to 'think' through a problem; these are often hidden from the user but incur computational cost
Indirect Prompt Injection: An attack where the adversary hides instructions in external data (e.g., a website) that the LLM retrieves, rather than in the user's direct prompt
Decoy Task: A benign but computationally intensive problem (e.g., Sudoku, Markov Decision Process) injected to force the model to spend resources solving it
ICL-Evolve: The authors' proposed optimization algorithm that uses In-Context Learning to evolve decoy tasks into more computationally expensive variants
Context-Aware Attack: Integrating the decoy task smoothly into the existing text of the retrieved source so it blends in
Context-Agnostic Attack: Injecting a general decoy template into a source without tailoring it to the specific surrounding text
Scratchpad: The internal buffer where an RLM writes its reasoning steps before generating the final response