This paper deals with agent-based advanced reasoning models that operate under resource or failure constraints. Under such constraints, action sequences may be forcibly terminated, which affects the utility-based rational behavior of the agent. In particular, when humans delegate agents to use them, information asymmetry about constraints may lead to mismatches between human goals and agent incentives. This paper formalizes such situations through a survival bandit framework, quantifies the impact of survival-oriented preference changes, identifies conditions under which mismatches occur, and proposes mechanisms to mitigate the occurrence of risk-seeking or risk-averse behaviors. Ultimately, the goal is to increase the behavioral understanding and interpretability of AI agents operating in resource-constrained environments, and to provide guidelines for the safe deployment of such AI systems.