PRISM Eval presented its LLM robustness leaderboard and submitted a technical report for the Paris AI Action Summit. This report introduces the PRISM Eval Behavior Elicitation Tool (BET), an AI system that performs automated adversarial testing via dynamic adversarial optimization. BET achieved a 100% attack success rate (ASR) on 37 of 41 state-of-the-art LLMs. Beyond simple pass/fail evaluations, we proposed a fine-grained robustness metric that estimates the average number of attempts required to induce harmful behavior, demonstrating a more than 300-fold difference in attack difficulty between models. We also introduced baseline vulnerability analysis to identify the most effective jailbreaking techniques for specific risk categories. This collaborative evaluation with trusted third parties from the AI Safety Network provides a practical path toward distributed robustness evaluation across the community.