This paper proposes the Optimistic Risk-Averse Actor-Critic (ORAC) algorithm to address the problem of conservative exploration in risk-averse constrained reinforcement learning (RaCRL), which leads to suboptimal policy convergence. ORAC constructs an exploration policy that maximizes the upper confidence interval of the state-action reward-value function and minimizes the lower confidence interval of the risk-averse state-action cost-value function. It encourages exploration of uncertain regions to discover high-reward states while satisfying safety constraints, and demonstrates improved reward-cost trade-offs compared to existing methods in continuous control tasks such as Safety-Gymnasium and CityLearn.