We discovered that analyzing the internal representation of a large-scale language model (LLM) can quantify the model's inference progress. Based on this, we propose a two-step method for fine-tuning an existing inference model to explicitly generate a progress estimate (0-100%) during the inference process. The proposed model exhibits an average error of 10% from the actual progress for sequences less than 16K tokens.