This paper presents for the first time a large-scale brain-inspired language model (BriLLM), which is different from existing methods such as Transformer or GPT. BriLLM is a neural network based on the definition of signal fully connected flow (SiFu) on a directed graph, and unlike existing models that are limited to input and output, it provides interpretability for all nodes in the entire graph of the model. Tokens are defined as nodes in the graph, and signals flow between nodes according to the principle of “least resistance.” The next token is the target of the signal flow, and since the model size is independent of the input and prediction length, it theoretically supports infinitely long n-gram models. The signal flow provides re-activation and multi-modal support similar to the cognitive patterns of the human brain. Currently, the Chinese version of BriLLM (4000 tokens, 32-dimensional node width, 16-token-long sequence prediction) has been released, and it shows similar performance to GPT-1.