This paper identifies the limitations of pre-tokenization, the initial step in modern tokenization pipelines, and proposes BoundlessBPE, a novel BPE algorithm to overcome these limitations. Pre-tokenization generates tokens by segmenting text based on spaces and punctuation, which leads to a bias in token distribution toward common words. BoundlessBPE relaxes pre-token boundaries, merging semantically unconnected pre-tokens to create "superwords." This achieves a more even token distribution than standard BPE and enables up to 15% more effective text compression.