With the increasing presence of adolescents and children online, it is crucial to evaluate algorithms designed to protect them from physical and mental harm. This study measures the bias introduced by emerging slurs found in youth language on existing BERT-based hate speech detection models. The research establishes a novel framework to identify language bias within trained networks, introducing a technique to detect emerging hate phrases and evaluate the unintended bias associated with them. As a result, three bias test sets are constructed: one for emerging hate speech terms, another for established hate terms, and one to test for overfitting. Based on these test sets, three scientific and one commercial hate speech detection models are assessed and compared. For comprehensive evaluation, the research introduces a novel Youth Language Bias Score. Finally, the study applies fine-tuning as a mitigation strategy for youth language bias, rigorously testing and evaluating the newly trained classifier. To summarize, the research introduces a novel framework for bias detection, highlights the influence of adolescent language on classifier performance in hate speech classification, and presents the first-ever hate speech classifier specifically trained for online youth language. This study focuses only on slurs in hateful speech, offering a foundational perspective for the field.