Since AI models are trained on Reddit data, is it safe to assume someone has already been “shittymorph”ed by now?
Exploring the Impact of Reddit Data on AI Language Models: Have We Reached the ‘Shittymorph’ Stage?
As artificial intelligence models continue to evolve and improve, a significant portion of their training data derives from social media platforms, including Reddit. This has led to intriguing questions within the AI community: Have these models been exposed to the more chaotic, edgy, or explicit content commonly found in Reddit’s more unfiltered subcultures? Specifically, could someone have already “shittymorphed” an AI—an expression referring to a model mimicking a deliberately coarse or vulgar style—by now?
Motivated by this curiosity, I decided to test how well current AI systems recognize and respond to such niche linguistic phenomena. I prompted a leading AI model, Gemini, to generate a response in the “shittymorph” style—a colloquial and heavily informal manner often associated with certain Reddit subcultures—and was met with expected, yet amusing, results.
This experiment raises broader questions: How deeply have these models internalized the diverse and sometimes obscure content circulating within Reddit’s communities? Could probing further into hidden or lesser-known Reddit lore shed light on the extent of their knowledge and stylistic versatility? Exploring these boundaries not only enhances our understanding of AI capabilities but also highlights the importance of diverse training data.
What are your thoughts? Are there particular Reddit narratives or subcultures you’d recommend investigating to better understand how AI models absorb and replicate online vernacular?



Post Comment