Posted to r/ClaudeAI, the original poster (OP) describes an experiment where they tested several AI models, including different versions of Claude and ChatGPT 5.3, with the same code critique prompt. The OP was surprised to find that the initial paragraphs of the responses from ChatGPT 5.3 and Claude Opus 4.7 were almost identical, leading to speculation about potential cross-model training or one provider secretly using another's API. While the responses diverged later, the OP found the initial similarity striking. Community comments suggest this phenomenon of AI models sounding similar is a known issue, possibly due to models distilling each other.
Want claims fact-checked?
Sign up free to run a Deep Sift on this post — verifies every claim with web-grounded research.
Sign Up FreeAI-generated assessment. Verdicts on this page were produced by language models with web search and may contain errors, hallucinations, or out-of-date information. They reflect Bullsift's automated analysis, not editorial judgment. Read the linked sources before relying on any verdict. How this works ·
Claims & verdicts
The original poster tested various AI models, including Sonnet 4.6, Opus 4.6, Opus 4.7, and ChatGPT 5.3, using the same prompt.
unverifiedThe original poster's primary objective was to determine which AI models would accurately identify the most significant issue within a provided example code.
unverifiedChatGPT 5.3 and Claude Opus 4.7 produced initial paragraphs that were nearly identical in wording.
unverifiedThe original poster considered the possibility that one AI model was extensively trained using data from another.
unverifiedThe specific wording similarities observed were too precise for the original poster to attribute them solely to training data overlap.
unverifiedThe responses from ChatGPT 5.3 and Claude Opus 4.7 showed greater divergence after their initial paragraphs.
—