OpenAI engineer claims that Codex with /detectaibugs command outperforms Claude
Mood
skeptical
Sentiment
negative
Category
tech
Key topics
AI
Codex
Claude
Code Comparison
An OpenAI engineer claims Codex outperforms Claude with a specific command, but the community is skeptical of the comparison's fairness.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
N/A
Peak period
1
Hour 1
Avg / period
1
Based on 1 loaded comments
Key moments
- 01Story posted
11/18/2025, 6:26:45 PM
15h ago
Step 01 - 02First comment
11/18/2025, 6:26:45 PM
0s after posting
Step 02 - 03Peak activity
1 comments in Hour 1
Hottest window of the conversation
Step 03 - 04Latest activity
11/18/2025, 6:26:45 PM
15h ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Of course if you give codex extra help by using a command/plugin then it’s going to outperform a claude agent that has no extra help. Funny how when asked if he tried to use the same command/plugin with claude code, he stated that he didn’t have time to. The OpenAI engineer clearly just wants to slander claude. OpenAI people are always doing this kind of crap. I’m going to try to see if I can do a local performance comparison when I have the chance this week. I’ll intentionally try to give a vague feature-request prompt to both claude and codex, no extra plugins/commands allowed, and see which one produces more bugs. I’ll edit this post on the weekend. Unfortunately, I don’t have a Mac anymore, so I can’t even test the performance of claude with the plugin vs codex with the command, so that don’t expect that from me (sorry).
Plus, has anyone tried codex 5.1? I want to be respectful and fair here… but respectfully…it’s complete garbage compared to Opus.
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.