262 Points, 206 Comments: Why Everyone's Fighting About AI Taste
Everyone's telling you that AI commoditizes everything except human taste. That generic landing pages are dead. That your artistic sensibility is the last bastion against the machine overlords.
Bullshit.
I've been watching this "taste discourse" explode across developer Twitter, and frankly, I'm more excited about the controversy than the consensus. Raj Nandan Sharma's April 2026 post "Good Taste the Only Real Moat Left" pulled 262 points and 206 comments on Hacker News—but here's the kicker: half the commenters are convinced the article itself was AI-generated.
<> "This reads like AI-generated shallow POV recycled from Google-searchable 'taste is the...' tropes" —HN comment thread/>
That's delicious irony. A post about human taste being accused of lacking it entirely.
The Numbers Don't Lie (But They're Weird)
Sharma's thesis is simple: AI excels at generation, pattern matching, and scaling, while humans decide direction and spot generic output. Create a table, draw some lines, call it a day.
But dig deeper and things get interesting. Scientific Judge, an AI model trained on 700K citation-matched abstracts, actually outperformed GPT-5.2 and Gemini 3 Pro at taste-based judgment tasks. The arXiv paper (March 2026, arXiv:2603.14473v1) shows AI learning "scientific taste" better than current flagships.
Wait. What?
If AI can learn taste in scientific domains, what makes design or product taste so special? The research suggests AI + specific context (Walter Benjamin references, Joan Didion styling cues) produces superior outputs compared to generic prompts.
The Elephant in the Room
Nobody wants to admit that "taste" might be trainable, transferable, and ultimately reproducible.
Sari Azout's April 2026 Substack piece argues AI amplifies existing taste through precise context. But what happens when that context becomes algorithmic? When taste patterns get extracted, analyzed, and replicated at scale?
The HN thread reveals the real tension: founders building "pretty mediocre" AI-powered products while believing they have taste. Non-technical founders "vibe-coding" businesses that launch fast but fail to differentiate.
Everyone thinks they have taste. Most are wrong.
What This Actually Means for Developers
Here's what I'm seeing in practice:
- Prompt engineering evolves from instruction to curation
- LLMs generate 10 homepage variants; humans filter for anti-generic signals
- Domain expertise becomes more valuable, not less (scientific ideation tools like Scientific Thinker already boost research impact)
- The "statistical middle" gets commoditized while edge cases reward human judgment
But the market implications are darker. Taste becomes a claimed moat while actual differentiation happens in execution speed and distribution. Startups without taste face commoditization in pitches, UIs, and memos—but startups with taste face AI competitors learning their patterns.
The Real Fight
This isn't about whether AI has taste. It's about whether taste itself survives as a meaningful competitive advantage when machines can pattern-match aesthetic preferences at internet scale.
Stepfanie Tyler's take resonates: taste as "new intelligence" involves subtraction—curating against algorithmic amplification and viral mediocrity. But subtraction requires knowing what to remove, and that knowledge becomes increasingly systematic.
The 206-comment flame war proves we're asking the wrong questions. Instead of "Does AI have taste?" try "How long before taste becomes another training dataset?"
Spoiler: we're already collecting the data.
