Methodology
How We Score
Hundreds of sources. A proprietary weighting algorithm. A model that never stops learning. One score out of 100.
The Source Universe
The score starts with the reading list.
Watt'n Potatoes ingests reviews from hundreds of sources across the hi-fi press. The legacy pillars are all in there: What Hi-Fi, Stereophile, Hi-Fi News, The Absolute Sound, Hi-Fi Choice. The specialists too — Analog Planet for vinyl, Audio Science Review for measurements, SoundStage Network across its family of sites. If a publication has been covering this gear seriously for years, it's in the pool.
Beyond print and digital, there's the video layer. YouTube reviewers with real track records — not unboxing channels, but the accounts that have been doing careful, consistent work long enough to build a signature. And then the forums: Audiogon, Steve Hoffman Music Forums, Reddit's r/audiophile, the ASR community. Forum signal is noisier by nature, but at volume it reveals things no single reviewer catches.
Not every source carries equal weight. Credibility is weighted, track record matters, and a hot take from a new channel doesn't count the same as a considered review from a publication with thirty years of archives behind it.
A hot take from a new channel doesn't count the same as a considered review from a publication with thirty years of archives behind it.
The Algorithm
The model took months to build and is still being refined every week.
The hard part isn't collecting reviews — it's making them comparable. Five-star ratings, ten-point scales, percentage scores, and prose-only reviews all have to be normalised into a common representation before anything useful can happen. A four-star review in one publication is not the same thing as a four-star review in another, and the model knows it.
From there it gets more interesting. The algorithm accounts for reviewer bias, house sound, and gear pairings — the fact that a reviewer known for warm tastes will score a bright DAC differently than a reviewer who chases detail. It weights sources by credibility, discounts thinly sourced outliers, and penalises scores built on shaky evidence.
And it keeps learning. Every new review refines the model, every disagreement between sources is a data point, and every product that gets re-reviewed sharpens the picture. The algorithm today is not the algorithm six months from now.
The hard part isn't collecting reviews. It's making them comparable.
What The Score Actually Means
A WnP score is not a simple average.
It's a weighted consensus — a read on what the serious press, taken as a whole, actually thinks about a piece of gear. Quantity of reviews matters, but so does the credibility of each source, the agreement between them, and how much real evidence sits behind each rating.
There's also a confidence penalty for thin sourcing. A product reviewed by thirty credible outlets lands on a firmer number than one reviewed by two. A 92 from thirty sources and a 92 from two sources are not the same 92, and the score reflects that.
A 92 from thirty sources and a 92 from two sources are not the same 92.
The Four Tiers
Every score lands in one of four buckets, named with a little affection for the site's unofficial mascot. Deliciously Crispy (90–100) is the top shelf — gear the press broadly agrees is exceptional. Nicely Roasted (80–89) is the heart of the hobby: very good components with clear strengths and minor compromises. Just Okay (70–79) means the consensus is mixed — workable gear, but rarely anyone's first pick. Kinda Soggy (below 70) is what it sounds like, and the score exists mostly so you know to keep moving.
Deliciously Crispy
90–100Nicely Roasted
80–89Just Okay
70–79Kinda Soggy
Below 70The tiers are casual. The math behind them isn't.
Editorial Independence
The firewall between the commerce side and the editorial side is the single most important thing about this project.
WnP scores are editorially independent. The score is built by the algorithm from public sources. No brand, distributor, or advertiser has any input into or visibility into the scoring process before publication. HiFiHub, as a separate commercial platform, maintains brand and retail partnerships — but those relationships have zero access to the WnP editorial layer. The firewall is structural, not just a policy.
If that ever changes, the whole project stops working. So it won't.
The firewall is structural, not just a policy.
Why Aggregation Beats The Single Review
Every reviewer has a house sound. That's not a flaw — it's what makes them worth reading.
The reviewer who loves warmth will always hear warmth first. The one who chases resolution will always weigh detail highest. Taken individually, each review is a single well-informed opinion with a known tilt. Taken together, the tilts cancel and the signal sharpens.
Aggregation neutralises individual bias at scale. Where the sources agree, you can trust the consensus more. Where they disagree, that disagreement is itself useful information — usually a sign the gear is polarising, or pairing-dependent, or doing something unusual that some ears love and others don't.
The underlying thesis is simple. The more the press writes, the sharper the consensus gets. Our job is to read all of it and hand you the result.
Every reviewer has a house sound. Aggregation is how you hear past it.