I bought two units of a product on two occasions - not too far apart in time, from am.uk.
One unit was rubbish, or at least not suitable for the application, so I wrote a one star review and returned the item.
The other unit seemed OK, so I wrote a more positive review, and kept the item. As this was for a different order I thought I might just be allowed two different reviews.
Amazon has now removed the first review - but that wasn't really my intention at all. I could modify the now misleading positive review by downgrading it to a mediocre rating of two or three stars, but that would also be a misrepresentation. The application circumstances were slightly different, so one unit deserved a very low rating, while the other unit, which actually worked for the intended purpose, deserved a significantly higher one. Not everyone reads all the different star reviews, or even a sample of the reviews at different star ratings. Probably very few people read the average (3 star) reviews. The point is that a rating may be only a very crude measure of satisfaction.
For CDs, which I suppose some round here might like, it is not uncommon for a CD to be rated highly for performance, but (particularly if it is a "historic" recording) much lower for sound quality. That's OK, and much more helpful than simply averaging the scores and "giving it foiv".
Another example of where averaging doesn't work well was an example of an investigation into which height of seat lorry drivers preferred. It turned out that there were two popular seat heights - so designers then produced lorries with seats which were at the mean of the two heights. Those seats turned out to not be popular with very many drivers. Using arithmetic means for bimodal (or even multi-modal) distributions just does not make sense, yet customer rating tools must often force people to make such "judgements" when evaluating products if they are just based on one scale.
One unit was rubbish, or at least not suitable for the application, so I wrote a one star review and returned the item.
The other unit seemed OK, so I wrote a more positive review, and kept the item. As this was for a different order I thought I might just be allowed two different reviews.
Amazon has now removed the first review - but that wasn't really my intention at all. I could modify the now misleading positive review by downgrading it to a mediocre rating of two or three stars, but that would also be a misrepresentation. The application circumstances were slightly different, so one unit deserved a very low rating, while the other unit, which actually worked for the intended purpose, deserved a significantly higher one. Not everyone reads all the different star reviews, or even a sample of the reviews at different star ratings. Probably very few people read the average (3 star) reviews. The point is that a rating may be only a very crude measure of satisfaction.
For CDs, which I suppose some round here might like, it is not uncommon for a CD to be rated highly for performance, but (particularly if it is a "historic" recording) much lower for sound quality. That's OK, and much more helpful than simply averaging the scores and "giving it foiv".
Another example of where averaging doesn't work well was an example of an investigation into which height of seat lorry drivers preferred. It turned out that there were two popular seat heights - so designers then produced lorries with seats which were at the mean of the two heights. Those seats turned out to not be popular with very many drivers. Using arithmetic means for bimodal (or even multi-modal) distributions just does not make sense, yet customer rating tools must often force people to make such "judgements" when evaluating products if they are just based on one scale.
Comment