>I think HackerNews and other link aggregators (e.g. reddit) have a kind of recency problem, where there is a lot of great content, but people only see the recent stuff.
With Reddit and HackerNews, I want a relative ranking index. I can search by the top content, but something 5th today could have more votes than the top submission of 2015 because of forum/subreddit growth.
I want them ranked by something like the ratio of views to upvotes or upvotes compared to total upvotes for that day.
> I want them ranked by something like the ratio of views to upvotes or upvotes compared to total upvotes for that day.
Yeah, this is definitely possible with public data. I did something similar on reveddit [1] for removed reddit content. Hovering over the graph shows the item with the highest vote ratio [3], and clicking skips to that point in time. Code is here [2] for anyone interested and I apologize in advance..
Absolutely. HN user 'EvanMiller had a lot to say about this, 11 years ago. His tl;dr is that the ranking score should be "the Lower bound of Wilson score confidence interval for a Bernoulli parameter."
I believe HN's ranking system is extremely creative and works great for the main page and day-to-day use (my understanding is that it additionally makes use of time-decay terms for comments and stories). Like you said, it's really just historical search (i.e. algolia) that seems broken.
Agreed! The problem I see is that the score does not correlate well with the perceived quality of the community. I'm researching that for some time now and am in the process of preparing a blog article with data analysis and solution approaches.
With Reddit and HackerNews, I want a relative ranking index. I can search by the top content, but something 5th today could have more votes than the top submission of 2015 because of forum/subreddit growth.
I want them ranked by something like the ratio of views to upvotes or upvotes compared to total upvotes for that day.