I created this for the HN community and I'm very happy to see so many people enjoying it :)
If you have any questions about how I built this or if you'd like to suggest something new, let me know!
The front-end specifically asks for pages in 200-chunks. So whether you slide to page 299,300 or 385, it will request 200-400 from the API. This means I can then very easily serve these requests out of Redis  and during usage spikes, requests never hit the disk :)
i7 with 16 gig RAM
I would love to see a list of all the Hacker News "meta tools".
I posted my own last week: http://www.HnEasy.com sorts hn all posts and comments by upvotes from the last day... to last 5 years.
Perhaps one minor bonus to HackerSlide for now is anyone can take the data collected. The URLs for the JSON archives are formatted like: http://hackerslide.com/data/2012-11-01-23.json (YYYY-MM-DD-HH) although even better long term are the once a day versions, e.g.: http://hackerslide.com/data/2012-11-01.json
I've learnt two things from this project in particular. First, that most similar projects don't seem to stick around very long (the Reddit one it was based on disappeared after a few months as have several others - http://hckrnews.com/ is an exception I can recall). Second, these tools seem to be popular at first but then rarely used over time. Luckily I still find it useful to catch up after vacations, etc ;-)
Indeed, may I make a suggestion here:
1) Why not make a datadump so people wouldn't need to scrape ~800*24 json files individually?
2) OP ought to load this data into his version so the timeline goes back further
3) It seems quite a few people get the urge to tinker like this with HN, I'm sure pg doesn't mind the scraping, but it strikes me as vastly more efficient if some sort of shared resource was setup and perhaps added to the footer, in the vain of HNsearch, so people don't waste time get crawling data setup.
I'm sure somebody else has a dataset just like yours that goes back further still. :)
Also, thank you for making this and OP for making his. Fun.
Publicly available JSON files was just a side effect of the implementation. But it's easily to tar and gzip it up, so there's now such a file at http://secretshenanigans.s3.amazonaws.com/hnfrontpages.tar.g... (32MB).
>> most similar projects don't seem to stick around very long ...
>> ... http://hckrnews.com/ is an exception I can recall)
I would be sad if it disappeared.
This is where the scraping happens. The code is a little uglier than I'd like, but that's largely to do with the hard-extract-data from HN markup. I'll look at adding more data soon.
All in all, Awesome.
Does the keymaster live demo at http://madrobby.github.com/keymaster/ work for you?
I originally just dumped the raw scraped HTML into the DOM but that proved to be too space-inefficient (even with gzip).