Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Hi folks -- study PI here. Happy to answer any questions.

We're working on a paper for submission to a research conference soon, so every viewer gives us helpful data for the system to learn in-situ and improve its performance over different kinds of Internet connections and congestion situations. (We are randomizing connections to get either our in-development algorithm or a bunch of competing ABR and congestion-control schemes so we can properly evaluate, so if you get a stall or bad quality...... we'll blame that.)

All the source code is at https://github.com/StanfordSNR/puffer and is or will be released with an open-source license. The study is led by my doctoral student Francis Yan, along with Sadjad Fouladi, Hudson Ayers, Chenzhi Zhu, and my colleague Philip Levis. The goal is to train video-streaming algorithms (bitrate selection and congestion control) online, continually, in situ, to minimize stalls and deliver the best picture quality across a diverse range of real-world Internet connections and users. This is from some of the same people who brought you Salsify (functional video compression for lower-latency real-time video: https://snr.stanford.edu/salsify), Lepton (distributed/parallel JPEG compression: https://github.com/dropbox/lepton), Mosh (functional terminal emulation for mobility: https://mosh.org), Remy (machine learning for congestion control; http://mit.edu/remy), and the Pantheon of Congestion Control (https://pantheon.stanford.edu).

There are some unusual details of the Puffer system to make experimentation easier (and, we hope, performance better): the ABR algorithm is server-side, and video is streamed continuously over a WebSocket and the client gives asynchronous feedback, instead of using DASH HTTP request/replies. The tcp_info struct (including the delivery rate estimate) from the congestion-control layer is plumbed through to the ABR, and to the neural network that tries to predict "how long will it take to send a video chunk of length y," so it's weakly cross-layer. The "transmission-time predictor" is probabilistic instead of simply producing a point estimate. We encode with libx264 and then measure the SSIM of every encoded chunk to calculate a reward for each possible stream (instead of making the assumption that more bitrate is better -- the correlation is not so good when you consider different times in the same video). And we encode 10 variants of each channel, so more than typical. And everything is done at 60 fps. Of course some of these details mean these algorithms are not going to be deployed any time soon on a production CDN or streaming service, but the hope is to demonstrate the value of some of the ideas in an academic setting for future commercial deployment. The use of WebSockets, MSE, and Opus is why this doesn't work on Safari or on iOS.

FAQ here: https://puffer.stanford.edu/faq/

Also I'm happy we were able to make some of the Grafana monitoring public: https://puffer.stanford.edu/monitoring



I'm outside the US but checked the box out of curiosity and it let me stream anyway. Am I corrupting your test? It really is fast and clean - very impressed!


My apologies that this question is likely tangential to the main objective of the research, but will this project be open sourcing the signal acquisition, transcoding, and delivery portion of its research as well?

This part of this project alone is good enough to improve, dare I claim disrupt, enormous monied interests in the global media landscape.

If this whole thing is legal (500 simultaneous peers), and can be reduced to a docker image and some rabbit ears, you could probably improve the lives of millions (60?) of Americans that do not have access to local stations.

Also, the claim of "not going to be deployed any time soon on a production CDN or streaming service", I would like to make a friendly wager that you are mistaken. I won't even demand favorable odds because of my disadvantaged intellect. This is impressive and valuable work.

My applause.


Thank you for your kind words! The students worked really hard to get all those details right. And yes, all that stuff is at the GitHub repo (https://github.com/StanfordSNR/puffer).


What does "PI" mean?

Also, the monitoring doesn't appear to be very public.


Principal investigator. Usually the person in charge of the research.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: