Hacker Newsnew | past | comments | ask | show | jobs | submit | darkvertex's commentslogin

Or import MIDI into a music box sheet converter? https://musicboxmaniacs.com/create


Yeah! They sell many packs by genre (inventions, music, movies, science, etc) but what's neat is you can mix the cards of multiple genres and the game still works all the same. Very elegant concept.


WebRTC excels at sub-second latency peer to peer, so you can do near-realtime video, so anywhere that is useful.

Say you wanted to do a virtual portal installation connecting views from two different cities with live audio, you could have ffmpeg feed off a professional cinema or DSLR camera device with a clean audio feed and stream that over WebRTC into a webpage-based live viewer.

Or say you wanna do a webpage that remote controls a drone or rover robot, it would be great for that.


The irony is that you don't _actually_ need WebRTC to get subsecond latency; you can fairly reliably get ~100–200ms (plus network latency) with a completely normal TCP stream.[1] But since browsers have effectively standardized on HLS, whose design is completely antithetical to low-latency (you _can_ do low-latency HLS, but only with heroic efforts), low-latency streaming video has never really been part of their bread and butter. So instead, we abuse a _much_ more complicated protocol (WebRTC), because that happens to hit a path that was meant for low-latency videoconferencing.

(I did sub-100ms glass-to-glass streaming with VLC back in the day, so it is eminently possible. But the browser is in your way.)

[1] Much less than that is going to be tricky under non-perfect network conditions, because once you start having any sort of packet drop, you want to go from TCP's retransmission regime and instead start dropping packets, take the artifacts for a little while, and then go on.


  node1# nc -u node2 12345 < /dev/fb0
  node2# nc -lu 12345 > /dev/fb0
The "sub-second latency" thing is the standardized punchline coming from WebRTC folks, but yes, it's confusing. Nothing can make video flow faster than above, only thing you can do by inventing a new standard is to minimize the overhead you must add for your purposes.


I saw this also. WebRTC just is the path of least resistance/highest adoption at this point.

We could go make a better/simpler standard for video streaming that is TCP. What a giant battle though that would never seen adoption .

I have accepted/embraced the challenge of making WebRTC as accessible as possible. Stuff like WebRTC for the Curious in hopes of making it less painful for everyone dealing with the complexity :)


> We could go make a better/simpler standard for video streaming that is TCP. What a giant battle though that would never seen adoption .

What do you mean? <video> in HTTP against a stream works, you don't need a new standard. But it's not a low-latency path (you cannot control the buffer).


Explanation: HTTP Live Streaming slices the bitstream into “segments” (traditionally 6–10 s each) and only starts playing after it has downloaded several of them. Out of the box that means 30–60 s of startup and live-edge latency—fine for linear TV, terrible for anything interactive.

Apple’s LL-HLS spec shrinks those segments into “partial segments” and uses CMAF to let the player start decoding while a chunk is still arriving. With careful encoder tuning, HTTP/2/3 push, CDN support, and a compatible player you can reach 2–5 s, sometimes a bit lower—but every link in the chain has to cooperate, so implementations are still called “heroic” for a reason.

Safari plays HLS natively and on other browsers you can bolt on hls.js via Media Source Extensions. DASH, RTMP, SRT, etc. all need extra code or plugins, so HLS became the “safe default” for on-demand and broadcast-style streaming even though it isn’t low-latency friendly.


And in real world (e.g. mobile networks) there is going to be packet loss, so TCP is a non-starter for production-quality real-time streaming.


You should try the new V2 of Meat's incredible DepthAnything model: https://depth-anything-v2.github.io/

It blows MiDaS out of the water imho.


*Meta, not Meat. Damn autocorrect!


Oh excellent! This looks really good, thank you.


If you wanna be ultra cheap and your subject is a still scene you can literally just use any camera like your smartphone and take two pics side by side in short succession, distanced apart by your human IPD (Inter-Pupillary Distance.)

Did you know your index finger is roughly 1 IPD from nail to knuckle? This is useful:

Hold your phone sideways, look where your lens is in the horizontal axis and align your left index finger's knuckle such that you're holding the phone edge with it but your knuckle is lined up just above the lens... Snap a pic. (This will be your left eye pic.) — Now keep your left index straight as-is and with your right hand gently grip your phone and slide it horizontally such that the lens previously lined up above your knuckle is now aligned just below the tip of your left index finger nail. Now snap another pic. (This is your right eye pic.)

Now put them in a collage side by side in Photoshop or whatever, and make it small and cross your eyes. If you kept your alignment fairly you should now have a respectable 3D photo.

Voila! 3D photos on the cheap, as long as your subjects stand still between the two pics.


You don't need a family calendar. You can make extra calendars and in the settings on web desktop you can share just the one calendar with another friend's Google account, and if you pick the good permissions either of you can make events in it, and it syncs to all who have it, magically: https://support.google.com/calendar/answer/37082?sjid=171696...


Oh you are right, thank you ! I don't know why I didn't found this out earlier... Still, it's a shame that this feature is only usable on the desktop web app that most users will never use


What do you use for validating pronunciation?

Or do you just use a Speech To Text model and hope the text that comes of it is the expected phrase?



Huh.. someone at Google is just blindly copying whatever Microsoft is enshitting with AI. Nevermind, it's their CEO :)


Haha, I loved that album when it came out!

Indeed it was unsearchable though. I had to buy the CD.


It's nice but I prefer https://github.com/extrawurst/gitui cause it's a little snappier.


Of course it’s written in rust


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: