Hacker News new | past | comments | ask | show | jobs | submit login
Microsoft’s Azure Kinect AI Camera (techcrunch.com)
163 points by panabee 7 days ago | hide | past | web | favorite | 44 comments





I have 5 of these sitting in my office.

The quality of image (depth and rgb ) is staggering compared to everything else I've used (zed, realsense, kinect v2, and a couple of others).

They can be chained together for near microsecond timing accuracy.

The mic array is awesome. The IMU is awesome.

But right now I've set them all back in their boxes.

Why? Because without being a vision specialist, there's nothing I can do with these devices.

The SDK and sample code is so incredibly bare bones, it is almost laughable.

There's no way to make use of those mics for anything. Its literally not in the SDK.

There's no way to make use of multiple devices in any practical manner. No point cloud merging, no calibration or shared space alignment.

Then there's the problem that buries deep in the SDK is a binary blob that is the depth engine. No source, no docs, just a black box.

Also, these cameras require a BIG gpu. Nothing is seemingly happening onboard. And you're at best limited to 2 kinects per usb3 controller.

All that said, I'm still a very happy early adopter and will continue checking in every month or two to see if they've filled in enough critical gaps for me to build on top of.

If any devs in Seattle want to collaborate (or know computer vision well enough to fill in some of these gaps for the OSS community) let me know :)


> The SDK and sample code is so incredibly bare bones, it is almost laughable.

This is why I held off ordering a few. I took an hour looking through the docs and concluded the overall offering isn't fully baked yet. The hardware looks incredible, but the software looks anemic.


Why did you get 5?

One of the key features for me is the hardware chaining to make realtime PC merging easier to resolve. All of the scenarios I can about are realtime rather than post processing reconstruction.

i would assume some full body mo-cap or realtime avatar scanning.

I can line you up with many people working with them around Seattle if you're interested. Got a contact?

Ditto here, though some of these people might end up distracted by their full time jobs. ;)

I have a similar problem :) - azurekinect@gmail.com

Sure and thanks :) - azurekinect@gmail.com

I have one of these as well and exploring interesting use cases for the tech. If you don’t mind me asking, what do you have in mind for those?

Can you still use the body tracking from previous kinects?

>The system requirements are Windows® 10 PC or Ubuntu 18.04 LTS...

I must say: I'm liking this new Microsoft.

[0] https://azure.microsoft.com/en-us/services/kinect-dk/


> 7th Generation Intel® Core™ i3 Processor (Dual Core 2.4 GHz with HD620 GPU or faster), USB 3.0 port, and 4 GB RAM.

Raspberry Pi or a Jetson Nano are probably not gonna work... Seems to be x86 only.


Given Microsoft's stance on supporting "non-native" setups I think we should be expecting ARM Linux driver's if this device catches on.

Me on the other hand i remain very skeptical with the new Microsoft.

https://news.ycombinator.com/item?id=20412905


It's just a matter of following the money. A bigger and bigger slice of Microsoft's income is coming from Azure, and an increasing proportion of Azure users are running Linux. This gives them strong incentive to be the business leaders in open source software.

The other increasing slice is services, where they get you to buy stuff from their storefront, or subscribe to their game pass, or office 365, or make personal interactions with Bing and Windows (so they can sell targeted ads). This naturally gives Microsoft the incentive to know everything about you.


Following the money is exactly why I am skeptical. As Linux becomes a revenue stream for Microsoft it will also become a Microsoft product. From the company that was built on Embrace, Extend, Extinguish I think that is cause for concern.

This is really cool! Back in college I used the Kinect dev kit to build proof-of-concept special effects for live theatre as an independent study project. I used the Kinect pointed at the face of an actor off stage, ran the resulting 3D data points to some cool algorithms, and then projected the result onto a screen on stage. The idea was that, for example, in Hamlet instead of having an actor in make up play the ghost of King Hamlet, you could have this larger-than-life projection on stage.

This camera is way better quality, so it'l be neat to see the sort of projects can be done now.


https://azure.microsoft.com/en-us/services/kinect-dk/

Here is the actual Microsoft link just in case you don't want blog spam that is nearly unreadable on a phone.


How does it compare to iPhone sensors used for Face Id? I’m wondering if mounted to a workstation it could be used to implement a Face id system under linux.

yes. the hard part is implementing FaceID! Lots of papers on the topic: https://scholar.google.com/scholar?hl=en&as_sdt=0%2C5&as_ylo...

I was just checking specs on the two devices and it seems that the new Kinect depth even higher resolution, so it looks like a go in terms of raw input data.

So my next question would be why would it be significantly harder than regular facial recognition approaches as found in say OpenCV, naively one would think more data makes it easier not harder, neglecting hardware requirements/performance, but just from accuracy perspective from a trivial refactor of current facial identification algorithms.

I’m not talking about identification of people moving or far away but straight looking at from a fairly close distance.


Right now the SDK supports body pose estimation without finger joints. Does anybody know or work on a model including finger joints?

Does anyone know if this will work with Mac running Windows in pArallels? Really wanna tinker with it but stuck with a Mac :(

I did some Kinect V2 experiments on Bootcamp and it worked like a charm

It supports Ubuntu too, you can dual boot.

This youtube video is definitely a chapter inspired/stolen from Apple's ads. Good on Microsoft though.

https://www.youtube.com/watch?v=jJglCYFiodI


I remember seeing The V Motion Project [0] when it first came out. It'd be good to see other people approach this too!

[0] https://www.youtube.com/watch?v=YERtJ-5wlhM


The "Order" link goes to a 404 outside of the United States/for languages other an en-us

From what I understand, this is nice if you are trying to use multiple devices together. The Kinect's API makes that very easy with Azure (read: multiple cameras scanning a single location in real time, via a robot).

If you just need the same sensors for depth, but significantly cheaper, then look at occipital.

Link: https://occipital.com


Occipital uses structured light. Kinect uses time of flight. Completely different sensor technology.

I’m interested in buying one and doing a hack project. Anyone interested in brainstorming with me?

what about something like this (holoportation): https://www.youtube.com/watch?v=7d59O6cfaM0

Can they be used as a kinect replacement for xbox360/xenia?

How does this compare to Intel's RealSense?

What are some potential use cases?

Thinking about using vision technology for small scale aquaculture stock management. Sorting the fishes by weight while eliminating the scale. Probably a conventional camera might suffice with a little bit of ML training. Length does not correlate well with weight in the fish species we are using, but a 2D picture of the fish body might do it. A 3D picture with the new kinect will do the job for sure.

Classic Kinect sensors are used extensively for gesture detection in the niche world of digital interactive exhibitions. You see them a lot in things like science museums or popup events or digital art gallery kinda places.

This exhibit had at least a dozen Kinect cameras: https://nysci.org/home/exhibits/connected-worlds/


Motion capture for animation comes to mind. Another is to project peoples' motion into VR space.

I've been wanting to build gesture based control of my smarthome for a while.

This would definitely be overkill. You could do that with 5$ cameras and some ML.

[flagged]


>It’s meant to give developers a platform to experiment with AI tools and plug into Azure’s ecosystem of machine learning services (though using Azure is not mandatory).

Agreed. Doubt sdk will be OS/Linux friendly either. Previous Linux kinect drivers (libfreenect) were all community driven, it worked, but not as polished as the official stuff.

Ubuntu is officially supported by Microsoft.

>The system requirements are Windows® 10 PC or Ubuntu 18.04 LTS...

[0] https://azure.microsoft.com/en-us/services/kinect-dk/




Registration is open for Startup School 2019. Classes start July 22nd.

Guidelines | FAQ | Support | API | Security | Lists | Bookmarklet | Legal | Apply to YC | Contact

Search: