It seems like this is what Rabbit's LAM was supposed to be. It is interesting to see it work, and I wonder how it will work in practice. I'm not sold on using voice for interacting with things still.
Image Generation is gross, I really didn't want this. I am not excited to start seeing how many horrible AI images I'm going to get sent.
I like Semantic Search in my photos.
This does seem like the typical Apple polish.I think this might be the first main stream application of Gen AI that I can see catching on.
I like that they finally brought typing interaction to Siri. You won't always need to use voice.
This does look like a real-world implementation of the concept promoted by Rabbit. Apple already had the App Intents API mechanism in place to give them the hooks into the apps. They have also publish articles about their Ferret UI LLM that can look at an app's UI and figure out how to interact with it, if there are no available intents. This is pretty exciting.
It has, but it’s presently an accessibility affordance you have to enable first.
It’s found under device Settings > Accessibility > Siri > “Type to Siri” On/Off
Oh, go figure. Let that be a lesson: if you don't check out the accessibility options, you're missing at least half the cool stuff your phone can do to actually make your life easier.
I wonder how they will extend this to business processes that are not in their training set.
At https://openadapt.ai we rely on users to demonstrate tasks, then have the model analyze these demonstrations in order to automate them. The goal is similar to Rabbit's "teach mode", except it's desktop only and open source.
I had similar reactions, a couple add-ons to make:
1. Yes, App Intents feel like the best version of a LAM we'll ever get. With each developer motivated to log their own actions for Siri to hook into, it seems like a solid experience.
2. Image Gen - yeah, they're pretty nasty, BUT their focus on a "emoji generator" is great. Whatever model they made is surprisingly good at that. It's really niche but really fun. The lifelessness of the generations doesn't matter so much here.
3. Polish - there's so much polish, I'm amazed. Across the board, they've given the "Intelligence" features a unique and impressive look.
It seems like this is what Rabbit's LAM was supposed to be. It is interesting to see it work, and I wonder how it will work in practice. I'm not sold on using voice for interacting with things still.
Image Generation is gross, I really didn't want this. I am not excited to start seeing how many horrible AI images I'm going to get sent.
I like Semantic Search in my photos.
This does seem like the typical Apple polish.I think this might be the first main stream application of Gen AI that I can see catching on.