Seems like Gemini 2.0 Flash Thinking got silently updated in AI Studio to accept audio input, as well as image and text, making it the first reasoning model I'm aware of that works across audio. Trying it out on a few audio tasks (transcription, sound analysis) seems to perform a little better than 2.0 Flash or even Pro. Curious what you guys make of it!
Thank you for saying that! I have a much, much harder time documenting everything and writing out each decision in continuous text than actually writing the code. So it took a look time for me to write all of this down - so I'm happy you appreciate it! =)