Hacker News new | past | comments | ask | show | jobs | submit | bryanh's comments login

Confirmed that mini uses ~30x more tokens than base gpt-4o using same image/same prompt: { completionTokens: 46, promptTokens: 14207, totalTokens: 14253 } vs. { completionTokens: 82, promptTokens: 465, totalTokens: 547 }.


Huh. I am so confused.


Page 7 of their technical report [0] has a better apples to apples comparison. Why they choose to show apples to oranges on their landing page is odd to me.

[0] https://storage.googleapis.com/deepmind-media/gemini/gemini_...


I assume these landing pages are made for wall st analysts rather than people who understand LLM eval methods.


True, but even some of the apples to apples is favorable to Gemini Ultra 90.04% CoT@32 vs. GPT-4 87.29% CoT@32 (via API).


This isn't apples to apples - they're taking the optimal prompting technique for their own model, then using that technique for both models. They should be comparing it against the optimal prompting technique for GPT-4.


Showing dominance in AI is also targeted at their entreprise customers who spend millions on Google Cloud services.


We've been using Braintrust for evals at Zapier and it's been really great -- pumped to try out this proxy (which should be able to replace some custom code we've written internally for the same purpose!).


Thanks for all the support Bryan and team!



This was a ton of fun to build! We'll also be releasing a NLA enabled version of our Chrome extension [0] within the next couple of days which will be similar (but way more convenient than) the demo on the landing page above.

Anyone can start hacking on NLA today, go check out https://nla.zapier.com/get-started/ after you log in!

We're super bullish on LLMs for pulling "no-code" forward, helping more knowledge workers build automations. Already, folks are using our OpenAI [1] + ChatGPT [2] integrations to build very cool automations with summaries, categorization, copy writing, and more. We think there is a ton more to do here.

If anyone is interested in this problem space, shoot me an email bryan@zapier.com!

[0] https://chrome.google.com/webstore/detail/zapier/ngghlnfmdgn... [1] https://zapier.com/apps/openai/integrations [2] https://zapier.com/apps/chatgpt/integrations


Thanks for making this for us! Congrats on your big day


Joined the wait list 1 week ago - any news on when it will open up or how fast you are taking new users?



So it does.

The code there implies cl100k_base has a vocab size of 100k (I guess it's in the name lol) which means it is more comprehensive than GPT-2's 50k, so fewer tokens will be necessary.


Lesics on YouTube is a not too dissimilar modern incarnation of this style.


This isn’t exactly true. The Django ORM can be used with care in the async views found in FastAPI and Django (see sync_to_async and run_in_threadpool helpers).

Plans exist to make the Django Queryset async, so it’ll be exciting when that day comes!


Very true, and I have been following this pull request implementing Async Queryset with interest:

https://github.com/django/django/pull/14843

So far they are only going down the the queryset level, it then uses a thread pool for the db connector, next job would be to support async db connections.



Can someone educate me on why these 2 links resolve to localhost?


I think the first is the 32 bit int version of 127.0.0.1, and 0177 is the octal representation of 127.


Along those lines, you can also use hex https://0x7f.0.0.0x01:443/


Someone on Reddit recreated the audio effect in Python, pretty fun and well done!

Github: https://github.com/equalo-official/animalese-generator

Video Explainer: https://www.youtube.com/watch?v=RYnI_ZLj5ys


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: