Hacker News new | past | comments | ask | show | jobs | submit | GraemeMeyer's comments login

I think Garnet might be what you're looking for: https://www.microsoft.com/en-us/research/project/garnet/


Doesn't support lua at the moment. Some connectors to Redis like .NET IDistributedCache use lua.


It turns out they didn't actually need LUA for IDistributedCache so it's being removed.


In my case it’s because companies would rather waste 6 months of person-hours than get out the credit card


What does this mean? Paying back loans / reducing debt? Or more like dividends?


The way that's written, "funnel fresh sources of funding to investors", appears to me to be alleging something that sounds like a pyramid scheme or a Ponzi scheme.


The filing (1) claims they are a Bing Search and Ads reseller that invests almost nothing into their actual search engine, and instead repays investors with every new funding round.

(1) From page 137: https://regmedia.co.uk/2024/02/26/us_v_google.pdf


Interesting that you see a slower response time with a large input - I don't see any speed degradation at all. Is that maybe just on the free tier of ChatGPT?


I'm on paid (rich, I know) and the performance is all over the place. Sometimes it'll spit out a whole paragraph almost instantly and other times it's like I'm back to my 2400bps modem.

I haven't noticed prompt size having an impact jut I'll test that.


This reflects my experience. Sometimes I'll provide a single sentence (to GPT-4 with the largest context window) and it will slowly type out 3 or so words every 5 seconds, and in other cases I'll give it a massive prompts and it returns data extremely fast. This is also true of smaller context window models. There seems to be no way to predict the performance.


Oh hey... leep an eye on your CPU load. The problem might be on the near end. In my case on a slower machine it slows down if you're dealing with a very long chat.

(DO report this as a bug if so)


I think that's not the issue here but I do notice the browser going crazy after a while of chatting with ChatGPT. The tab seems to consume a baseline CPU while doing nothing. I just brush it off and close it... bad JavaScript maybe. I should look into this and report as a bug, thanks for the advice.


This is basically how I respond to requests myself. Sometimes a single short sentence will cause me to slowly spit out a few words. Other times I can respond instantly to paragraphs of technical information with high accuracy and detailed explanations. There seems to be no way to predict my performance.


Early on, I noticed that if I ask ChatGPT an unique question that might not have been asked before, it'll split out a response slowly, but repeating the same question would result in a much quicker response.

Is it possible that you have a caching system too so that you are able to respond instantly with paragraphs of technical information to some types of requests that you have seen before?


Yes, search for LLM caching and semantic searches. They must be using something like that.


I cannot tell if this comment was made in just or in earnest.

As far as I understand, the earlier GPT generations required a fixed amount of compute per token inferred.

But given the tremendous load on their systems, I wouldn’t be surprised if OpenAI is playing games with running a smaller model when they predict they can get away with it. (Is there evidence for this?)


I'm guessing there are so many other impacts of own on the model that size of print probably gets lost. I can see a future where people are forecasting updates to ChatGPT like we do with the weather.


Yeah. It has so many moving parts that I doubt anyone can make a science out of it, but people will try for sure. Just like with most psycology/social experiments and SEO. I'm flooded with prompt engineering course spam these days.


I typically notice the character by character issue with complex prompts centered around programming or logic. It feels kind of like the model is thinking, but my guess is that the prompt is being dispatched to an expert model that is larger and slower.


If you mean the “analyzing” behavior, the indicator can be clicked on to show what it’s doing. It’s still going character-by-character, but writing code that it executes (or attempts to) to get the contents of a file, the solution for an equation, etc. Possibly an expert model but it seems like it’s just using an “expert prompt” or whatever you want to call it.


Interesting, no I'm on the pro tier aswell. So you're telling me you never get the character-by-character experience?

Edit: What prompt sizes are we talking about?

Even with small prompts I occasionally get rather slow responses but it becomes unbearable at 2000-3000 characters (the upper limit of custom instructions), at least for me it does.


In the UK you basically only see iPads (95%+). Android tablets aren’t common


I see a lot of Surfaces in commuter hours.


Kindle Fire excepted; loads of those in the hands of kids all over.


Two-column layout in Microsoft Word, large header, smaller footer, with appropriate font choices would get you basically all the way there.


You need to use Obsidian Sync if you have an iPhone. IIRC it’s because the third-party storage APIs are limited on iOS in such a way that makes it non-feasible for Obsidian. Something about bulk editing/access.


Obsidian works just fine with iCloud.


He does seem to be able catalyse a reaction in thousands of journalists just by farting, so it does seem to be a limitless supply


Fadell also spoke about this story on the Decoder podcast recently


Imgur links always do this for me on Safari: https://ibb.co/vcYy458


When on mobile I end up missing out on a lot of content because using imgur on mobile is much more painful than it should be.

This is compounded by the problem that mobile devices are often considered second class citizens by developers, so even finding out why imgur is behaving differently between mobile & desktop versions is a mission as there are usually no developer tools or ability to 'view source' or even see the error console.

This trend is being fully embraced by web developers who've taken the poison pill (e.g. Reddit, Imgur) where everything needs dumbing-down with options and normal functionality removed even when the desktop site could be adapted with a few small tweaks.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: