I think modern overlay networks can navigate CG-NAT fine now. Other options include free cloudflare, or just a wireguard tunnel to a free tier VPS. On a similar point, I don't think enough people talk about how most western home internet connections now also have similar bandwidth as entire datacentres had in the 2000's too.
We still take for granted how hard basic web technology is for people who don't consider themselves technology people though.
I'm all for personal website and these sentiments which regularly come up here around self hosting. This one seems a bit disproportionately confused and angry though.
If we're going to have any large aggregation or social media businesses where individuals trade data ownership for convenience, being able to put your opening hours and rates on the the internet without having to figure out how to have a website seems like the optimal use case.
I think we should aim for a sensible mid ground where social media provides just the things it provided before around 2011, like updates and communication with people you know and want to interact with already.
An "all personal websites" web that OP is calling for is just pushing the exclusion they feel onto the people they're complaining about.
We should have websites. We should also use the appropriate tool for the appropriate job, and running your own website isn't the best tool if you just want to get your business rates and opening hours on the web.
> Really, LLMs are kind of like convenient, wildly inefficient proxies for useful processes. But I'm not convinced they should often end up as permanent fixtures of logical pipelines. Unless you're making a chat bot, I guess.
I think I agree with this. It's made me realise LLMs are great for prototyping processes in the same way that 3D printers are great at prototyping physical things. They make it quick and easy to get something close enough to see the unforeseen problems a proper solution might have.
3d printing is a great analog because there are so many critical considerations that are often missed or can't be accounted for in the prototype, but, it's alright because it's a prototype. The strain testing, durability, manufacturing at scale; none of that is properly addressed. Those might involved some serious, expensive challenges, too. But it's alright because you've got something in your hand that informs you whether or not those challenges are worth contending with. I really love this about LLMs and 3d printing.
This feels like a ridiculous thread that captures everything wrong with modern Javascript ecosystem.
It's grown into a product of cults and attempted zingers rather than pragmatic or sensible technical discussions about what we should and shouldn't expect to be able to do with an individual programming language.
edit: to clarify, I assume there needs to be a basical level of comprehension of programming languages to debate the nuance of one, and if you can't think of a single reason as to why someone would want types removed, that's a possible indicator you don't have that necessary level yet, and I think the most effective way for you to learn that is to Google it. Sorry for coming across as rude if you genuinely don't know this stuff.
If you already know many reasons as to why types would be removed, then it seems disingenuous to ask that question, other than to make the point that you feel types shouldn't be stripped. If you think that, say it, and explain why you think they shouldn't be stripped.
The current state of Javascript is you _have_ to remove types; I was pointing out I can think of reasons why I sometimes wouldn't want to. (Admittedly in a glib manor; though on this site many prefer that to four paragraphs)
I really like solutions in this space, and this is quite nice. Seeing people try create solutions like this really tickles my brain a lot. Even if I think more into it and conclude it has catastrophic issues, I still really get a weird kick learning about novel decentralised networks. I really can't explain it. Fancy combinations of encryption and decentralisation just really do it for me, to an abnormal and uncomfortable extent. Hopefully someone else relates to this.
Anyway, I really like this idea, it's cool. When I think about this one though, I feel there's too much friction in the follow/unfollow process. Having unfollowing requiring reenecrypting and rebuilding the entire website for everyone seems cumbersome. It's not a killer in itself, but combined with this:
> If the original post is inaccessible (e.g. the viewer doesn’t follow the author), the reply is hidden entirely. A user only sees replies from people they follow — this is the spam prevention mechanism.
I think this is going to prevent it from scaling in any desirable way. I know it's not intended to scale, and is targetted at smaller freinds networks, not influencers, but again, even small friendship networks grow complex, and I can see the experience on S@t turning into the worst parts of activitypub where you can only read half of the interesting replies because not being friends, and it being a pain to then become mutual friends.
But, I really, really do like that s@t feels like a combination of RSS, activity pub and static sites, having a browser heavy client is interesting to.
It does feel a bit like s@t wants stuff to be easily locked down between a dynamic list of friends though, and it feels a bit weird to have the foundational tech of such a protocol be static sites, which by definition make it hard to lock stuff down to a dynamic list of friends. Hmmmm, I really do love/hate static site architecture
Fedify is really fun to mess around with. The fedify tutorial was also really great for learning about developing with Activity pub and the fediverse in general.
I don't use Discord generally, but the fedify Discord is particularly useful, and I see how some discussions there have evolved into features in this release which is nice too!
I think the appeal and use case for Graphene and similar OS for most users is the Google/privacy/ownership type argument.
I do understand your point that people at risk of state level attacks might get a false surface level appearance of defence from this. But then anyone who's a target of state level attacks and is making OS decisions based on a surface level understanding of the tech is not going to have a good time anyway.
Personally I run an ollama server. Models load pretty quickly.
There's a distinction between tokens per second and time to first token.
Delays come for me when I have to load a new model, or if I'm swapping in a particularly large context.
Most of the time, since the model is already loaded, and I'm starting with a small context that builds over time, tokens per second is the biggest impactor.
It's worth noting I don't do much fancy stuff, a tiny bit of agent stuff, I mainly use qwen-coder 30a3b or qwen2.5 code instruct/base 7b.
I'm finding more complex agent stuff where multiple agents are used can really slow things down if they're swapping large contexts. ik_llama has prompt caching which help speed this up when swapping between agent contexts up until a point.
tldr: loading weights each time isn't much of a problem, unless you're having to switch between models and contexts a lot, which modern agent stuff is starting to.
I always felt the idea of trying to align your code, policy, software and infrastructure so it's easy to do compliance is the bread and butter of devops and devsecops in a regulated environment,
Is this an article by someone who's just done ISO 27001 for the first time and realised that?
I think modern overlay networks can navigate CG-NAT fine now. Other options include free cloudflare, or just a wireguard tunnel to a free tier VPS. On a similar point, I don't think enough people talk about how most western home internet connections now also have similar bandwidth as entire datacentres had in the 2000's too.
We still take for granted how hard basic web technology is for people who don't consider themselves technology people though.
reply