2024 is going to shift into a tough year for AI. Business minded folks are already starting to deeply question where the value is relative to the amount of money spent on training. Many/most of the GenAI companies have interesting ideas but no real business plan. Many of the larger AI companies look very shaky in terms of their governance and long term stability. Stability is showing itself to be very unstable, OpenAI has its mess that still seems not fully resolved, Inflection had a bunch of strange stuff go down earlier this week, and more to come.
I’m a huge fan of the tech, but as reality sets in things are gonna get quite rough and there will need to be a painful culling of the AI space before sustainable and long term value materializes.
> Business minded folks are already starting to deeply question where the value is relative to the amount of money spent on training
Kinda. In my experience, the bigger issue is the skillset has largely been diffused.
Overtraining on internal corpora has been more than enough to enable automation benefits, and the ecosystem around ML is very robust now - 10 years ago SDKs like Scikit-learn or PyTorch were much less robust than they are now. Implementing commercial grade SVM or <insert_model_here> is fairly straightforward now.
ML models have largely been commodified, and for most usecases, the process of implementing models fairly straightforward internally.
IMO, the real value will be on the infrastructural side of ML - how to simply and enhance deployment, how to manage API security, how to manage multiple concurrent deployments, how to maximize performance, etc.
And I have put my money where my mouth is for this thesis, as it is one that has been validated by every peer of mine as well.
I agree with you. My reference in business questioning value was on the flood of money going into building new models. That’s where we will see a significant and painful culling soon as it’s all becoming quite commoditized and there’s only so much room in the market when that happens. Tooling, security services and other things to build on top of a commoditized generative AI market opens up other doors for products to be built.
In my experience, there really hasn't been that significant of a flood of money in this space for several years now, or at least not to the level I've seen based on discussion here on HN.
I think HN tends to skew towards conversations around models for some reason, but almost all my peers are either funding or working on either tooling or ML driven applications since 2021.
-------
I've found HN to have a horrible noise to value signal nowadays, and people with experience (eg. My friends in the YC community) deviating towards Bookface or in person meetups instead now.
There was a flood of new accounts in the 2020-22 period (my hunch is it's LessWrong, SSC, and Reddit driven based on the inside jokes and posts I've been seeing recently on HN) but they don't reflect the actual reality of the industry.
I agree that the quality of the posts has decayed dramatically since the start of the COVID pandemic and lots of the Reddit type memes and upvotes have added horrible levels of noise. I can still find gems in it and it’s still miles ahead of Twitter, but I do question my time using it passively and would much rather have a smaller and more focused community again.
Been on HN off and on for about 9 years with different accounts. I do believe the comments are getting more difficult to parse (find the good vs noise) but still the best site I have found - is hand curated links and people willing to offer solid advice given a question in the comments. I use Reddit from time to time but only for pure time killing purposes. I have learned more from the comments on HN than any course, book or (other) web site. Hopefully it will stay that way for years to come. A smaller community could offer more quality responses but less chance of finding the person who has the answer to your question.
I’ve been in crypto since 2010 and HN hasn’t predicted anything. At least with AI the congnoscenti here is on board. With crypto / blockchain the current plan is to pretend it doesn’t exist and never speak of it, as the industry never imploded as was long predicted.
Deep down I fundamentally believe it's a solution searching for a problem. That said, if people want to put money into it, who am I to judge.
That said, IMO crypto as a growth story is largely done. Now that Coinbase has IPOed and the industry consolidated or wiped out (eg. Kraken, FTX, OpenSea), it's not as attractive an industry anymore unless some actual fundamental problems are found that can't be remediated by existing financial infrastructure (legal and illegal).
Hard disagree but even if all of the growth is done, you never would have captured a penny of it reading HN. It went from “scam” to “ponzi” to “government will surely ban it” to “ok it’s over no need to look at or speak of it”.
I'm not denying that there's money in the industry - every portion of the software industry can rake cash due to the extremely high margins that software has compared to just about any other industry.
The magic is finding which subsegments might have even higher margins than others.
Crypto used to have fairly high margins, but all the easy gains have been claimed by larger firms as it's a much more mature industry now, but portions of the ML space still have much more opportunity for growth, so it makes sense to deploy capital there (this is a very high level view so take with a grain of salt - B2C and SMB B2B and Enterprise B2B have entirely different GTM motions and path to profitability).
But at least for me, I can't justify participating in a Series A round for a crypto startup compared to an MLOps startup today.
It's still a scam even if bitcoin is currently benefiting from the creation of institutional ETFs. FTX was one of the largest financial frauds in history and the founder was disgraced and convicted. There's clearly a large amount of appetite for regulatory action in government right now and afaik the funding in the space has completely dried up. What did we get for all that trouble? A really inefficiet payments system? "Digital Gold"? A fun way to fund terrorists and drug running? Crypto proponents are delusional if they think they can just shake off the reputational damage they brought onto themselves in the past several years.
VCs learning nothing or cynically continuing to try and cash in on this scam isn't exactly an airtight argument, even if my original statement was wrong. FTX also raised a ton of money.
It’s just further proof that people commenting on HN confidently about crypto, do not have any idea what is happening in crypto. This website is not plugged in and you won’t pick up any information about it here.
Meh. Crypto is still a waste of resources that had no reason to exist. The modern art market hasn't crashed yet either, and homoeopathy is booming, too last I checked. Just because it's stupid and useless doesn't mean you can't sell it. (And if you underestimate the possibility of selling useless things, you end up incorrectly predicting the demise of the market a lot.)
Not wanting to be part of that is just having some morals. Grabbing money where you can isn't a sign of a successful life.
There are definitely still homeopathic treatments that work, but have not received the investment necessary to go through clinical trials (or haven’t caught on due to lack of recognition from insurance companies).
Yes. Each query needs to generate enough revenue to pay for the cost of running the query, a proportional cost of the cost to train the original model, overhead, SG&A, etc. just to break even. Few have shown a plan to do that or explained in a defensible way how they’re going to get there.
A challenge at the moment is a lot of the AI movement is led by folks that are brilliant technologists but have little to no experience running viable businesses or building a viable business plan. That was clearly part of why OpenAI has its turmoil in that some where trying to be tech purists where others knew the whole AI space will implode if it’s not a viable business. To some degree that seems to behind a lot of the recent chaos inside these larger AI companies.
It's always fascinating to see. The business portions are much easier to solve for with an engineering mindset, but it seems to be a common issue that engineers never take it into account.
This is saying nothing about "technologists" (or as they're starting to become derided as "wordcels": people that communicate well, but cannot execute anything themselves).
It would be... not trivial, but straightforward to map out the finances on everything involved, and see if there is room from any standpoint (engineering, financial, product, etc.) to get queries to breakeven, or even profitable.
But at that point, I believe the answer will be "no, it's not possible at the moment." So it becomes a game of stalling, and burning more money until R&D finds something new that may change that answer (a big if).
Inference will continue to shift toward profitability, Google and the rest are going to choke the cost to gain traction but as TPU costs come down and GPU costs scale up and models become more efficient the cost to infer will drastically scale down into profitability.
Many here know I personally despise AI, but to take off my tech critic hat for a moment, if subscription models don't work out, we still ultimately do have a bunch of pretty powerful base models to build on. If you can push inference to the edge and run on local machines, you can make your users bear the cost.
I'm fairly sure this is why there's a rush to get AI related capabilities into processors for consumer devices, to offload that computing so they don't bear the ongoing cost and it'll probably be more responsive to the user.
>Many/most of the GenAI companies have interesting ideas but no real business plan
-
This is most likely the reason being the SAMA firing, to be able to re-align to the MIC without terrible consequence, from a PR perspective.
no criticism here aside from the fact that we will see the AI killer fully autonomous robots will be here and unfettered by 'alignments' much sooner than we expected...
And the MIC is where all the unscrupulous monies without audits will come from.
> no criticism here aside from the fact that we will see the AI killer fully autonomous robots will be here and unfettered by 'alignments' much sooner than we expected...
What exactly do you mean with this sentence? That less woke/regulated companies will suddenly leapfrog the giants now? What timeframe are we talking here?
And do you mean for example non public mil applications from US/China or whatever or private models from unknown players?
One thing i've been wondering is that if GPT-4 was 100 mil to train, then there's really a lot of plutocrats, despots, private companies and states for that matter that could in principle 10x that amount if they really wanted to go all in, and maybe they are right now?
The bottleneck is the talent pool out there though, but i'm sure there's a lot people out there from libertarians to nation states that don't care about alignment at all, which is potentially pretty crazy / exciting / worrying depending on view.
Aren’t they? Pretty sure that most tactical and strategical decisions are automated to the bottom. Drones and cameras with all sorts of CV and ML features. You don’t need scary walking talking androids with red eyes to control battlefields and streets. The idea of “Terminator” is similar to a mailman on an antigrav bicycle.
I don’t think the bottleneck is the talent pool. There are plenty of folks smart enough to do what OpenAI did, the question is, did they have enough access to capital?
> That less woke/regulated companies will suddenly leapfrog the giants now?
What kind of "leapfrog" do you think is necessary to produce a "killer fully autonomous robot"?
We've actually had "autonomous killer robots," machines that kill based on the outcome of some sensor plus processing, for centuries, and fairly sophisticated ones have been feasible for decades. (For example it's trivial to build a "killer robot" that triggers off of face or voice recognition.)
The only thing that's changed recently is the kind of decisions and how effectively the robot can go looking for someone to kill.
The interesting thing is who's actively working on this, any despots, private companies, foreign nations, who from the talent pool, criminal orgs or even western military which mostly works for the western elite classes.
I think Stability is in an interesting situation. A few suggestions on its direction and current state:
1. Stability AI's loss of talent at the foundational research layer is worrying. They've lost an incredibly expensive moat and there's enough unsolved problems in the foundation layer (faster models, more energy efficient models, etc.) to ensure Stability provides differentiated offerings. Step 1 should be rectifying the core issues of employment and refocusing this more into the AI lab space. I have no doubt this will require a re-steering of the ship and re-focusing of the "mission".
2. Stability AI's "mission" of building models for every modality everywhere has caused the company to lose focus. Resources are spread thin. With $100M in funding, there should be a pointed focus in certain areas - such as imaging or video. Midjourney has shown there is sufficient value capture already in just 1 modality. E.g. StableLM seems like early revenue rush and a bad bet with poor differentiation.
3. There is sufficient competition on the API layer. Stability's commitment to being open-source will continue to entice researchers and developers but there should be a re-focus on improvements in the applied layer. Deep UX wrappers for image editing and video editing while owning the end to end stack for image generation or video generation would be a great focal point for Stability that separates itself from the competition. People don't pay for images, they pay for images that solves their problems.
> Deep UX wrappers for image editing and video editing while owning the end to end stack for image generation or video generation would be a great focal point for Stability that separates itself from the competition. People don't pay for images, they pay for images that solves their problems.
Recently, during an interview [1], when questioned about OpenAI's Sora, Shantanu Narayen (Adobe CEO) gave an interesting perspective on where value is created. His view (paraphrased generously)..
GenAI entails 3 'layers': Data, Foundational Models and the Interface Layer.
Why Sora may not be a big threat is because Adobe operates not only at first two layers (Data and Foundational model) but also at the interface layer. Not only Adobe perhaps knows better than anyone else what is need and workflow of a moviemaker, but I guess most importantly they already have moviemakers as their customers.
So product companies like Adobe (& Microsoft, Google etc.) are in better position to monetize GenAI. Pure-play AI companies like OpenAI are perhaps in B2B business. Actually, they maybe really in api business, they would have great data, would be building great foundational models and giving results of those as APIs; which other companies who are closer to their unique set of customers with their unique needs would be able to monetize and some part of those $$ flows back to pure-play AI companies
Thanks for linking. I agree and strongly believe product companies are in the best position to monetize Gen AI. Existing distribution channels + companies being extremely fast to add AI features.
Where start-ups like Stability need to be rising to compete will have to be AI-native e.g. products re-thought of from the ground up like an AI image editor or as foundation-level AI research companies, agents or AI infrastructure companies.
There's no reason Stability can't play in both B2B and API if planned and strategized well and OpenAI can definitely pull it off with their tech and talent. But Stability has a few important differentiators from OpenAI where I believe if they launch an AI-native product in the multimodal space, they stand to differentiate significantly:
- People join because they believed in Emad's vision of open source so it is their job to figure out a commercial model for open source. They can retain AI talent by ensuring a commitment to open source here. If they need to ensure their moat is retained and can commercialize, they should delay releasing model weights until a product surrounding the weights has been released first. Still open source and open weights but give them time to figure out a commercial strategy to capitalize their research. However because of this promise, they will not be able to license their technology to other companies.
- Stability's strong research DNA (unsure about their engineering) is so badly fumbled by a lack of a cohesive product strategy that it leads to sub-par product releases. In agreement to the 3 'layers' argument, that's exactly Stability's greatest strength and weakness. Their focus on foundational models is incredibly strong and has come at the cost of the interface layer (and ultimately the data layer as it has a flywheel effect).
The company currently screams a need for effective leadership that can add on interface and data layers to their product strategy so they can build a strong moat outside of a strong research team which has shown it can disappear at any moment...
Many years ago it was a good offering but it’s becoming increasingly clear with outsourcing and talent drain that the current teams working on the likes of Photoshop, After Effects and Premier do not actually understand how the core tool, both in its inner workings or even how it draws its own UI works at all and couldn’t either recreate it or even change its existing behavior.
Every major change in the last 6 years has either been weird window dressing changes to welcome panels or new document panels, in all cases building sluggish jank heavy interfaces, try navigating to a folder in the premier one and weep as clicks take actual seconds to recognize.
Or just silly floating tooltips like the ones in Photoshop that also take a second to visible draw in.
All tangible tool changes exist outside the interface or you jump to a web interface in a window and back with the results being passed between in a way that makes it very obvious the developers are trying to avoid touching the core tools code.
Very clear Narayens outsourcing and not being a product guy has lead to this
It's been like this since at least late 90s. At this point Photoshop is similar to Windows in that it has at least 6 mismatching UIs from 6 different eras in it. (or maybe more)
Have been building a generative video editor and doing interviews with enterprise creative cloud users. Basically there’s a large knowledge investment in their tools, and adobe (so far) has shown that their user’s knowledge investment won’t become useless because adobe will continue to incorporate the future into their existing knowledge stack. Adobe doesn’t have to be the best, but just show they won’t miss the boat entirely with this next generation of tools.
I don't know Adobe's business so could be wrong, but maybe "creatives" are not their key customers? If they're focusing on enterprise sales, they're selling to enterprise decision makers.
Every user hates using microsoft products, and don't get me started on SAP. But these are gigantic companies with wildly successful products aimed at enterprise customers.
Only because they've never had a chance to experience the competition.
Having worked in IBM and had to use the Lotus Office Suite I can tell you Microsoft won fair and square. And I'm not even talking about the detestable abomination that is Lotus Notes.
This comment distracted me as I tried to recall a product I saw my mother using when I was younger. I fell asleep after I found it and never got back here until just now: Microsoft Works
This one was so obscure to find because it seems to exist in a weird space of being Microsoft Office but not.
I work with Adobe products everyday. They are mostly trash, at least in the video arena. People use them but they are painful to use. Premiere is known for poor stability, After Effects has just started being updated again after maybe 10+ years of nothing useful being added. Some parts of After Effects are clearly bolted on, and incompatible with each other.
They’re terrible products for the most part. But they buy the competition (Substance) or they develop some half baked substitute (Adobe XD) that people might use since it comes with the subscription.
Sure there maybe scope of improvement in the products, but the point is they have $Billions in sale (year on year) to those customers (Ad agencies, movie studios etc.).
> Figma could build an illustrator killer in 6 months if they wanted to and it would be obliterated
Statements like this are almost always wrong, if for no other reason that a technically superior alternative is rarely compelling enough by itself. It that weren’t the case you would see it happen far more often…
That’s a different scenario: could a system be built to eat X’s lunch.
The scenario I was calling out is more: Company A is great at X, so clearly that technology could be used to easily be great at Y, and doom company B in 6 months. The problem with that thinking is that often the technology is the easy part (and sometimes superficial similarity, also).
They tried to ponder to the open source CEO’s but as much as open source is an ideal, it’s a pretty sure way to failure for the most part. The only way open source works is if a rich company open sources parts of their non revenue forming items.
Yes - open source is a good go to market strategy. They should not open source if there's no good, strong business strategy around how they plan to commercialize. But there are many ways to commercialize an open source model. For many it's lead generation into a paid offering or a managed offering of the open source model.
He's leaving to work on decentralized AI? That's exactly what Stability AI was doing before it became clear the economics no longer work out in practice, and starting a new company wouldn't change that. (Emad is an advisory board member to a decentralized GPU company, though: https://home.otoy.com/stabilityai/ )
Obviously this is the polite way to send him off given the latest news about his leadership, but this rationale doesn't track.
Instead of wasting all the compute on bitcoin we pretrain fully open models which can run on people's hardware. A 120b ternary model is the most interesting thing in the world. No one can train one now because you need a billion dollar super computer.
SETI made sense because there is a lot of data where you download chunk and do expensive computation and return thin result.
Model training is unlike that. It's large state that is constantly updated and updates require full, up to date state.
This means you cannot distribute it efficiently over slow network with many smaller workers.
That's why NVIDIA is providing scalable clusters with specialized connectivity so they have ultra low latency and massive throughput.
Even in those setups it takes ie. a month to train base model.
Converted to distributed setup this same task would take billions of years - ie. it's not feasible.
There aren't any known ways of contributing computation without access to the full state. This would require completely different architecture, not only "different than transformers" but "different than gradient descent", which would be basically creating new branch in machine learning and starting from zero.
Safe bet is on "ain't going to happen" - better to focus on current state of art and keep advancing it until it builds itself and anything else we can dream of to reach this "mission fucking accomplished".
You're right that parameter updates typically require full state, but during my PhD I've explored some possibilities to address this limitation (unfortunately, my ideas didn't pan out in the time I had). That said, there is research that has explored this topic and made some progress, such as this paper:
Unfortunately it's hardly progress. MoE expert models are still large, have to be trained in usual, linear way, this approach requires training set classification upfront, each expert model is completely independent, each has to relearn concepts, your overall model is as good as dedicated expert, scale is in low numbers ie. 8, not thousands (otherwise you'd have to run inference on beefed up cluster only, experts still have to be loaded when used) etc.
But if we think of mixture of experts models outperforming "monolithic" models, why not? Maybe instead of 8 you can do 1000 and that is easy to paralellize. It sounds worth exploring to me.
I think the MoE models are trained together just like any other network though, including the dispatcher layer that has to learn which "expert" route each token to. Perhaps you could do some kind of technically worse model architecture that is trained separately and then a more complex dispatcher that then learns to utilize the individually trained experts as best as it can?
During MoE training you still need access to all weights.
1k experts would mean 30 TB of state to juggle with on 7B params. Training and inference is infeasible at this size.
If you'd want to keep the size while increasing number of experts, you'd end up with 7b -> 56m model. What kind of computation can you do on 56m model? Remember that expert model in MoE runs the whole inference without consulting or otherwise reusing any information from other experts. Thin network at the top just routes it to one of experts. But at this small size those are not "experts" anymore, it'd be more like Mixture of Idiots.
To put it in other way, MoE is optimization technique with low scaling ceiling that is more local maximum solution that global one (this idea works against you quickly if you want to go more that direction).
I don't think MoE allows for that either. You'd have to come up with a whole new architecture that allows parts to be trained independently and still somehow be merged together in the end.
Cool paper. It's more independent than dense or normal MoE but I think it's still far away from the distributed training you're looking for because you still need a seed LM which is trained normally and when fine-tuning each expert from the seed LM, you still need enough GPUs or VRAM to fine-tune that LLM so you're still limited to large GPU clusters which is the problem we're trying to avoid.
In the case of the paper, they are using OPT-6.7b as the seed LM which requires 8xV100 GPUs for fine-tuning each expert. That's a combined total of 256GB of VRAM for a single expert while the 3090 only has 24GB of VRAM and is still one of the most expensive GPUs out there.
Maybe we could use something like PEFT or QLoRA in combination with this technique to make each expert small enough for the community to fine-tune and make a worse Mixtral 8x7b, but I don't know enough to say for sure.
Or maybe it turns out we can make a good MoE model with thousands of smaller experts. Experts small enough for a separate member of the community to independently fine-tune on a normal GPU, but idk.
To have both a performant and distributed LLM trained from scratch, we still need a completely different architecture to do it, but this work is pretty cool and may mean that if nothing else, there is something the community can do to help move things forward.
Also, I was going to say the MoE routing on this technique was lacking, but I found a more recent paper[0] by Meta which fixes this with a final fine-tuning stage.
Base model was still trained in usual, non distributed way (by far the most cost).
Fine tunes were also trained in usual, non distributed way.
Proposed approach tries out several combinations to pick one that seems to perform better (where combination means ie. adhoc per layer operation).
Merging is not distributed as well.
There is not much distribution happening overall beyond the fact that fine tunes were trained independently.
Taking weight averages, weighted weight averages, trimming low diffs, doing arithmetic (subtracting base model from fine tune) etc. are all ad hoc trials throwing something on the wall and seeing what sticks the most. None of those work well.
For distributed training to work we'd have to have better algebra around this multidimentional/multilayer/multiconnectivity state. We don't have it and it has many problems, ie. evaluation is way too expensive. But solving "no need to rerun through whole training/benchmark corpus to see if my tiny change is better or not" problem will mean we solved problem of extracting essence of intelligence. If we do that, then hyper-efficient data centers will still keep beating out any distributed approach and it's all largely irrelevant because that's pure AGI already.
That's wrong. What you described is data parallelism and it would indeed be very tricky to e.g. sync gradients across machines. But this is not the only method of training neural nets (transformers or any other kind) in parallel. If we'd like to train, say, a human brain complexity level model with 10^15 parameters, we'd need a model parallelism approach anyways. It introduces a bit of complexity since you need to make sure that each distributed part of the model can run individually with roughly the same amount of compute, but you no longer need to worry about syncing anything (or have the entire state of anything on one machine). The real questions is if you can find enough people to run this who will never be able to run it themselves in the end, because inference alone will still require a supercluster. If you have access to that, you might as well train something on it today.
Lack of data parallelism is implied by computation that is performed.
You gradient descend on your state.
Each step needs to work on up to date state otherwise you're computing gradient descend from state that doesn't exist anymore and your computed gradient descent delta is nonsensical if applied to the most recent state (it was calculated on old one, direction that your computation calculated is now wrong).
You also can't calculate it without having access to the whole state. You have to do full forward and backward pass and mutate weights.
There aren't any ways of slicing and distributing that make sense in terms of efficiency.
The reason is that too much data at too high frequency needs to be mutated and then made readable.
That's also the reason why nvidia is focusing so much on hyper efficient interconnects - because that's the bottleneck.
Computation itself is way ahead of in/out data transfer. Data transfer is the main problem and going in the direction of architecture that dramatically reduces it by several orders of magnitude is just not the way to go.
If somebody solves this problem it'll mean they solved much more interesting problem – because it'll mean you can locally uptrain model and inject this knowledge into bigger one arbitrarily.
Your gradient descent is an operation on a directed acyclic graph. The graph itself is stateless. You can do parts of the graph without needing to have access to the entire graph, particularly for transformers. In fact this is already done today for training and inference of large models. The transfer bottleneck is for currently used model sizes and architectures. There's nothing to stop you from building a model so complex that compute itself becomes the bottleneck rather than data transfer. Except its ultimate usability of course, as I already mentioned.
Your DAG is big. It's stateless for single pass. Next one doesn't operate on it anymore, it operates on new, updated one from previous step. It has fully connected sub DAGs.
There is nothing stopping you from distributing assembly/machine code for CPU instructions, yet nobody does it because it doesn't make sense from performance perspective.
Or amazon driving truck from one depo to other to unload one package at a time to "distribute" unloading because "distributing = faster".
Yes, if there was something interesting there you'd think since 2017 something would happen. Reinforcement Learning (that is compared with) is not particularly famous for its performance (it is it's biggest issue and reason for not being used that much). Also transformers don't use it at all.
OpenAI has turned for profit and stopped releasing any tehcnical details regarding architectures or training. So how do you know that nothing has happened? Because they didn't release it? Do you see the issue here?
Wonder if that can be avoided by modifying the training approach. Ideas offhand: group by topic, train a subset of weights per node; figure out which layers have the most divergence and reduce lr on those only.
Read his Wikipedia page and tell me he doesn’t sound like your run of the mill crypto scammer.
> He claims that he holds B.A. and M.A. degrees in mathematics and computer science from the University of Oxford.[7][8] However, according to him, he did not attend his graduation ceremony to receive his degrees, and therefore, he does not technically possess a BA or an MA.[7]
In the US attending your graduation ceremony has zero bearing on whether the university recognizes if you achieved a degree or not. Is the UK or Oxford different in this regard? Who cares if someone attended a ceremony. This sounds fraudulent at first glance. People with legit credentials don't need to add technicalities to their claim.
Kinda like Deltec's "Deputy CEO"? (Tether's bank), or even Deltec itself:
At the start of 2021, according to their website, it was a 55 year old bank. By the end of 2021, it was a 70 year old bank!
The bank's website is a WordPress site. And their customers must be unhappy - online banking hasn't worked for nearly two years at this point.
Anyway, their Deputy CEO gave this hilarious interview from his gaming rig. A 33 year old Deputy CEO, who by his LinkedIn claimed to have graduated HEC Lausanne in Switzerland with a Master of Science at the age of 15... celebrating his graduation by immediately being named Professor of Finance at a university in Lebanon. While dividing his spare time between running hedge funds in Switzerland and uhh... Jacksonville, FL.
The name of his fund? Indepedance [sic] Weath [sic] Management. Yeah, okay.
In this hilariously inept interview, he claimed that people's claims about Deltec's money movements being several times larger than all the banking in their country was due to them misunderstanding the country's two banking licenses, the names of which he "couldn't remember right now" (the Deputy CEO of a bank who can't remember the name of banking licenses), and he "wasn't sure which one they had, but we might have both".
Once the ridicule and all this started piling on, within 24 hours, he was removed from the bank's website leadership page. When people pointed out how suspicious that looked, he was -re-added-.
The bank then deleted the company's entire website and replaced it with a minimally edited WordPress site, where most of the links and buttons were non-functional and remained so for months thereafter.
I mean fuck it, if the cryptobros want to look at all that and say "seems legit to me", alright, let em.
I didn’t go to Oxford, but going to your graduation ceremony isn’t usually a requirement for possessing a BA. The university just mails your diploma to you.
Pretty simple background check would answer that question. If he’s claiming those credentials without actually having them I would assume it be common knowledge by now.
Someone became a US House Rep while lying about an education they did not have and a completely falsified resume. I wouldn't be so quick to assume that if he was lying everyone would know by now.
but now that the crypto boys are back en vogue and are returning from hibernation / ai-vacations due to price levels you can combine 2 hype trends into one and capture the imagination & wallets of 2 intersecting circles of fools!
so if these days someone is talking about decentralized anything i'd bet it involves coinshit again
The underlying reason is that current AI businesses and models are failing to capture any significant economic value. We're going to get there, but it will take some more work. It won't be decades, but a few more years would be helpful
Midjourney is the most popular discord channel by far with 19.5M+ members, $200M in revenue in 2023 with 0 external investments and only 40 employees.
The problem has nothing to do with commercializing image gen AI and all to do with Emad/Stability having seemingly 0 sensible business plans.
Seriously this seemed to be the plan:
Step 1: Release SD for free
Step 2: ???
Step 3: Profit
The vast majority of users couldn't be bothered to take the steps necessary to get it running locally so I don't even think the open sourcing philosophy would have been a serious hurdle to wider commercial adoption.
In my opinion, a paid, easy to use, robust UI around Stability's models should have been the number one priority and they waited far too long to even begin.
There's been a lot of amazing augmentations to the stable diffusion models (ControlNet, Dreambooth etc) that have propped up, lots of free research and implementations because the research community has latched onto the stability models and I feel they failed to capitalize on any of it.
Leonardo.ai have basically done exactly this and seem to be doing OK.
It’s a shame because they’re literally just using stable diffusion for all their tech but built a nicer front end and incorporated control net. No-where else has done this.
Controlnet / instantID etc are the really killer things about SD and make it way more powerful than Midjourney, but they aren’t even available via the stability API. They just don’t seem to care.
InstantID uses a non-commercial licensed model (from insightface) as part of its pipeline so I think that makes it a no-go for being part of Stability's commercial service.
Yes, and MJ has no public API either. Same for Ideogram, I imagine they have at least 10m in the bank, and aren't even bothering making an API despite being SoTA for lots of areas.
That’s not true. He was pretty open about the business plan. The plan was to have open foundational models and provide services to governments and corporations that wanted custom models trained on private data, tailored to their specific jurisdictions and problem domains.
Was there any traction on this? I cannot imagine government services being early customers. What models would the want?Military -- maybe, for simulation or training, but that requires focus, dedicated effort and a lot of time. My 2c.
I've heard this pitch from a few AI labs. I suspect that they will fail, customers just want a model that works in the shortest amount of time and effort. The vast majority of companies do not have useful fine tuning data or skills. Consultancy businesses are low margin and hard to scale.
Heres a Stable Diffusion buisness idea: sign up all the celebrities and artists who are cool with AI, and provide end users / fans with an AI image generation interface, trained on their exclusive likenesses / artwork (loras).
You know, the old tried and true licensed merchandise model. Everybody gets paid.
I think the following isn't said often enough: there must be a reason why there are extremely few celebrities and artists who are cool with AI, and it cannot be something abstract and bureaucratic as copyright concerns although those are problematic.
It's just not there yet. GenAI outputs aren't something audiences wants to hang on a wall. It's something that evoke sense of distress. Otherwise everyone's tracing them at least.
Most people mix up all the different kinds of intellectual property basically all the time[0], so while people say it's about copyright, I (currently) think it's more likely to be a mixture of "moral rights" (the right to be named as the creator of a work) and trademarks (registered or otherwise), and in the case of celebrities, "personality rights": https://en.wikipedia.org/wiki/Personality_rights
> It's just not there yet. GenAI outputs aren't something audiences wants to hang on a wall.
People have a wide range of standards. Last summer I attended the We Are Developers event in Berlin, and there were huge posters that I could easily tell were from AI due to the eyes not matching; more recently, I've used (a better version) to convert a photo of a friend's dog into a renaissance oil painting, and it was beyond my skill to find the flaws with it… yet my friend noticed instantly.
Also, even with "real art", Der Kuss (by Klimt) is widely regarded as being good art, beautiful, romantic, etc. — yet to me, the man looks like he has a broken neck, while the woman looks like she's been decapitated at the shoulder then had her head rotated 90° and reattached via her ear.
> Der Kuss (by Klimt) is widely regarded as being good art,
The point is, generative AI images are not widely regarded as good art. They're often seen as passable for some filler use cases and hard to tell apart from human generations, but not "good".
It's not not-there-yet because AI sometimes generates sixth fingers, it's something another level from Gustav Klimt, Damien Hirst, Kusama Yayoi, or the likes[0]. It could be that genAI is leaving something that human artist would filter out, or because images are too disorganized that they appear to us to be encoding malice or other negative emotions, or maybe I'm just wrong and it's all about anatomy.
But whatever the reason is, IMO, it's way too rarely considered good, gaining too few supportive celebrities and artists and audiences, to work.
0: I admit I'm not well versed with contemporary art, or art in general for that matter
> The point is, generative AI images are not widely regarded as good art. They're often seen as passable for some filler use cases and hard to tell apart from human generations, but not "good".
> It's not not-there-yet because AI sometimes generates sixth fingers, it's something another level from Gustav Klimt
My point is: yes AI is different — it's better. (Or, less provocatively: better by my specific standards).
Always? No. But I chose Der Kuss specifically because of the high regard in which it is held, and yet to my eye it messes with anatomy as badly as if he had put 6 fingers on one of the hands (indeed, my first impression when I look closely at the hand of the man behind the head of the woman, is that the fingers art too long and thumb looks like a finger).
wait what? Isn't that missing the point of expressionism? Klimt's Judith I is basically a photo, surely he can draw sh*t if he wanted to?
But myriad predecessors such as Vermeer, Rembrandt, Van Gogh, da Vinci, et al., have done enough in realism, and also photography was becoming more viable and more prevalent, that artists basically started diversifying? Isn't that what lead to various forms of early 20th century arts like surrealism(super-real -ism), cubism, etc?
I don't mean offense but that's just, surely that level of understanding can't be basis of policy decisions when it comes to moral rights and licensing discussions and "artists should just use AI" and such???
I think you're conflating "good" in the sense of "competent" with "good" in the sense of "ethical" or "legal".
I am asserting here that the AI is (at its best) more competent, not any of the other things.
I suspect that the law will follow the economics, just as it often has done for everything else before — you're communicating with me via a device named after the job that the device made redundant ("computer").
But I said "often" not "always", because the business leaders ignoring the workers they were displacing 200 years ago led to riots, and eventually to the Communist Manifesto. I wouldn't discount this repeating.
--
I've just looked up "Judith I" (I recognise the art, just not the name), and I don't even understand why you're holding this up as an example of "basically a photo".
As for the other artists demonstrating realism: photography made realism redundant despite being initially dismissed as "not real art". Artists were forced to diversify, because a small box of chemistry was allowing unskilled people do their old job faster, cheaper, and better. Photography only became an art in its own right when people found ways to make it hard, for example by travelling the world and using it to document their travels, or with increasingly complex motion pictures.
I suspect that art fulfils the same role in humans as tails fulfil in peacocks: an expensive signal to demonstrate power, such that the difficulty is the entire point and anything which makes it easy is seen as worse than not even trying. This is also why forgeries are a big deal, instead of being "that's a nice picture", and why an original painting can retain a high price despite (or perhaps because of) a large number of extremely cheap prints being plastered onto everything from dorm rooms to chocolate wrappers.
Why would those celebs pay Stability any significant money for this, given they can get it for a one off payment of at most a few hundred dollars salary/opportunity cost by paying an intern to gather the images and feed it into the existing free tools for training a LoRA?
You can already do that with reference images and even for inpainting. No training required. Also no need to pay actors outrageous sums to use their likeness in perpetuity as long as you do business. The licensing still tricky anyways, because even if the face is approved and certified, the entire body and surroundings would also have to be. Otherwise you basically re-invented the celebrity deepfake porn movement. I don't see any A-lister signing up for that.
What's insane to me is the fact that the best interfaces to utilize any of these models, from open source LLMs to open source diffusion models, are still random gradio webUIs made by the 4chan/discord anime profile picture crowd.
Automatic1111, ComfyUI, Oobabooga. There's more value within these 3 projects than within at least 1 billion dollars worth of money thrown around on yet another podunk VC backed firm with no product.
It appears that no one is even trying to seriously compete with them on the two primary things that they excel at - 1. Developer/prosumer focus and 2. extension ecosystem.
Also, if you're a VC/Angel reading my comments about this, I would very much love to talk to you.
For a founder maybe , definitely not for employees .
AI startups need not an insignificant amount of startup capital , you cannot just spend weekends to build like you would a saas app . Model training is expensive so only wealthy individuals can even consider this route
Companies like that have no oversight or control mechanisms when management inevitably goes down crazy paths, also without external valuations option vesting structures are hard to ascertain value.
Sometimes you need to say fuck the money, I’ve already got enough, and I just want to do what I enjoy. It may not be an ideal model for HN but damn not everything in life is about grinding, P/E ratios, and vesting schedules
Yeah, that's easier to say when you have enough. A lot of employees might not be in that privilege position. The reality for some of the folks might be addressing education loans, families to take care of, tuition for kids, medical bills, etc.
As a counter-counter-point that gets rarely discussed on HN, VCs aren't taking as much of the pie as people think. In a 2-founder, 4-engineer company, it wouldn't be unusual to have equity be roughly:
20% investors
70% founders
2-3% employees (1% emp1, 1% emp2, 0.5% emp3, 0.25% emp4)
7% for future employees before next funding round
This is not a fair comparison because you are not taking into account liquidation preferences. Those investors don't have the same class of equity as everyone else. That doesn't matter in the case of lights out success but it matters a great deal in many other scenarios.
Sure. My point was that most employees think that VCs take 80+%, and especially the first few employees usually have no idea just how little equity they have compared to the founders.
There's money to be made for sure, and Stability's sloppy execution and strategy definitely didn't help them. But I think there are also industry-wide factors at play that make AI companies quite brittle for now.
>The vast majority of users couldn't be bothered to take the steps necessary to get it running locally so I don't even think the open sourcing philosophy would have been a serious hurdle to wider commercial adoption.
The more I think about the AI space the more I realize that open sourcing large models is pointless now.
Until you can reasonably buy a rig to run the model there is simply no point in doing this. It's no like you will be edified by setting the weights either.
I think an ethical business model for these business is to release whatever model can fit into a $10,000 machine and keeping the rest closed source until above machine is able to run them.
The released image generation models run on consumer GPUs. Even the big LLMs will run on a $3500 Mac with reasonable performance, and the CPU of a dirt cheap machine if you don't care about it being slow, which is sometimes important and sometimes isn't.
The `big' AI models are trillion parameter models.
The medium sized models like GPT3 and Grok are 185b and 314b respectively.
There is no way for _anyone_ to run these on a sub $50k machine in 2024, and even if you can the token generation speed on CPU is under 0.1 tokens per second.
It's just semantic gymnastics. I'm sure most people will consider LLaMa 70B a big model. Of course if you define big = trillion then sure big = trillion[1].
You can get registered DDR4 for ~$1/GB. A trillion parameter model in FP16 would need ~2TB. Servers that support that much are actually cheap (~$200), the main cost would be the ~$2000 in memory itself. That is going to be dog slow but you can certainly do it if you want to and it doesn't cost $50,000.
For 2TB and the server you're at $1698. You can get a drive bracket for a few bucks and a 2TB SSD for $100 and have almost $200 left over to put faster CPUs in it if you want to.
That's stinking Optane, would work if you're desperate. Normal 128GB LRDIMMs cost more than other DDR4 DIMMs. You can, however, get DDR4 RDIMMs for ~$1/GB:
You can get a decent approximation for LLM performance in tokens/second by dividing the model size in GB by the system's memory bandwidth. That's assuming it's well-optimized and memory rather than compute bound, but those are often both true or pretty close.
And "depending on the task" is the point. There are systems that would be uselessly slow for real-time interaction but if your concern is to have it process confidential data you don't want to upload to a third party you can just let it run and come back whenever it finishes. And releasing the model allows people to do the latter even if machines necessary to do the former are still prohibitively expensive.
Also, hardware gets cheaper over time and it's useful to have the model out there so it's well-optimized and stable by the time fast hardware becomes affordable instead of waiting for the hardware and only then getting to work on the code.
Why would increasing memory bandwidth reduce performance? You said "You can get a decent approximation for LLM performance in tokens/second by dividing the model size in GB by the system's memory bandwidth"
Indeed! Also, Mixtral 8x7b runs just as well on older M1 Max and M2 Max Macs, since LLM inference is memory bandwidth bound and memory bandwidth hasn't significantly changed between M1 and M3.
ChatGPT is 20B according to Microsoft researchers, also the fact that big AI models are trillion parameter models is mostly speculation, about GPT-4 it was spread by geohot.
To be precise, ChatGPT 3.5 turbo being 20B is officially a mistake from a Microsoft Researcher, quoting a wrong source published before the release of chatgpt3.5 turbo. Up to you to believe it or not. But I wouldn’t claim it’s a 20B according to Microsoft Researchers.
I think it became apparent when mixtral came out. I've noticed too during training that my model overwrites useful information so it makes sense for these types of models to have emerged.
Disagree. A few weeks ago, I followed a step-by-step tutorial to diwnlad ollama, which in turn can download various models. On my not-soecisl laptop with a so-so graphics card, Mixtral runs just fine.
As models advance, they will become - not just larger - but also more efficient. Hardware advances. Large models will run just fine on affordable hardware in just a few years.
I’ve come to the opposite conclusion personally - AI model inference requires burst compute, which particularly suits cloud deployment (for these sort of applications).
And while AIs may become more compute-efficient in some respects, the tasks we ask AIs to do will grow larger and more complex.
Sure you might get a good image locally but what about when the market moves to video? Sure chat GPT might give good responses locally, but how long will it take when you want it to refactor an entire codebase?
Not saying that local compute won’t have its use-cases though… and this is just a prediction that may turn out to be spectacularly wrong!
I wonder if MidJourney is still ripping. I’m actually curious if it’s superior to ChatGPT’s Dall-E images… I switched and cancelled my subscription when ChatGPT added images, but I think I was mostly focused on convenience.
If you have a particular style in mind then results may vary but aesthetically Midjourney is generally still the best, however Dalle-3 has every other model beat in terms of prompt adherence.
Image quality, stylistic variety, and resolution are much better than ChatGPT. Prompt following is a little better with ChatGPT, but MJ v6 has narrowed the gap.
>It won't be decades, but a few more years would be helpful
I see it this way to be honest:
- companies will aggresively try to use AI in the next 2-3 years, downsizing themselves in the meantime
- the 3-5 year launch mark will show that downsizing was an awful idea and took too many hits to really be worth it. I don't know if those hits will be in profits (depends on the company) but it will clearly hit that uncanny valley.
- 6-8 year mark will have studios hiring like crazy to get the talent they bled back. They won't be as big as before, but it will grow to a more sane level of operation.
- 10-12 year mark will have the "Apple" of AI finally nail the happy medium between efficiency and profitability (hopefully without devastating the workers, but who knows?). Competitors will follow throw and properly usher the promises AI is making right now.
- 15 year mark is when AI has proper pipelining, training, college courses, legal lines, etc. established and becomes standard faire, no stranger than using an IDE.
As I see it, companies and AI tech alike are trying to pretend to be the 10 year mark all the while we're currently in legal talks and figuring out what and where to use AI to begin with. In my biased opinion, I hope there's enough red tape on generative art to make it not worth it for large studios to leverage it easily (e.g. generative art loses all copyright/trademarkability, even if using owned IPs. Likely not that extreme, but close).
Companies are aware of the current AI generation being a tool and not a full replacement (or they will be after the first experiments they perform).
They will not downsize, they will train their workforce or hire replacements that are willing to pick up these more powerful and efficient tools. In the hands of a skilled professional there will be no uncanny valley.
This will result in surplus funds, that can be invested in more talent, which in turn will keep feeding AI development. The only way is up.
Not allowing copyright on AI generated work is a ridiculous and untenable decision that will be overturned eventually.
You greatly overestimate how quickly companies learned. Outsourcing has been a thing for decades and to this day some people are still trying to do it to cut costs. I guess that's what happens when you don't value retention nor document previous decisions. You repeat the cycle.
Sure, the smart companies will use it as a tool, but most companies aren't smart, or just don't care. It'll vary by industry. There is already talks of sizing down VFX/Animation for a mix of outsourcing and AI reliance, for example. And industry that already underpays its artists.
>Not allowing copyright on AI generated work is a ridiculous and untenable decision that will be overturned eventually.
Maybe, once the dust settles on who and what and how you copyright AI. It'll be a while, though. But I get the logic. No one can (nor wants to) succinctly explain what sources were used in a generative art work right now, and that generative process drives the art a lot more than the artist for most generative art. Even without AI there is a line between "I lightly edited this existing work on photosshop" and "I significantly altered a base template to the point where you can't recognize the template anymore" where copyright will kick in.
Still, my biased hopes involve them being very strict with this line. You can't just give 2 prompts and expect to "own" an artwork.
Models are already general, and you can set them to recursively self improve narrowly and still need humans in the loop for general improvement... but those humans are only going to get less... so change your years into months instead and multiply by rand(2)... and change the hire-backs into more startups doing more things...?
>so change your years into months instead and multiply by rand(2)... and change the hire-backs into more startups doing more things...?
I disagree completely. But I should note I was referring to medium to large scale companies. Nothing in those companies happens in "months" these days.
Maybe some startups rise from being first to market much faster, but given the huge legal issues I'm not seeing it. Microsoft et al. can afford A lengthy legal battle and make backup plans. A startup can't.
With the current limitations of AI in mind, it often looks like a solution looking for a problem. It's too unreliable, slow, dumb, and too expensive for a lot of the tasks companies would like to use it for.
And that becomes part of the problem because it's hard to sell unreliable technology unless you design the product in a way that plays well with the current shortcomings. We will get there, but it's still a few iterations away.
I don't think it's that SD and LLMs are solutions looking for problems, it's that there are very clear problems to which they provide 90% of a solution and make it impossible to clear the last 10%.
They're the new WYSIWYG/low-code. Everyone that doesn't fully understand the problem space thinks they're some ultimate solution that is going to revolutionise everything. People that do are responding with a resounding 'meh'.
Stable Diffusion is a great example. Something that can generate consistent game assets would be an absolute game changer for the entire game industry and open up a new wave of high tech indie game development, but despite every "oh wow" demo hitting the front page of HN, we've had the tech for a couple of years now and the only thing that's come out of it is some janky half solutions (3D meshes from pictures that are unworkable in real games, still no way to generate assets in consistent styles without a huge amount of complex tinkering) and a bunch of fucking hentai lol.
Human work is much more deterministic than AI as it encompasses a lot more constraints than what the task specified. If you take concept art creations, while the brief may be a few sentences, the artist knows to respect anatomy and perspective rules, as well as some common definitions (when the brief says ship, you know that it’s the ship concept approved last week). As an artist, I’ve used reference pictures, dolls, 3d renders and one of the most aspect these tools had was consistency. I don’t see Large Models be consistent without another models applying constraint to what they’re capable of producing, like rules defining correct anatomy and extracting data that defines a character. The fact is we do have tools like MakeHuman [0], Marvelous Designer [1], and others that let you generate ideas that are consistent in their flexibility.
I look at Copilot and it’s been the same for me. I’m either working on a huge codebase and most of the time, it means tweaking and refactoring, which is not something I trust a LLM with. Or it’s a greenfield project and I usually write only the necessary code for a task and boilerplate generation is not a thing for me. Coding for me is like sculpting and LLM-based solutions feel like trying to do with bricks attached to my feet. You can get something working if you’re patient enough, but it’s make more sense and it’s more enjoyable to just use your fingers.
And even a lot of the hentai is fucking worthless for the same reasons! Try generating some kinbaku. It's really hard to get something where all the rope actually connects and interacts sensibly because it doesn't actually know what a knot is. Instead, you end up with M. C. Escher: Fetish Edition.
I can shed some light on this phenomena: These models are trained on many images but no thought is put into the "generalisation" aspect the ML community was so obsessed with during the deep-learning era.
It's very easy to create a "Stochastic Parrot" but I'm quite sure these models are capable of learning underlying information such as correct layout of a knot - given the right data and curriculum of course. Maybe slight architecture tweaks.
I'm sure this is the reason we're starting to see a normal amount of fingers or ability to write text. Proof of concept was 2015 until 2022 now we're starting to see interesting things come out of the workshops.
Pretty much what happened with Speech Recognition for 30 years. That last 10% had to be handled manually. Even if you get 90% right, it still means ever second sentence has issues. And as things scale up the costs of all that manual behind the scenes hacking scale up too. We underestimated how many issues involved Ambiguity - where N people see the same thing and have N different interpretations. So you see a whole bunch of Speech Rec companies rising and falling over time.
Now things are pretty mature, but it took decades to get there but there is still a whole bunch of hacks upon hacks behind the scenes. Same story will repeat with each new problem domain.
We use Whisper for automatic translation, supposedly SotA, but we have to fix its output, I would say, very often. It repeats things, translates things for no reason, has trouble with numbers.. it's improved in leaps and bounds but I'd say that speech recognition doesn't seem to be there yet.
A game with no consistency in the art is probably enabled. We've crossed the threshold where something like Magic the Gathering could be recreated by a tiny team and a low budget.
I don't think the limiting factor here is the software; it looks like we got AI-generated art pretty much as soon as consumer graphics cards could handle it (10 years ago it would have been quite hard). I'd be measuring progress in hardware generations not years and from that perspective Stable Diffusion is young.
Current AI is entirely incapable of generating the balanced and fun/engaging rule sets required for a MtG style game. Sure the art assets could be generated with skilled prompting and touchup but even that is nowhere close to the strong statement you made.
OP likely meant that a Midjourney-level AI can easily generate all the card art.
Obviously, current AIs cannot generate game rulesets because the game feel is an internal phenomenon that cannot be represented in the material domain and therefore AIs cannot train on it.
It would be negligent for them to aim for profitability right now.
They're doing what they should: growing the customer base while continuing to work on the next generation of the core technology, and developing the support code to apply what they have to as broad a cross-section of problems as they have the potential to offer a solution for.
So the problem might be two-fold. First, there's an oversupply of companies trying to use AI relative to the current technological capabilities. Second, even when value is created, the companies don't capture the value to profit out of it.
Pretty sure YouTube has been losing a lot of money every year before they very aggressively ramped up ad frequency and duration as well as subscriptions in recent years. They could only do that thanks to Google’s main cash cow; YouTube would have been dead if not acquired and subsidized by Google.
studios are the target here, not consumers. Pareto principle applies when you need more than a passive user. 20% or less of the serious studios (which is already a minority) will end up providing 80% of the value of any given AI solution.
No. There can be tremendous value in AI, you just won't find it here.
The reason is that Stability.ai gave away everything for free. Until recently, they didn't even attempt to charge money for their models.
I've heard the only reason they're not already closed up is that they're reselling all of the rented GPU quota they leased out years ago. Companies are subletting from Stability, which locked in lots of long term GPU processing capacity.
There's no business plan here. This is the Movie Pass of AI.
90% of startups fail, so it's not just a matter of waiting for the tech to get better and having more value - most of the current players will simply fail and go out of business.
Really sad to see. Emad pivoting to posting crypto nonsense on twitter made me think the writing is on the wall for Stability, but I still didn't expect it so soon. I expect they'll pivot to closed models and then fade away.
> Really sad to see. Emad pivoting to posting crypto nonsense on twitter made me think the writing is on the wall for Stability
Open-source AI is a race to zero that makes little money and Stability was facing lawsuits (especially with Getty) which are mounting into the millions and the company was already burning tens of millions.
Despite being the actual "Open AI", Stability cannot afford to sustain itself doing so.
Wow much of what he’s saying in that video is a lie with regards to the history of latent diffusion, creating an open source GPT3, etc. Just taking credit for a bunch of work he didn’t have much to do with.
You know it never ceases to amaze me how even the most respected fall prey to this money laundering scheme. If people even spent some time to read about Tether they would not touch this stuff. It's blood money.
You should probably post some citations to that. Tether most probably have backing for every single dollar they have, and then plenty more. What about it is “blood money”?
I mean, is AI a less sketchy space in 2024 than crypto/blockchain in 2024? Two or three years ago sure, I guess, but today?
The drama around OpenAI is well documented, there are multiple lawsuits and an SEC investigation at least in embryo, Karpathy bounced and Ilya's harder to spot than Kate Middleton (edit: please see below edit in regards to this tasteless quip). NVIDIA is pushing the Dutch East India Company by some measures of profitability with AMD's full cooperation: George Hotz doesn't knuckle under to the man easily and he's thrown in the towel on ever getting usable drivers on "gaming"-class gear. At least now I guess the Su-Huang Thanksgiving dinners will be less awkward.
Of the now over a dozen FAANG/AI "boomerangs" I know, all of them predate COVID hiring or whatever and all get crammed down on RSU grants they accumulated over years: whether or not phone calls got made it's pretty clearly on everyone's agenda to wash all the ESOP out, neutron bomb the Peninsula, and then hire everyone back with at dramatically lower TC all while blowing EPS out quarter after quarter.
Meanwhile the FOMC is openly talking about looser labor markets via open market operations (that's direct government interference in free labor markets to suppress wages for the pro-capitalism folks, think a little about what capitalism is supposed to mean if you are ok with this), and this against the backdrop of an election between two men having trouble campaigning effectively because one is fighting off dozens of lawsuits including multiple felony charges and the other is flying back and forth between Kiev and Tel Aviv trying to manage two wars he can't seem to manage: IIRC Biden is in Ukraine right now trying to keep Zelenskyy from drone-bombing any more refineries of Urals crude because `CL` or whatever is up like 5% in the last three weeks which is really bad in an election year looking to get nothing but uglier: is anyone really arguing that some meme on /r/crypto is what's pushing e.g. BTC and not a pretty shaky-looking Fed?
Meanwhile over in crypto land, over the same period of time that AI and other marquee Valley tech has been turning into a scandal-plagued orgy of ugly headlines on a nearly daily basis, the regulators have actually been getting serious about sending bad actors to jail or leaning on them with the prospect (SBF, CZ), major ETFs and futures serviced by reputable exchanges (e.g. CME) have entered mainstream portfolios, and a new generation of exchanges (`dy/dx`, Vertex, Apex, Orderly) backed by conventional finance investments in robust bridge infrastructure (LayerZero) are now doing standard Island/ARCA-style efficient matching and then using the blockchain for what it's for: printing a consolidated, Reg NMS/NBBO/SIP-style consolidated tape.
As a freelancer I don't really have a dog in this fight, I judge projects by feasibility, compensation, and minimum ick factor. From the vantage point of my flow the AI projects are sketchier looking on average and below market bids on average contrasted to the blockchain projects, a stark reversal from even six months ago.
Edit: I just saw the news about Kate Middleton, I was unaware of this when I wrote the above which is in extremely poor taste in light of that news. My thoughts and prayers are with her and her family.
I and many many people around me use ChatGPT every single day in our lives. AI has a lot of hype but it’s backed by real crap that’s useful. Crypto on the other hand never really did anything practical except help people buy drugs and launder money. Or make people money in the name of investments.
I love their models and I love how they have changed the entire open source AI ecosystem for the better, but the writing was always on the wall for them given how unprofitable they are.
I don't think much of the AI startup scene or socials groups like e/acc would have existed if it weren't for the tech that they just gave away for free.
Its interesting how Stability AI and their VC funding have done a much better job of acting effectively as a non-profit charity (because they don't have profits. lol) to speed up AI development and open source their results as compared to other companies that were supposed to have been doing that from the beginning.
They really were the true ActuallyOpenAI.
Related to this, if you are an aspiring person who wants to improve the world, tricking a bunch of VC investors to fund your tech and then giving away the results to everyone free of charge is the single best way to do it.
Hey, let me know when the doomers have built anything that has mattered because they choose to, and not because they are forced to my market forces.
At least the open source AI people have code that you can use, freely without restriction.
The doomers, on the other hand, don't do anything but try and fail to prevent other people from releasing useful stuff.
But, in some sense I should be thanking the doomers because I rather that people with such incompetence were the enemy as opposed to people who might have a chance of succeeding.
Not being e/acc doesn’t make one a doomer. It means being someone with a healthy relationship with technological progress, who hasn’t given up on the prospect of useful regulation where it’s helpful.
The US is great but the best AI researchers aren’t gonna want to live here if it becomes a hyper libertarian hellscape. They want to raise a family without their kids being exploited by technologies in ways that e/accs tell us we should just accept. It’s not sustainable.
You may want to read Marc Andreessen’s manifesto. E/acc isn’t just about “releasing cool AI tech”.
It means unrestricted technological progress. Unrestricted, including from annoying things like consumer protection, environmental concerns, or even national security.
If e/acc was just about making cool open source stuff and posting memes on the Internet, you wouldn’t need a new term for it, that’s what people have been doing for the past 30 years.
If Marc writes something it doesn’t become the definition of e/acc. Marc is hyperbolic and he gets a lot of clicks and eyeballs. As a VC though, he does it for his interest.
E/acc has many interpretations. In the most basic sense it means “technology accelerates growth”. One should work on better technology and making it widely distributed. Instead of giving away money, one can have the biggest impact on humanity with e/acc.
we’ve been effectively accelerating for the past 200 years.
Ok, thats nice and all. But the actual results of all of this is memes and people making cool startups.
Regardless of whether a couple people who are taking their own jokes too seriously truly believe that they are going to, I don't know, create magic AGI, the fact remains that the actual measurable results of all this is only:
1: funny memes
2: cool AI startups
Anything other than that is made up stuff in either your head, or the heads of people who just want to pump up the valuation of their startups.
> Marc Andreessen’s manifesto
Yes, I'm sure he says a lot of things that will convince people to invest in companies that he also invests in. It is an effective marketing tactic. I'm sure he has convinced other VCs to invest in his companies because of these marketing slogans.
But regardless of what a couple people say to hype up their startup investments, that is unrelated to the actual real world outcomes of all of this.
> you wouldn’t need a new term for it
The fact that me and you are talking about it, actually proves that yes some marketing terms both make a difference and also don't result in, I don't know, the government being overthrown and replaced by libertarian VCs or whatever nonsense that people are worried about.
Exactly. There are two groups of people: ones that defend Effective Accelerationism online with a straight face, and ones that take memes too seriously
> This is one of the worst advices I've ever seen.
Really? Because stability AI caused a very large amount of good in the world.
It arguably kicked off the entire AI startup industry.
> Tricking people to obtain money? How's this not fraud?
Its not fraud because you don't have to lie to anyone. You can tell VCs exactly what you plan on doing. Which is to open source all of your code... and... uhhh... yeah that will totally make the company valuable.
There are lots of ways of making sales pitches about open source, or similar, that will absolutely pass regulatory scrutiny and are "honest", and yet still have no hope of commercial success and also provide a huge amount of value to the public. Like what stability AI did.
A company that sets out to obtain VC money and blow it all on open source software without turning a profit, is going to leave behind smoking guns. Those will turn up in discovery and make one's life rather difficult.
Sure they did. They have been open from the start that they were releasing everything open source. They have been very up front about that!
> is going to leave behind smoking guns. Those will turn up in discover
No it won't, and it didn't. The VCs all hopped on board onto a very transparent open source giveaway. Good on them! Nobody lied to anyone about their open source plans.
Looks like many AI startups are experiencing a bit of some turbulence and chaos in the recent months:
First the OpenAI rebellion in November, then the Inflection AI acqui-hire from Microsoft not willing to pay the over-valued $4B and deflected that to $600M instead (after making $0 revenue) and now a bit of in-stability at Stability AI with the CEO resigning after many employees leaving.
What does that say about the other AI companies out there who have raised tons of VC cash and aren't making any meaningful amount of revenue? I guess that is contributing to the collapse of this bubble with only a very few companies surviving.
Amazon nailed high revenue growth from the very beginning, just reinvesting in growth & deferring the margin story. They could have stopped at any time.
Google nailed high traffic from the beginning, so ad sales was always a safe Plan B. The founders hoped to find something more aesthetic to them, failed, and the conservative path worked.
The reason I write this is misleading is b/c this is very different from a ZIRP YC era thinking that seems in line with your suggestion:
- Ex: JustinTV used their VC $ to pivot into Twitch, and if that didn't work, game over.
- Ex: Uber raised bigger & bigger VC rounds until self-driving cars could solve their margins / someone else figured it out. They seem to be figuring out their margins, but it was a growing disaster and unclear if they could with such an unpredictable miracle.
In contrast, both Amazon & Google were in positions of huge cash flows and being able to switch to highly profitable growth at any time. They were designed to control their destinies, vs have bankers/VCs dictate them.
Amazon was famous for not taking profit, and instead putting profit back into the company, but revenue? Amazon was generating gobs and gobs and gobs of revenue from day 0.
I am quite sure Google had a lot to offer from day one, which is why they were encouraged to start the business. There wasn’t any open source Google’s.
Yes maybe the business model wasn’t perfect but ad revenue was already well and truly a thing by the time Google invented a better search engine. All they had to do was serve the ads and the rest is history.
The big companies fighting to win gen AI have forgotten how to build good products. There seems to be enormous confusion about AI applications — few seem to be willing to commit to a singular focus or application, instead trying to be everything to everyone. The tried and true way to build something big is to bootstrap a killer app that solves a real problem and then expand horizontally to other problem spaces. These companies are trying to do too much too soon (probably because they have too few financial constraints). A lack of focus.
This phrasing makes me feel like we're living in some techno-future where corporations are de-facto governance apparatuses, squashing rebellion and dissidents :^)
Colonial companies did this all the time, e.g. the British East India Company, and the 1800s American railroads had the Pinkertons. There's also the phenomenon of the company town.
Decentralized systems, peer to peer, Blockchain, smart contracts, are all important technologies with real use cases. It is not accurate to refer to any of them as simply "crypto" especially in this context.
It's true. Latin America already knows it, USA and Europe are yet to catch up.
Most of crypto is bullshit, but at the very least Bitcoin is a massive, massive use case.
Commercial banks make up most central banks, and if you believe commercial banks aren't grifters or con artists, you've obviously never heard of LIBOR for starters.
As for bad solutions, I'll give you incorrect economic forecasts on inflation and using interest rates as a lever.
“ In reality, Mostaque has a bachelor’s degree, not a master’s degree from Oxford. The hedge fund’s banner year was followed by one so poor that it shut down months later. The U.N. hasn’t worked with him for years. And while Stable Diffusion was the main reason for his own startup Stability AI’s ascent to prominence, its source code was written by a different group of researchers. “Stability, as far as I know, did not even know about this thing when we created it,” Björn Ommer, the professor who led the research, told Forbes. “They jumped on this wagon only later on.” “
“ “What he is good at is taking other people’s work and putting his name on it, or doing stuff that you can’t check if it’s true.”
Emad had the genius idea of essentially paying for the branding rights to an AI model. This was viewed as insane in the pre-ChatGPT era, and only paid off massively in retrospect.
Also all those 'controversies' were mostly the result of an aggrieved co-founder/investor who decided to sell their shares before SD1.4's success. Emad may not have proven to be competent enough to run an large AI lab in the long run, but those complaints are just trivial 'controversies'.
They are saying it was not clear "buying branding rights to an image model" would lead to any investments, any kind of high valuation, or any other financial success. It is only clear in hindsight.
Maybe… he certainly was good at taking credit for it. Not course if they stepped in, rebranded something they didn’t make and threw a bunch of AwS GPUs they couldn’t actually afford at it though
People forget, moving things around and funding them and making it all work is what makes an entrepreneur. Elon Musk did that and all great founders do that. Emad doesn't have to write the code himself.
Thats a hit piece, but whatever, isn't that what the most prominent/funded academics do? As far as i know he is known generally as CEO of the company that makes SD, not as the creator of SD. It does look like without him these models wouldnt have evolved so much
> What he is good at is taking other people’s work and putting his name on it, or doing stuff that you can’t check if it’s true.
I have to say, this is a quite common ignorant statement that's said about almost every CEO.
I'm not sure if there's more to it in this particular case, but no, CEOs aren't stealing your work. Similarly, marketers aren't parasites. Designers aren't there to waste your time. Many engineers seem to hold similar belief that others are holding them down or taking advantage of their work. This is just a congnitive bias.
Emad jumped on the train after the major inventions were invented and PoCs were made. He could not have contributed to them unless he had a time machine. (Yes he contributed to the training costs of SD1.4 but the time point when he made the decision was not early research.)
You made a point about devs underappreciating the work of other professionals, like CEOs and designers. You made this point in the context of Emad and the success story that is Stable Diffusion. The implication in your point is that Emad surely contributed to the success, even though the CEO is not a developer or researcher. My counter to your point is that Emad wasn't there when the inventions were made. He joined the party after success was already evident. Emad's main contribution to the success of Stable Diffusion is that he funded the training of a large model. That's great, but this thing would've happened with or without him. The inventions were made before he funded the training of SD1.4.
If Emad had been supporting the research team from the beginning, one might argue that he created the conditions for their success, or whatever. But he wasn't there at that time.
None of this is related to whatever Emad is accused of. I'm just making a point about how we attribute contribution for success.
This is 100% correct. Emad tried to lure me into the trap that Eleuther eventually fell into, and I lucked out by blowing him off after getting weird vibes from him. This was back when he was unknown, but was running around offering a bunch of researchers massive GPU cluster time for seemingly altruistic reasons but were in fact creepy reasons. In reality he wanted his name on their work.
I have the DMs to prove this, and have not ever said something like this about someone. I wouldn’t make this accusation lightly, for whatever it’s worth. In the HN discussion of that article, I had left a comment, which Emad DMed me on Twitter about, saying no no, he never lied to investors, and tried to convince me that what I was saying wasn’t true. I was wondering why he cared so much. In retrospect it’s probably because it was correct.
I’ve never worked with him, to be clear, and the few colleagues who have worked at Stability have had generally positive things to say. But there was one that was screwed over by them hard (he was doing contract work, and never got paid for it), and I can think of at least four other alarming data points that all point to the same thing.
It’s unsettling not knowing whether to speak up about this. On one hand it doesn’t really matter that much. On the other hand, it’s the fundamental difference between a CEO that tends to IPO vs one that tends to fail. I hate seeing people fail, and I genuinely thought that my feelings about Emad were mistaken since empirically they were doing fine. Turns out, nope, not fine, and the original impression was right. Weird experience.
Not surprised at all to see how unpopular this sentiment is on HN. For some reason HN seems to love one dimensional stereotypes for every job that isn't theirs.
Actually that's wrong, even the idea that engineers are smarter than managers is very prevalent here.
Emad is such an obvious grifter it’s honestly mad that he attracted so much VC money.
He couldn’t even get his own story straight regarding his education and qualifications which should be a pretty clear disqualifying red flag from the outset.
The Forbes article from last year was dismissed on here as a hit piece but the steady flow of talent out was the clear sign, capped by the original SD authors leaving last week (probably after some vesting event or external funding coming through).
That's my guess too. Emad teased SD3 while it looked like he was looking for more money, but without convincing rationale for not releasing it already. The samples may have been heavily cherry-picked, we don't know if it's actually a decent model in practice.
> Earlier today, Emad Mostaque resigned from his role as CEO of Stability AI and from his position on the Board of Directors of the company to pursue decentralized AI.
Reading explanations of "decentralized AI" [0] sounds like a sales pitch for investors that missed out on the AI and crypto hype. Technically, it sounds like a distributed file storage which already exists.
I interviewed at Stability AI a while ago and that interview was a complete shit show. They quite literally spent 40 minutes talking about Emad and his "vision". I think we actually talked about what they wanted me to do there for like 15 minutes.
I was not feeling confident about them as a company that I wanted to work for before that interview, afterwards I knew that was a company I wouldn't work for.
6 months ago, I expressed my doubt about the viability of Stability business model on HN News. Mostaque answered the questions with a resounding yes and claimed the business of Stability.AI is better than ever.
There have been a series of high profile staff/researcher departures, implied to be partially as a result of Emad's leadership. The latest departures of the researchers who helped develop Stable Diffusion could be fatal: https://news.ycombinator.com/item?id=39768402
Given all the talk here, it seems like the state of the business and the state of the stock market are grossly out of whack. Of course, that happens from time to time.
lol Emad was always seemed like an obvious fraud to me. Not quite SBF level but same vibe. Whenever someone goes overboard on the nerd look it’s always a red flag.
I think when parsing that statement it's important to understand his (and your) definition of "programmer".
We (I) tend to use the term "programmer" in a generic way, encompassing a bunch of tasks waaay beyond "just typing in new code". Whereas I suspect he used it in the narrowest possible definition (literally, code-typer).
My day job (which I call programming) consists of needs analysis, data-modelling, workflow and UI design, coding, documenting, presenting, iterating, debugging, extending, and cycling through this loop multiple times. All while collaborating with customers, managers, co-workers, check-writers and so on.
AI can do -some- of that. And it can do small bits of it really well. It will improve in some of the other bits.
Plus, a new job description will appear- "prompt engineer".
As an aside I prefer the term "software developer " for what I do, I think it's a better description than "programmer".
Maybe one day there'll be an AI that can do software development. Developers that don't need to eat, sleep, or take a piss. But not today.
(P.S. to companies looking to make money with AI - make them able to replace me in Zoom meetings. I'd pay for that...)
That's right. We invented programming AI a very long time ago, and called it an "assembler". All you had to do was tell the assembler what kind of program you wanted, and it would do the programming work for you!
Then we invented another AI to tell the assembler what kind of program you wanted, and called it a "compiler". All you had to do was tell the compiler what kind of program you wanted it to tell the assembler you wanted, and it would do all the not-exactly-programming work for you!
P.S. Visual Basic with its GUI designer was a quite effective way to rapidly build apps of questionable quality but great business value. Somebody should bring that paradigm back.
My day job is programming in an environment which originated in the mid 90s. A contemporary of the Visual Basic era, but somewhat more powerful, and requiring substantially less code.
While I, and a few thousand others still use it (and it gets updated every couple years or so) it has never been fashionable. Ironically because it's perceived as 'not real programming'.
We routinely build systems with hundreds of thousands of lines of code, much of it founded in the 90s and having been added to for 25 years. Most of it was built, and worked on, by individuals, or very small teams. Much of it today is still active doing the boring business software that keep the lights on.
But its not "main stream" because programmers pick language based on popularity, and enterprises pick programmers based on language. A self-fueling cycle of risk aversion.
A lucky few though hot off the treadmill a long time ago and "followed a path less travelled by". And that has made all the difference.
You mean the CEO of the company that just rode the AI hype wave to become one of the top 3 most valuable companies in the world? It's his fiduciary duty to say things like that, whether or not he believes them, the same as every other CEO in the AI space.
There is no fiduciary duty to lie, or make up stories you do not believe.
People really oversell fiduciary duty. Yet the whole point of top-level corporate roles is to steer a company predicated upon opinion, which means that you have great latitude to act without malfeasance.
We have, there are a bunch of bottom of the barrel executives that have done hiring freezes and layoffs under the assumption that AI would replace everything. It will go the exact same way as the outsourcing craze that swept the industry in the mid aughts. The executives that initiate the savings will be praised and rewarded lavishly, and then when everything blows up and falls apart those responsible for the short sighted and ineffective cuts will be long gone.
“Mostaque had embezzled funds from Stability AI to pay the rent for his family's lavish London apartment” and that Hodes learned that he “had a long history of cheating investors in prior ventures in which he was involved”
It also seems like the company just isn't doing very well, it looks like there have been constant problems since Stable Diffusion was released - not enough funding, people leaving, etc. Which I don't get - you create a massive new piece of software that is a huge leap forward and you can't just get more funding? There have to be big structural issues at Stability.
Yet for the 6 months before that they were talking about running out of money, and even the month after they got funded were considering selling the company due to money issues.
I’m practicality shaking my head in disbelief at all the red flags this guy has and people are still defending him. Stability and its work are great. We should support an open community and ethos. And Emad can still be a shady narcissist con man. These are all compatible views.
Stability models are mostly open source. While it was never going to last, Stability put the entire industry in a race to the bottom, all the while building up the open ecosystem.
The most obvious potential real reason is Stability isn't making money, and the people that matter don't think Emad was going to be able to turn it around.
There's nothing to enshitify, their models were open sourced. Now they may no longer release future models for free, but its entitlement to think we'll just get free improvements forever.
Also not all CEO replacements turn out bad. Uber certainly has turned itself around.
What I'm saying is that if it's really true that he is going out because "the people that matter don't think Emad was going to be able to turn it around" then it's very likely that the replacement will be more able to turn it around, because that will be the selection criteria.
I looked this up and apparently the controversy is that Stable Diffusion was trained on child porn? How can the model itself not be considered objectionable material then? Does the law not apply some kind of transitive rule to this? And don't they want to arrest someone for having the child porn to train it on?
To say it was “trained on child porn” is just about the most “well, technically….” thing that can be said about anything.
Several huge commercial and academic projects scraped billions of images off of the Internet and filtered them best they knew how. SD trained on some set of those and later some researchers managed to identify a small number of images classified as CP were still in there.
So, out of the billions of images, was there greater than zero CP images? Yes. Was it intentional/negligent? No. Does it affect the output in any significant way? No. Does it make for internet rage bait and pitchforking? Definitely.
There's two controversies that are the opposite of each other.
1. Some people are mad that Stable Diffusion might be trained on CSAM because the original list of internet images they started with turned out to link to some. (LAION doesn't actually contain images, just links to them.)
This one isn't true, because they removed NSFW content before training.
2. Some other people are mad that they removed NSFW content because they think it's censorship.
That actually isn't the legal issue I meant though. It's not that they trained on it, it's that it contains adult material at all (and can be shown to children easily), and that it can be used to generate simulated CSAM, which some but not all countries are just as unhappy about.
"AI"s are still pretty much vaporwares like 40 years ago. When people get tired of these toys, the bubble will simply burst, and nothing valuable left.
There are a few options. I work with a REPL, so I usually load the answer from a scratch file and put some representative data into it. When the result is wrong, which often happens, I feed ChatGPT the error, and it corrects the code accordingly. Iterating this results in a working function about 80% of the time. Sometimes it loses the plot and I either give up or start over with more detailed instructions.
You can also ask it to write tests, some of which will pass, some of which will fail. It's pretty easy to eyeball whether or not a test is valid, and they won't always be valid, I just fix those by hand.
Here's an example of some thing I tried at random a few weeks ago. I have a bunch of old hand written notes that are mind maps and graphs written on notebooks from 10+ years ago. I snapped a picture of all of the different graphs with my phone, threw them into chatGPT and asked it to convert them to mermaid UML syntax.
Every single one of them converted flawlessly when I brought them into my markdown note tool.
If you're using chatGPT as nothing more than a glorified fact checker and not taking advantage of the multimodal capabilities such as vision, OCR, Python VM, generative imagery, you're really missing the point.
Exactly how you would verify the result that your human underling yielded. You can even delegate the googling and summation to the AI and just verify the verification.
I’m a huge fan of the tech, but as reality sets in things are gonna get quite rough and there will need to be a painful culling of the AI space before sustainable and long term value materializes.