they are designed to elucidate the author's thought process - not the reader's learning process
No, it’s exactly the opposite: when I write papers I follow a rigid template of what a reader (reviewer) expects to see. Abstract, intro, prior/related work, main claim or result, experiments supporting the claim, conclusion, citations. There’s no room or expectation to explain any of the thought process that led to the claim or discovery.
Is that reason because Buffalo is the 81st most populated city in the United States, or 123rd by population density, and Waymo currently only serves approximately 3 cities in North America?
We already let computers control cars because they're better than humans at it when the weather is inclement. It's called ABS.
I would guess you haven't spent much time driving in the winter in the Northeast.
There is an inherent danger to driving in snow and ice. It is a PR nightmare waiting to happen because there is no way around accidents if the cars are on the road all the time in rust belt snow.
I get the feeling that the years I spent in Boston with a car including during the winter and driving to Ithaca somehow aren't enough, but whether or not I have is irrelevant. Still, I'll repeat the advice I was given before you have to drive in snow, go practice driving in the snow (in eg a parking lot) before needing to do so, esp during a storm. Waymo's been spotted driving in Buffalo doing testing, so it seems someone gave them similar advice. https://www.wgrz.com/article/tech/waymo-self-driving-car-pro...
There's always an inherent risk to driving, even in sunny Phoenix, Az. Winter dangers like black ice further multiply that risk but humans still manage to drive in winter. Taking a picture/video of a snowed over road and judging the width and inventing lanes based on the width taking into account snowbanks doesn't take an ML algorithm. Lidar can see black ice while human eyes can not, giving cars equipped with lidar (wether driven by a human or a computer) an advantage over those without it, and Waymo cars currently have lidar.
I'm sure there are new challenges for Waymo to solve before deploying the service in Buffalo, but it's not this unforeseen gotcha parent comment implies.
As far as the possible PR nightmare, you'd never do self driving cars in the first place if you let that fear control you because, you you pointed out, driving on the roads is inherently dangerous with too many unforeseen complications.
I feel like that range is a on the high side, at least for jobs outside the Bay Area. Unless aerospace pays a lot better than software jobs at government contractors. When I was laid off less than 2 years ago and applying to a wide variety of places, the gov contractors were only offering in the $120-$140 ballpark for senior sw engineering positions. They tended to play up gold plated benefits, high 401k match, etc as well but there’s no way that would get it up to $263k.
That is outrageous that two seemingly developed countries could have such a huge compensation gap. A senior aerospace engineer in the UK can make as little as 45k GBP, or 56k USD? One fifth as much as the lowest-paid American??
The take-home on that is £35k. The median rent in London is £26k. I suppose the person making £45k doesn't likely live in London, but still pretty grim.
Americans are often blown away and kind of ignorant of how, relative to the rest of the world, they are really wealthy and well paid. Like, people have way less disposable incomes in other parts of the world, even developed countries. The purchase power of the USD and the power of the US economy is absolutely insane.
Yes, Aerospace Engineers don't live in London because there are very little (if any) aerospace jobs in London. Biggest aerospace employers in the UK are BAE Systems and Airbus, and both have factories in much cheaper locations (Wales, Northwest of England).
You're basically comparing "super specialised job in the middle of nowhere with very low cost of living" vs a "super specialised but much more needed job in multiple high cost of living locations" (Seattle metro area, LA metro area to name a few).
The UK is poor because they decided to financialize the economy in the 90s and stop making things. It's like canada where the GDP per capita goes down every year. I'm amazed there hasn't been a revolution.
Sure, but expanding the definition of “Canadian” to include people who were already poor is a bit different from people who were already Canadian becoming poorer.
I don't think the average migrant salary is much different from the average UK citizen salary. Then again, I also don't find the "financialisation" argument very compelling. Plus, the GDP per capita visibly does not go down every year.
According to Stats Canada, “Real GDP per capita has now declined in five of the past six quarters”, so fair to say it’s currently declining. This was news to me.
"another respected data journalist, John Burn-Murdoch, calculated that without London, the UK would be poorer, in terms of GDP per capita, than even the poorest US state, Mississippi."
Yes, it's true that the UK economy is very London-centric, but the original poster was talking about the UK as a whole vs the US as a whole. (The flip side of this is that the figures would look better if you compared London to a major US city.)
None of this changes the fact that US software engineering salaries are a poor comparison to use to illustrate wealth disparities between the US and other countries, as they are an outlier.
Regardless, Americans are not five times richer than Brits by any reasonable measure. The salaries in the comparison upthread are outliers. The exact figure obviously depends on which stat you look at, but Americans are around 50% richer by most measures.
The U.S. engineer can be fired on a whim immediately and lose their health care (COBRA) and the company that fires them can even contest their unemployment benefits (that the employee paid into) if they feel motivated enough. That's one of the reasons they get paid much more.
I’ve been fired before by a major American tech company. I was underperforming, unmotivated, and depressed about it. They gave me a substantial severance payment in exchange for quitting voluntarily, and for signing an agreement that basically said I wouldn’t sue them. They let me pick my last date, they paid my health insurance through the next three months, and my manager told me I could use my last month of employment to find a new job. I was quickly hired into a better-paid position at another company, with a better manager, and I did well there.
I realize this story sounds absurd to anyone who hasn’t experienced it, but my understanding is that this form of firing (“managing out”) is basically the norm for low performers at top-tier tech companies.
To get actually fired, you usually have to fuck up big time, like sexually harassing a coworker, stealing trade secrets, or trying to start a union. (That last one is a joke, sort of)
Quite! A top 10% earner in Finland, a supposedly very developed country, by saving all of their net-income spending zero on food and letting their SO pay the bills, could in 2-3 years afford a new Skoda.
I don’t know if you’re making a joke or not, but getting NHS isn’t worth $200,000 USD per year.
Most Americans get employer-provided health insurance, which costs money (the amount specified in the DD section of the W2), and its often in the $1500/month range. That DD amount isn’t part of your income or the salary Glassdoor mentions. It’s an added benefit of top of that.
In the UK and elsewhere, around $500/month/person in taxes pays for your healthcare. That’s essentially subtracted from your income. So the uk income is even lower when you subtract the taxes the NHS costs.
> I don’t know if you’re making a joke or not, but getting NHS isn’t worth $200,000 USD per year.
Nope, but NHS + no/less student loans + no car dependency + cheaper childcare + time off + a ton of other things shave quite a bit off that $200k. Not equal, and not in every personal case, but a lot.
Isn’t housing extremely unaffordable in the UK though? That erases a lot of these benefits, doesn’t it? (I’m aware this is true of a lot of HCOL areas in the U.S. as well.)
Canadian and not UKian, but our public healthcare is definitely not worth 50% of my take home cash, I get much better access to care in the US right now. it still says Canada on my passport so I can get healthcare if I get fired or chronically ill
You say this ironically, but someone who’s been working hard 15 hours 7 days a week in a niche, 50 weeks a year, from age 15 to age 29 has clearly a much higher potential than a 45 year old following the normal path in life.
And almost certainly a higher employable value too unless they have catastrophically bad social skills…
Someone who works for 15 out of 18 of their waking hours, leaving 3 hours to eat, exercise, and have any semblance of social interactions or secondary interests, for FOURTEEN YEARS is not a genius. They are actually an idiot, wasting their life.
The implication was that someone who dedicated all of their time as physically possible to working and studying, would not have had time to develop social skills
>… been working hard 15 hours 7 days a week in a niche, 50 weeks a year, from age 15 to age 29…
Developed the same level of social skills as the average individual who lived a more normal schedule?
I have to ask before you even answer that. Do you believe that social skills are something to be practiced and built upon, are they some waste of time they only hormones bother with, or some other option I haven’t considered?
I think you might be delusional if you think that the people who can do all of this at the same time and don’t come out maladjusted to society is anything beyond a fraction of a fraction of a percent of outliers
This is just magical thinking on your end. I’ve met some of these “literal geniuses” making 500k at faangs and most of them are completely socially maladapted once you’ve taken them out of the pipeline they’ve lived in since high school to getting their first job mid or late 20s after their masters or PhD.
Secondly you started off this chain with talking about how someone working hard for 15 hours a day for decades is going to be more valuable and they’ll just be able to pick up every skill a human could have or need because they’re “geniuses”.
If they’re really geniuses why do they need to grind?
If you’re implying that they are only part of the set of geniuses that grind that long and there is another set of geniuses that didn’t, then how does that track with geniuses being a very small fraction of society?
> But the majority of them do exceed that very low bar, so it’s simply not that critical of a hinderance most of the time.
Describing not having catastrophically bad social skills as a “very low” bar is not a valid take when it comes to the world of computer science. I remember when visiting Carnegie Mellon as a senior in high school and evaluating their comp sci program, how the guides suddenly got very serious when they informed our parents(not the prospective students) that a course on hygiene was required freshman year and could not be waived. I’ve also worked with near limitless number of engineers who think they have the social skills down and then don’t understand why no one wants to work with them when they will do shit like call someone else’s project they’ve worked on for months pointless or useless in a group setting without even trying to approach said coworker with even a modicum of social awareness.
Those kinds of behaviors don’t show up in a population where having non catastrophically bad social skills is a “very low bar”
> You appear to be reading absolute implications into my comments, and/or inserting your own conjectures which aren’t there on a plain reading.
I think we’re coming at this with different axioms. You seem to believe that social skills are trivial and don’t matter next to the hard sciences that people grind away on. I am coming from one where I have to constantly make excuses or apologies for various people in software engineering or comp sci because they appear to be literally incapable of empathy or understanding that other people might have a different viewpoint than theirs.
Given my axiom I think your are handwaving away a lot, and that’s where you see my statements as inserted conjectures.
Realistically, there are plenty of competent people they could hire for any low six figures amount (unless they are directly in the DC area, in which case add 20-30k for cost of living). 500-600k or half of that is a unicorn salary that doesn't apply to those industries or areas and is irrelevant to the discussion. Even if they offered you that salary you wouldn't take it because the work environment would be radically different from working at a bloated web tech firm, or working at a silicon valley startup.
Totally different markets. You wouldn't be interested in that job, and they wouldn't want to hire you even if you were interested. Even the tone of your post makes that obvious.
We should be discussing early to mid career folks from somewhere other than silicon valley startup or big web tech land. Aka "meat and potatoes" tech jobs. That is what's being discussed.
I don't know what their problem is with hiring either and I agree with you that it could be partially compensation related. But not being able to compete with Silicon Valley on compensation is not where I would be going with that argument....I think it's more likely to be related to environment and interview style and notions of what "experience" means. In other words...bad hiring practices...not necessarily raw compensation issues. The compensation for non "big tech" firms can sometimes be quite good in comparison to other career paths especially when located outside of the valley, so being unable to hire talent makes me suspicious of hiring practices more than compensation (assuming they are reasonably large and hit market rate for the area and are in a reasonably large metro).
I am also disappointed, and I have not missed the context. The talk is empty for anyone who follows the field for more than two years, and especially for those who are familiar with his 2014 paper. Yes, he had amazing insight and intuition behind modern LLM breakthroughs, and yes, he probably earned the right to sound "prophetic", but he could have provided some interesting personal anecdotes about how the paper was written, or some fresh ideas in "What Comes Next" section of his presentation.
No, and they haven't been for at least half a year. Utterly optimized for by the providers. Nowadays if a model would be SotA for general use but not #1 on any of these benchmarks, I doubt they'd even release it.
I've started keeping an eye out for original brainteasers, just for that reason. GCHQ's Christmas puzzle just came out [1], and o1-pro got 6 out of 7 of them right. It took about 20 minutes in total.
I wasn't going to bother trying those because I was pretty sure it wouldn't get any of them, but decided to give it an easy one (#4) and was impressed at the CoT.
Meanwhile, Google's newest 2.0 Flash model went 0 for 7.
I didn't see a straightforward way to submit the final problem, because I used different contexts for each of the 7 subproblems.
Given the right prompt, though, I'm sure it could handle the 'find the corresponding letter from the landmarks to form an anagram' part. That's easier than most of the other problems.
You're saying the ultimate answer isn't 'PROTECTING THE UNITED KINGDOM'?
Can you please give an example of a “completely illogical statement” produced by o1 model? I suspect it would be easier to get an average human to produce an illogical statement.
Haha, you are right, I just asked Copilot, and it replied this:
> This is a classic riddle! The surgeon is actually the boy's mother. The riddle plays on the assumption that a surgeon is typically male, but in this case, the surgeon is the boy's mother.
> Did you enjoy this riddle? Do you have any more you'd like to share or solve?
Ha, good one! Claude gets it wrong too, except for apologizing and correcting itself when questioned:
"I was trying to find a clever twist that isn't actually there. The riddle appears to just be a straightforward statement - a father who is a surgeon saying he can't operate on his son"
More than being illogical, it seems that LLMs can be too hasty and too easily attracted by known patterns. People do the same.
It's amazing how great these canned apologies work at anthropomorphising LLMs. It wasn't really in haste, it simply failed because the nuance fell below noise in its training set data but you rectified it with your follow-up correction.
Well, first of all it failed twice: first it spat out the canned riddle answer, then once I asked it to "double check" it said "sorry, I was wrong: the surgeon IS the boy's father, so there must be a second surgeon..."
Then the follow up correction did have the effect of making it look harder at the question. It actually wrote:
"Let me look at EXACTLY what's given" (with the all caps).
It's not very different from a person that decides to focus harder on a problem once it was fooled by it a couple of times already because it is trickier than it seems. So yes, surprisingly human, with all its flaws.
But thing is it wasn't trickier than it seemed. It was simply an outlier entry, like the flipped tortoise question that tripped the android in the Bladerunner interrogation scene. It was not able to think harder without your input.
Grok gives this as an excuse for answering "The surgeon is the boy's mother." :
<<Because the surgeon, who is the boy's father, says, "I can't operate on this boy, he's my son!" This indicates that there is another parent involved who is also a surgeon. Given that the statement specifies the boy's father cannot operate, the other surgeon must be the boy's mother.>>
Sounds plausible and on the first read, almost logical.
Easy, from my recent chat with o1:
(Asked about left null space)
‘’’
these are the vectors that when viewed as linear functionals, annihilate every column of A . <…> Another way to view it: these are the vectors orthogonal to the row space.
‘’’
It’s quite obvious that vectors that “annihilate the columns” would be orthogonal to the column space not the row space.
I don’t know if you think o1 is magic. It still hallucinates, just less often and less obvious.
Sure average humans don’t do that, but this is hackernews where it’s completely normal for commenters to confidently answer questions and opine on topics they know absolutely nothing about.
"On two occasions I have been asked, 'Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?' I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question."
So the result might not necessarily be bad, it's just that the machine _can_ detect that you entered the wrong figures! By the way, the answer is 7.
> Can you please give an example of a “completely illogical statement” produced by o1 model? I suspect it would be easier to get an average human to produce an illogical statement.
Following the trail as you did originally: you do not hire "ordinary humans", you hire "good ones for the job"; going for a "cost competitive" bargain can be suicidal in private enterprise and criminal in public ones.
Sticking instead to the core matter: the architecture is faulty, unsatisfactory by design, and must be fixed. We are playing with the partials of research and getting some results, even some useful tools, but the idea that this is not the real thing must be clear - also since this two years plus old boom brought another horribly ugly cultural degradation ("spitting out prejudice as normal").
> For simple tasks where we would alternatively hire only ordinary humans AIs have similar error rates.
Yes if a task requires deep expertise or great care the AI is a bad choice. But lots of tasks don't. And in those kinds of tasks even ordinary humans are already too expensive to be economically viable
Do you have good examples of tasks in which dubious verbal prompt could be an acceptable outcome?
By the way, I noticed:
> AI
Do not confuse LLMs with general AI. Notably, general AI was also implemented in system where critical failures would be intolerable - i.e., made to be reliable, or part of a finally reliable process.