The objection seems to be based on the falasy that technological progress equals social or political "progress". Why on earth would we expect AI descision making to display a lack of prejudice when human decision making is suffused with it.
The only people who expect technology to act like a benevolent god are the ones who have replaced their god with it. All technological progress does is to increase the power and influence of human beings. The progress the writer seems to want is socio- political, not technological.
Particularly when we consider what we mean by prejudice, which is presumably something like Making a decision on grounds which we deem it important to ignore. This is a very complex concept. It's a function of society, and changes with society. It's not something with a rigorous definition.
Obvious example: reasonable modern people know it's indefensible to make an engineering hiring decision on the grounds of ethnicity, regardless of whether there are any correlations associated with ethnicity. This is even enshrined in law in many countries.
To make a decision on the grounds of someone's qualifications and job experience, however, does not count as prejudice.
We should expect a machine learning system to act as a correlation-seeker (that is after all what it is designed to do), without a nuanced understanding of what prejudice means.
We've seen this issue crop up in the context of an AI having a say in parole decisions.  also relevant discussion at 
One likely outcome is that increased usage of ML will lay bare more gray-area correlations that force us to confront broader systematic prejudice in an uncomfortable way. Now whether we make material changes based on these discoveries or whether we leverage our significant ability for post-hoc rationalization to paper over the cognitive dissonance is another question entirely.
Or, perhaps we discover that correlations don't necessarily point to systematic prejudice, and continue to argue over which correlations are more indicative of prejudice than others.
Some point at the proportion of women in top software engineering job and claim prejudice in hiring. Others point at the hiring pool for top software engineering job and disagree. And around and around we go. Machine learning won't solve that.
We should expect machine learning systems to dismiss accidental indirect correlation for the benefit of the variables that are directly correlated. There are plenty of algorithms that achieve that, it's only that gradient descent doesn't.
The fact that our AIs are becoming biased is a bug. It should be fixed.
In many cases, it's the data that is biased. In that case it's impossible to differentiate between bias that the AI should learn and bias that the AI shouldn't learn.
Let's assume we have a database of incidents where a person was found to have cannabis. This database has the following data items: a timestamp, the persons name, the persons ethnicity and the amount of cannabis that was found.
Now assume further that black and white people have the same base rate of cannabis use (which according to the studies I found seems to be the case).
The last thing we have to assume in this case is that this database was created by racist policemen who arrest more black people for cannabis consumption.
An AI trained using this data would assume a higher base rate of cannabis consumption by black people. It's impossible for this AI to differentiate between correlations it should learn (for example that people who used cannabis multiple times and were found to have much of it are worth looking at) and (untrue) correlations that it shouldn't learn (that black people have a higher base rate).
The correct solution here is to use a dataset that is not biased, but it's hard to tell whether a dataset is biased.
You can add such a prior by adding a term to the loss function that penalizes any difference between the way the groups are treated. The math isn't hard, only the political decision of what is protected and what isn't.
I don't see why we should assume that this would reflect reality.
If a law has racist roots, i.e. if it is written to target a particular ethnicity, then we should expect that certain ethnicities really do break that law more than others.
I get what you're getting at. If the data collection is biased, massage the actor on the data to compensate. Got it.
...but if you're uncomfortable with the consequences of unleashing a Machine Learned system trained on your data as part of your decision making process from the get go, why in the heck would you want to anyway? There is no benefit to doing so when you realize that the system only works as well and unbiased as the sum of the biases of the data architects, data populators, and data interpreters combined, and that extra loss factor you added is, in fact, another bias. An unacceptable on to the minds of many no matter how disagreeable one may personally find it.
Better to just leave it in human hands to deal with on a case by case basis by that point. At least by doing that we stay ultimately in control of the overall decision making process.
People being in the loop is not a bug. Trying to solve social/political problems with poorly thought out applications of inexplicable technological systems is.
This is a very good point, but it's not enough to ensure your input data aren't reflecting existing biases in society. (I believe it's necessary, but not sufficient.)
In my other comment I gave the example of maternity leave. A woman who becomes pregnant won't be as productive that year. It's no-one's 'fault' that this is the case, and it doesn't reflect anyone's bias. It's still important to ensure, when making hiring decisions, that applicants are not eliminated on the grounds of 'high pregnancy risk'.
Let's consider an AI that weighs in on hiring decisions. Let's consider what it might make of maternity.
Decades ago, it wasn't unusual for a prospective employer to ask a female applicant whether she was planning on getting pregnant. Society decided that wasn't acceptable, and made the practice unlawful (as well as forcing employers to offer paid maternity leave), despite that, from a profit-seeking perspective, that is valuable information for the prospective employer to have.
Let's suppose an AI has been trained on a dataset of previous hiring decisions, and some measure of how much value those employees went on to deliver. Let's suppose it's also told the age and sex of the applicants. It might notice that female applicants within a certain age range, have a tendency to have a long period of zero productivity, some time after being hired. Having noticed this correlation, it would then discriminate against applicants matching that profile.
But wait!, I hear you think, Why did we provide it data on sex and age? Remove those data-points and the problem goes away. Not so. There could be non-obvious correlations acting as proxies for that information. The system might detect these correlations.
In a parole-board context, suspects' names and addresses could be proxies for ethnicity, which could correlate with recidivism rates. The 'correct' behaviour for a correlation-seeking AI is to detect those correlations, and then start to systematically discriminate against individuals of certain ethnicities, even when their crime and personal conduct is identical to another individual of a different ethnicity.
Back to our maternity leave example, then. It's an interesting exercise to come up with possible correlations. The ones that occur to me:
* Given names (if provided to the AI) generally tell you a person's sex
* Given names also fall in and out of fashion over the decades, so there's a correlation with age
* The topics a person chose to study at university, can correlate with the person's sex, as well as their age, as some fields didn't exist 40 years ago
* The culture a person is from, may impact the statistically expected number of children for them to have, and at what age. Various things could act as proxies for this.
humans may also act on the same information. The fact that you are less likely to be called if you have an arabic sounding name in europe is well documented.
at least with neural networks we can remove those variables.
I guess another problem could be that our current distribution is biased. so the system might give an advantage to new-grads from unviersity X becausea lot of old employees are from it. which is inherent disadvantage from people from tradtionally-black university Y.
but hiring based on alma mater is already a common practice.
Anything that doesn't get stuck on local maxima do.
> How can you tell a direct from an indirect correlation in data?
There is another variable, with stronger correlation, that explains the same change.
The problem for AI is these proxies can seem to be very useful. Want to know how fast a vehicle is, well price seems like a really good approximation when you start. Except, the more information you have the more likely price is going to fool you.
It is patently an act of incredible destruction to increasingly, globally and inescapably “control” important domains of human lives, relationships and society with extremely superficial centralized mathematical models of what humanity is.
There will be unintended consequences. They will be horrifying, and we may not even grasp what we've lost.
According to the article, that is how AI decision making is presented by "technocrats everywhere". For instance:
Empiricism-washing is the top ideological dirty trick of technocrats everywhere: they assert that the data “doesn’t lie,” and thus all policy prescriptions based on data can be divorced from “politics” and relegated to the realm of “evidence.”
The article is presenting, and supporting, the opposite opinion to that of "technocrats everywhere".
The ultimate goal of teaching, however, is teaching how we arrived at a given behavior. And amongst the most prominent themes in this is dealing with new situations and adapting to change. Teaching is not about imitation, it's about transcending the example.
And, from what I remember from my control theory classes, the integral part of a controller introduces lag, inertia, generally making the output more resistant to input changes.
(Also note that the "non-conservative" Proportional and Derivative components in a PID by definition don't learn - they react to input and its change.)
Our 20 year old millennium has seen a tremendous rate of change in both technology and social norms. Perhaps politics has some similar inertial dynamics.
See "The Cult of Information" by Theodore Roszak.
If you fitted a decision tree to predict recidivism risk, it would be extremely easy to interpret. But if black men are rearrested more often in your dataset, then black men will likely have a higher predicted risk on average—no matter the causes of that feature of your dataset.
I don't get to stand there and blame the model when I'm the one doing the slapping.
Edit: the auto correct had written "data doesn't like"
We've picked a metric, recidivism rate, that is believed to be inherently biased because cops arrest a lot of protected minorities. The model has correctly predicted that cops will arrest a lot of protected minorities. The general public has then turned around and shot the messenger rather than hold cops accountable for all that arresting they're doing.
Ironically, I think “conservative” is both an excellent descriptor of function approximators—they tend to conserve whatever bias they’re provided with—and a terrible word to use for popular writing, since it’s so easily confused with political conservatism (even though e.g. no politically conservative “AI” would autosuggest “on my face” as a completion of “can you sit”).
It, as the current top comment mentions, is a philosophical rather than technical argument.
It doesn't matter if all machine learning can do is describe the current state of things of it is presumed by those interpreting the output to provide insight into some great and universal truth.
Most science degree programs have a philosophy of science course as a degree requirement. As an engineering faculty member I have found huge resistance to and equivalent course becoming part of engineering or computer science degrees. Those degrees, instead, take an apparoch to scientific/technical knowledge that is less science and more religion.
The comments here are a reminder why such courses have a place.
Please read this thread https://twitter.com/yonatanzunger/status/975545527973462016 "computer science is a field which hasn't yet encountered consequences."
Unfortunately we're still on the rise of the current hype cycle around Machine Learning, and prescriptive solutions are precisely what nearly every company who use the phrase "machine learning" are promising.
It's a bit like dictionaries: they "learn" the current meaning of words. It gets annoying though when everyone starts running around policing the meaning of words based heavily on that dictionary: it would conservatively restrict the natural evolution of word meanings, when we know meaning of words do change over time, and sometimes should change over time.
Every time a random number is involved in the training process (stochastic selection of mini batches, noise injection, epsilon-greedy action selection), a form of exploration is taking place, finding novel paths of solving the problem at hand. The mixing of noise with data is paradoxically useful, even necessary in learning.
This only applies to reinforcement learning (and similar). Classification problems (and similar) are fundamentally conservative, because you feed them with a finite set of preexisting training data that they must extrapolate from.
The same applies to humans, no?
Some people are working on it : https://news.ycombinator.com/item?id=21728776 (dec 2019)
What I want to point out is that nothing he says should be attributed specifically to "machine learning". Machine learning is a set of techniques to make inferences from data automatically, but there is no implicit restriction on what the inferences should be. So machine learning is not "conservative" — almost all popular applications of it are. There is no inherent reason why an ML-algorithm should suggest the most similar videos to the ones you watched recently. The same way you can use (say) association learning to find most common item sets, you can use it to find the least common item sets with the given item, and recommend them instead. Or anything in between. But application designers usually choose the less creative option to recommend (understandably so) stuff similar to what you already got.
Sometimes it's ok: if the most popular thing to ask somebody to sit on nowadays is "my face" it's only logical to advice that, I see nothing wrong with this application. But many internet shops indeed could benefit from considering what a user has already bought (from this very shop), because it isn't likely he will want to buy a second similar, but different TV anytime soon. Or, when recommending a movie, you could try to optimize for something different than "the most popular stuff watched by people that watched a lot of stuff you did watch" — which is a "safe" thing to recommend, but at the same time not really interesting. Of course, finding another sensible approach is a lot of work, but it doesn't mean there isn't one: maybe it could be "a movie with unusually high score given by somebody, who also highly rated a number of movies you rated higher than the average".
This stationary distribution is the source of the conservativeness and central to algorithms that we call "machine learning". ML always tries to replicate the distribution when it makes infererences, so it is fundamentally biaed against changes that deviate from that distribution
The Future and the Past are structurally the same thing in these models! They are "missing" but re-creatable links.
AI is the broader term, but in pop culture AI and ML are very much synonymous.
I see why you assume that "stationary distribution is the source of the conservativeness", so maybe I should clarify this moment. It is kind of true in the most general sense: sure, when querying the stationary distribution we can only ever ask how things are in a timeless universe. How anything new can be obtained this way? The problem is, that if we are this general, then the word "conservativeness" loses any meaning, since everything in the [deterministic] Universe can be framed as a question of "how things are", everything is conservative, nothing new can be obtained anywhere, ever.
And we don't even need to get this general for the word "conservativeness" to lose practical sense. When you ask another human for an advice, all he ever does is, in essence, pattern recognition and querying his internal database of "how things generally are to the best of his knowledge". Yet you don't call every human advice ever "conservative": only the kind of advice that recommends the safest, most popular thing, thing that everybody likes, pretty much ignoring the nuance of your personal taste. In fact, even then, you call it "conservative" only if you can notice that, which means that the recommendation isn't new for you personally (and by that criteria alone most humans would lose to a very simple, absolutely currently implementable music recommendation algorithm, since they probably know much lesser number of "artists similar to what you like" than Spotify knows: the only thing Spotify has to do to win is not to recommend the most popular choice almost every time).
One more thing. I could probably convey to you that assuming "ML = conservativeness" is wrong much faster by invoking the example of reinforcement learning, since it is sort of intuitive: there is the obvious existence of "time" in that, you can imagine a "dialogue" where it adapts to what user wants using his feedback, etc. It is easy to see how it could behave "not conservatively". I intentionally avoided doing that, since it can lead to the false impression that RL is somehow different and less conservative than other algorithms. On the contrary, the point I'm trying to make is that every algorithm, even the purest form of memorizing the stationary distribution (like Naïve Bayes) is not inherently conservative. It all depends on what exactly you ask (i.e. how you represent inputs and outputs) and how you query the distribution you have (e.g. how much variability you allow, but not only that).
So, when you see the application that uses ML algorithm and behaves "conservatively", it isn't because of the ML algorithm, it is because of the application: it asks the ML algorithm wrong questions.
>Nor is machine learning likely to produce a reliable method of inferring intention: it’s a bedrock of anthropology that intention is unknowable without dialogue. As Cliff Geertz points out in his seminal 1973 essay, “Thick Description,” you cannot distinguish a “wink” (which means something) from a “twitch” (a meaningless reflex) without asking the person you’re observing which one it was.
Was this mathematically proven? It's definitely an interesting statement, since a lot of "AI" systems try to predict intention and do a piss poor job of it, but to quote that the anthropological ancestors have proclaimed for eternity that a computer can never know even the slightest fraction of intention from just observation seems hypocritical.
It is futile to talk about 100% correctness in this field
> ... machine which hasn't even the slightest semblance of a shared cultural context ...
Why is this universally true?
> ... it would likely be trained on a small subset of human cultures and would produce varying results.
As it has when you travel to foreign countries.
There is not such hing as understanding artificial intelligence. There is a task of understanding _intelligence_ and attempting to implement it without giving birth.
And don't get me started on supposedly universal facial expressions like smiles. Everytime I see a US citizen in a posed picture, I understand that they are supposed to be smiling, because it's a portrait etc, but what I perceive is an aggressive and half-mad rictus. [Edit: so as not to make this a thing about US people: British smiles I perceive as fake; Balkan smiles, as lewd and provocative; smiles of people from the subcontinent as forced and subservient; the only people whose smiles I recognise as smiles are Mediterranneans and Middle-Eastern folk, who smile like the people back home. But I always remind myself that the way people smile, or generally arrange their face is just how they have learned to arrange their face to express certain feelings, which is a different way than I have learned to arrange my face to express the same feelings. I probably look weird to them, too. Actually- because I smile with my eyes more than my mouth, I've been told in the past that I'm rude because I "don't smile". Wot! But I...]
There's no use assuming cultural consistency of winks, nods, facial expressions and gestures. A wink is a wink, except when it isn't and a twich is a twitch except when it isn't.
So, yeah, the only way to know what someone means by a gesture or a facial expression is to ask. And even then you can't be sure.
They are pointing out that the system under observation, the system of interest (a person, a conversation, a society) is extremely complex and when it emits a signal it is not possible in the general case to know what that signal “means” without some understanding of the system’s state and processes.
We are not capable of directly interrogating, representing or computing over the true state and processes of these systems, so we must use heuristics. Humans develop “theory of mind”, the cognitive ability to model other people's cognitive state and process, including their human values and motivations. People with a better theory of mind, which makes more accurate, more fine-grained inferences in a wider variety of contexts, have a decisive social competitive advantage.
This doesn't even touch on “theory of relationship”... when you take a system of humans, what are their shared human values? How are value conflicts negotiated in groups at every point of scale? What is the current state of the negotiated compromises?
A dataset and a machine learning algorithm is an observer, inferrer and predictor of human and social systems whose actual state and processes are irreducibly large and invisible, and whose human values (and the domain of possible human values) are mysterious.
Even if we gloss over the symbol grounding problem, it is clear that machine learning systems cannot be expected to operate in a way that is aligned with the human values of individual or groups.
In the worst case they may operate to extinguish those values, and as an unintended consequence hugely diminish quality of life.
And maybe we won't notice, because surely there is an “Overton window” of imaginable possible lives, and ML has the potential to slowly and insidiously shift that window until “being happy and fulfilled” becomes “being in the centre-right of the main distribution” across the major ML models in society.
And yes, I do realize that similar normalizing forces have been present in society forever, operating with similar effect. (And I recognize their value in reducing transaction costs of all kinds). But to date they have been based on human-driven mechanisms which lacked the ability to centrally micro-observe and micromanage behaviour on a global scale. Alternative value systems could evade these pressures and develop, compete and co-exist, at every point of social scale.
But with the advent of the global Internet and sufficient computational throughput, ML has the potential to micro-observe and micromanage everyone everywhere. It can know more about us than we know about ourselves, while entirely missing the point of being human.
That statement was easy to refute mathematically, but I don't know about Geertz. That seems more vague.
How would you prove that "mathematically"? Do you mean "scientifically proven"?
So, how does this relate to your question? The point is that predictive policing is solving the wrong problem. What's needed are not more accurate neural nets predicting crime, but techniques for addressing the underly sociological factors that cause crime.
Taking a step back and speaking broadly, Cory's point is that the focus on data and quantitative analysis is causing problems in two ways: (i) people are using quantitative methods to solve the wrong problems, and (ii) they seem to be oblivious to (and in some cases actively hostile to acknowledging) the harms being perpetrated by their methods. Both of these problems seem to be driven by a lack of understanding of well understood (but non-quantitative) social science literature.
Corollary - Big Data is political, but it can hide behind the pretence of objectivity.
This shouldn't surprise anyone, but for some reason it does - perhaps because we're used to thinking of AI as the automation of the scientific method, when in fact it's just as likely to be the automation of all kinds of other things, some of which are nasty, stupid, and wrong.
A hypothetical smarter-than-human AGI may be able to talk back and say "You're being stupid because you're not modelling the correct problem, and so your attempted solution is wrong."
But that assumes hypothetical smarter-than-human AGI is also more-ethical-than-human - which unfortunately might be a bit of a reach.
Bottom line - the limits of AI are set by the limits of human political intelligence. And since most human political systems operate at pre-scientific policy levels - the exception being the abuse of social and personal psychology to gain power - this isn't encouraging.
Slightly off context but I think that the solution to this is some kind of quota and or tier system to laws and policing. Basically to restrict the amount of policing that may be done for things like jaywalking and traffic violations while the violent crime rate is above a certain threshold.
I have been held at gun point twice in my life as part of armed robbery and hijacking and neither time did I even consider for a moment that the police would find the people who did it. But if I do 60 km/h in a 40 km/h zone they will follow me to the end of the earth to get me to pay them.
From an academic perspective, it is true that there is some debate about the efficacy of broken windows policing, but even the most supportive academic studies find only a small correlation between violations of "order" (like jaywalking and graffiti) more serious crimes. There's just isn't any evidence at all that the way to reduce serious crimes is by going after jaywalkers.
Here are three different examples that would meet your description, with very different answers:
1) The X community has exactly the same, or lower, crime rates than the entire nation. However, nationwide anti-X sentiment means that despite this, most convicted criminals are from the X community. This makes it look like the regions where the X community live are high-crime regions.
This is “accidentally racist”. Researchers know about this problem.
2) The X community is more prone to crime, and as this is purely an example, it just is and there’s no need to justify that.
Extra police in this scenario is not racist, though I suspect anyone who jumps right in and assumes it to be true about the world might well be racist themselves.
3) A confounding variable, such as income or education, means that members of the X community are more likely to commit crimes than the general population, albeit it at the same rate as the equivalent income/education/confounding sunset of the general population.
In this case, while it would not be racist to send in more police, it would totally be racist if politicians applied unequal effort to solve the confounding variable within the general population as compared specifically to group X.
Stopping cars just because of the skin color of the driver is not "perceived as racist", it is racist, since you are treating people worse due to their race.
Regarding ML based predictive policing specifically, it seems like a spectacularly bad (but efficient and statistically effective!) idea to naively apply a machine learning based classification approach to such a problem.
However, I would refine the claim "Machine learning is fundamentally conservative" to say "Reducing a distribution of predictions to only the most likely prediction is conservative."
For a product/content recommendation you could include the product that's maximally different, the second most similar, you could exclude products within the same category, you could choose randomly and weight by conversion rate. We need more creativity and experimentation here I believe.
This is interesting. Has anyone tried that or knows of a website that uses this?
The top-n removal was a small but noticeable amount better. That was in part due to the specific content in the database as well as the audience, but it's a thing to try I believe.
Another way: think of common human categorizations. Opposites are usually defined in terms of these categories.
Yes, I agree, but this does not have the be the final word.
There is a lot of room for design. For example, an autosuggestion tool can trade off between exploration (randomized recommendations) and exploitation (pure collaborative filtering).
More creativity, and ability to escape local minima, but at some cost when dealing with 'typical' cases and when making particularly damaging mispredictions.
ML Algorithm: yes!!
I mean they could have all simultaneously gone insane, or be infested by alien mind parasites, but going bungee jumping is the much more likely reason.
It's not always that a XKCD is such a good answer.
For example, people throw around the low labor productivity stat all the time to prove that no automation is happening, not realizing that GDP is the numerator of that stat. Well, GDP is a pretty terrible gauge of who is benefitting from technological innovation, as it is distributed incredibly unevenly, probably more so than income even. The problem with automation isn’t that it’s not happening, it’s that only a very small number of people are capturing the wealth that it is generating. Also, it is generating a small number of professional jobs, but mostly the jobs it is generating suck (Uber driver, AMZN warehouse worker, etc).
In other words, a worker is richer if they can buy more and better stuff even without any rise in their wage.
The fact that everyone can afford a smartphone, including those with jobs that “suck”, does not reconcile with “only a very small number of people are capturing the wealth that it is generating.“
Yes technology has made cool gadgets for consumption, but that isn't what wealth is about. From Wikipedia: "Wealth is the abundance of valuable financial assets or physical possessions which can be converted into a form that can be used for transactions."
The average worker hasn't seen a real pay ride in 40 years, and their ability to build wealth through savings has been greatly diminished.
Between 1979 and 2011, gross median household income, adjusted for inflation, rose from $59,400 to $75,200, or 26.5%. This compares with the Census' growth of 10%. However, once adjusted for household size and looking at taxes from an after-tax perspective, real median household income grew 46%, representing significant growth.
Instead of looking at nominal income (which was a smaller part of peoples’ wages decades ago), look at real wage growth (https://www.pewresearch.org/fact-tank/2018/08/07/for-most-us...). If we take in all the benefit cuts, the fact that college has become a necessity, pensions are dead, and healthcare inflation is insane, real housing inflation has been substantial, then you can begin to understand the situation. Income is only part of the equation. Yes, income has gone up, but we traded it for the mother of all benefit cuts.
> The average worker hasn't seen a real pay [rise] in 40 years
> So basically, improvements in productivity are only benefitting people in areas where productivity has improved?
Worker productivity has increased enormously in the last decades.
40% of Australia's population live in either Sydney or Melbourne, and these two cities are among the top 5 least affordable in the world .
1 - https://www.afr.com/companies/financial-services/australian-...
- There is a ballpark estimate that sum of all physical money in the world is 90.4$ trillion, let's distribute it equally to all 7.53 billion people: we get 12k$ per person. This is not enough to buy an apartment in most parts of the world.
- There is an estimate that 1% has 303.9$ billion, let's distribute it equally to US population of 327.2 million people: we get 928$ per person. Barely enough to buy status smartphone.
So I don't see how wealth distribution can solve anything. Even if we equalize wealth and prices at a same time, free market will quickly spiral down to similar pricing that we have today (things are barely purchasable, or not at all).
Could you explain how you got to that conclusion? It does not seem logical to me at all.
Current prices are in part a result of high inequality: It makes economic sense to sell a house for 10x the median income because they are still groups of people so high above the median that they can afford to pay that price - and of course as a business you'd very much like to sell to those people if in any way possible.
In a society with more equal income, the highest price for which you could find a buyer would likely be a lot closer to the median.
I wouldn't even be sad about the money I'd no longer be making, because right now a good chunk is going into rent, and an even larger chunk is going into savings so I can buy a home at NYC prices. So it's not like I get to spend it on other things. If I didn't have the money but I also didn't have to spend that money on housing, I'd be equally well off.
(And all of this is leaving aside that other people would be better off, which is net good for society, etc. etc. Even the first-order effects alone are selfishly valuable.)
-  https://www.businessinsider.com/the-1-percent-dont-know-what...
The world could make do with very little money if it could move faster. Quantity of money times velocity of money times unit of time gives you the total monetary transactions over a time period; if you increase the velocity, then you can decrease the quantity, and nothing else would change.
However, the total value of the S&P 500 (which is probably about half the world stock markets value) is about $26 trillion, so that is about $3400 per person or maybe double or triple for a family. And the dividend of 1.75% means an income of $60/year if you hold on to the stock.
So I think doing this redistribution would result in everybody to whom $3400 is a lot immediately selling (assuming they were allowed to). I'm not sure what the distribution of capital would end up being. Someone would buy, say "Barren Wuffet" who fortunately had all their investments in cash, bonds, and small cap stocks. So ownership would end up being concentrated again, but maybe not as concentrated.
Anyone who has a lot of money saved for retirement in a large-cap fund would be unhappy, but why should you be better off than an average citizen of the world?
It seems like a fine experiment to me.
Edit: but just to spell it out, obviously politically impossible given the consequences for first world people who are relatively privileged but don't think of themselves that way. The median IRA balance (note, not the average that is skewed by the rich) is somewhere around $20-30K, so redistribution would affect those people like a 90% market crash.
If your boss said you're not getting a pay rise this year but that's OK because you'll still be able to buy more would you accept that as a fair and reasonable argument?
Don't use it as an argument to keep poor people's wages low then.
I don't see everyday life improving several billion times from 1970 even though the number of switches I control in my laptop are several billion times (at least) larger than the number of switches that existed in the world in 1970.
Only that they are fully captured by a small class. It says nothing about the relative amount, and doesn't change anything about the wealth disparity picture, for those smartphones are cheaper for everyone but one group has a lot more to spend AND smartphones are cheaper for them
Because wealth is defined by claims on other people's productivity, not by the ownership of one particular consumer good.
If you'd said "Everyone now gets a significant proportion of their income from investments and uses some of it to buy a smartphone" you might have had a point.
If you try to look at how much much of societies gains go to which wealth bracket and how much they actually do for it, this gap widened considerably during the last century and has become a real threat to these western values of liberty and democracy, especially because the rules of civilization and democracy exclude those on top.
One doesn’t really have to be a marxist in order to want to go back to a resonable wealth/wage gap. One doesn’t have to be a marxist to demand the rich should pay their taxes and follow the rules. I’d argue that every conservative in the very sense of the word should stand behind this as well.
Many cannot afford these smart phones and gadgets but will take loans or buy them with plans that on the surface looks cheap.
Loans from pretty much fully automated ”banks” working on-line only. Send an sms and get $500 to buy that phone!
Further increasing the gap in the process...
In the 1990s and early 00s, automation happened mostly in manufacturing. The low productivity growth stats that stupid (mainstream) "economists" like Krugman cite to argue against people like Yang or Musk measure all labor. But manufacturing makes up less than 1/5 of the US economy. If you look at the labor productivity in the manufacturing sector alone, it's congruent with the automation argument .
Productivity numbers are lagging because investments in new tech need time to bear fruit. Automation of the service sector has already started and will show in the numbers soon enough. Given that the service sector is four times the size of manufacturing in the US, it will be even better for GDP and even worse for people if again nothing is done and people keep listening to those intellectual frauds who call themselves "economists."
Clearly, if the latter would happen, and we are on that path right now, we are bound for a distopian future.
Old economic models don't work anymore, we need radical changes. If we don't plan them, they will come in the form of either revolution, or brutal repression (ala China).
In the limit the only beneficiaries are shareholders, who now own all the wealth and capture all of the productivity.
You now have a choice between a neofeudal dystopia where a small percentage of the population lives like emperors while everyone else is relegated to a shanty economy where they fight over scraps, or you stop the nonsense and set up a humane economy where all productivity is redistributed and humans automatically have economic freedom as a birthright without having to work for it or inherit it.
Currently we're on our way to the first option. But the second is the only viable long-term outcome - because in the limit the wealthy will continue to play status games with wealth, and the circles of distribution will become smaller and smaller until eventually literally everything is owned by a small handful of individuals and - perhaps - their immediate families.
That's been the case for most of human history. I don't see why this isn't a viable option again. High technology does not mean a humane society. For the majority of the last 10,000 it was pretty much the reverse. Egalitarianism meant primitive, starvation and deprivation meant wealth.
Of course one only need read what happened to monarchs for most of the period. Being king is good, less so when you become furniture for the next king.
Also, you're wrong. Egalitarianism has meant a lack of luxury relative to what was possible at the time, but a better baseline for the population as a whole.
>Egalitarianism has meant a lack of luxury relative to what was possible at the time, but a better baseline for the population as a whole.
The population of China was larger than any of the time and the majority lived in poverty so great the Europeans who visited were shocked by it.
However, I do wish we would not use the word "conservative" as an epithet. I think it's quite likely that "conservative" is exactly what we should be looking for in algorithms and prediction engines.
The fact that their properties are misused and infused with reactionary (not necessarily conservative) biases by humans should not make us attach morally negative properties to being conservative.
FWIW, my conservatism leads me to be suspicious of employing these tools in the first place ...
I don’t understand what he’s arguing for, exactly. That we should have AGI all of the sudden to understand intent, and detangle causation vs correlation? I don’t think anyone in the machine learning community would argue against that, but the question is how.
What’s new here?
Cory is trying to inform people that AI and ML aren't a magical solution to all our wasteful systems because in some cases we need to throw away the system and start again, and AI can't tell us that. It can only tell us the best way to run the current system. That's what he means when he says AI is conservative.
It's not surprising at all if you understand AI but most non-tech people don't.
And he points to three other sources right in the first paragraph.
Furthermore, there is a psychological element at play around mirroring back your own intelligence at you (see Eliza / Rogerian Psychotherapy) in a way that will lead you to new thought.
I see this and other examples of "explainability" from time to time as proof why humans are not a "Black Box."
However it rests on two faulty assumptions.
1. The explanation will be truthful
2. The explainer can always reliably describe the actual cause of their own actions
For the purposes of theory, you could explain away #1. However a minute of introspection will make you realize that you would fail at #2 the vast majority of the time - using story telling and retroactive explanations to explain your behavior.
> This is what makes machine learning so toxic.
I'm saddened to learn about our toxic contributions to society and can't wait to hear about alternative mind-reading approaches for fridge recommendation in the next article.
Words like 'conservative' and 'toxic' are misleading because they imply that there are better alternatives that are not being chosen. Far better are the terms by commenters here, 'descriptive' and so on. That the article is not written for machine learning practitioners makes it even more misleading.
It could be any conclusion, not just those conservatives dislike. Using myself as a target, what if we eventually have enough sampling data to show that people of Irish extraction are more prone to alcoholism, and people of Scottish extraction tend to be statistically more thrifty? (This suggests that I would be a cheap drunk). How will we cope with that?
It is true, ML is prone to some black-boxiness, but it could be any statistical extraction. We might very well use other methods besides ML to show the correlates once suggested.
Will we simply put in a hard override to get the answers we want to get, the answers we find comfortable? History shows we have seen whole governments subscribe to this idea before. Ignore the results, publish what pleases.
I've no easy answers here, but my guess is that history will repeat itself.
Yep. That's already happening. See Google's papers on "debiasing AI" (if you look at their example goals, they mean biasing it away from what it learned and towards what Googlers wish was true).
What makes you think it hasn't already happened? Numerous times?
Biological differences between sexes? IQ differences between differentiable subsets of people? The gender equality paradox?
We already know what will happen when findings are unpalatable. We'll think of ways to explain why they don't matter and can't possibly be right, or that we weren't measuring the right thing in the first place. And these are just the more obvious cases!
Which has been true, actually. For example, human facial recognition systems that can't properly distinguish dark-skinned faces. What is your conclusion in those cases? That dark-skinned people don't have faces? All look the same? Aren't human? Or is it more likely that the system is flawed in some way? Is that flaw in line with existing biases? What then is the chance that the flaw is BECAUSE of our existing biases, in some way or fashion?
From what I've read on HN, many people involved in developing ML technology seem to be overly concerned with their systems spitting out "uncomfortable truths", and less concerned with flaws in their data capture, training, or system design processes. We're putting the cart before the horse, and the giddiness with which it's being done is troubling in and of itself.
It may be more relevant in developing ML technology because readers here are the people who have to (at some point) decide whether they've found an uncomfortable truth or falsehood. Or that it's potentially biased, but may or may not be relevant. If I'm doing this at work, I may need to be able to explain to my boss whether or not what we're doing is going to be called racist or sexist, and that's regardless of whether what's being done actually is racist or sexist.
It's easy if you're using ML to detect whether an image contains a bird or train. And it's easy when you point to an obvious example with bad data (e.g. inability to differentiate dark faces). But that's not all cases. Sometimes it's good data and the findings are unpalatable. That was the root of the question.
I agree that it's important to be aware of the potential for these sorts of findings, but I continue to disagree that our takeaway should be an impetus to gird ourselves for the backlash when The Machine spits out, "Black people are dumb," rather than an impetus to observe the bedrock notion of scientific inquiry (you know, skepticism).
>It's easy if you're using ML to detect whether an image contains a bird or train. And it's easy when you point to an obvious example with bad data (e.g. inability to differentiate dark faces). But that's not all cases. Sometimes it's good data and the findings are unpalatable. That was the root of the question.
I think the "Newspaper Error Rule" (or, I guess, the "Cockroach Rule") applies here: if you recognize one when you're looking, how many are you missing when you're not, particularly when you're in less familiar territory? The less obvious examples are the ones we should be more wary of. And if your boss is expecting a quick, cut-and-dry answer to questions that have vexed society for centuries, just because we've applied a sophisticated (but still limited) statistical model to it... I'm just saying, maybe you don't want to go down in history as the 21st century's version of a phrenologist.
> I agree that it's important to be aware of the potential for these sorts of findings, but I continue to disagree that our takeaway should be an impetus to gird ourselves for the backlash when The Machine spits out, "Black people are dumb," rather than an impetus to observe the bedrock notion of scientific inquiry (you know, skepticism).
There is no shortage of pressure to avoid ML conclusions like "black people are dumb." We are already girded there. This isn't the first HN article about recognizing ML biases, and it's not like we're wading through headlines about how racial stereotypes are justified due to some all-knowing ML algorithm. We know we should be interrogating biases.
That said, it will always be possible to dismiss any evidence-based science by claiming there are unfalsifiable flaws with data or methods. That was the point of the comment I responded to. Outside the context of ML, this already happens, so we can already predict how it will happen with ML. And it is.
Nobody's saying, "just trust the machine if it says black people are dumb." I'm saying, "if the machine says sickle-cell anemia is most common among those with sub-Saharan African ancestry, it's still worth checking your methods and data, but it shouldn't be ignored just because unfalsifiable reasons why the analysis could be wrong can be postulated."
Your last analogy is an example of this. The link between sickle cell anemia and ancestry is a causal link that could be shown conclusively with a genetic test for a SNP. Finding that gene may have been an involved effort undertaken with a great many resources, but that link was established with near-perfect data and a clear conclusion.
The most promising are applications for ML are situations on the exact opposite end of the spectrum: problems where data is imperfect, noisy, or incomplete, where conclusions are not certain, but simply likely, to some measure of confidence.
We want to use ML to come to answers approaching conclusive, when a solid conclusion is impossible or unlikely, or would too long to obtain by conventional means. And we won't know how that conclusion is reached. So, in terms of conclusions that could spark wars, genocide, untold suffering, maybe we need to be more serious about our vigilance in regards to what we can control.
To put it bluntly, these concerns people have seem akin to NASA being more worried about handling post-catastrophe PR and internal messaging than keeping the shuttle from exploding in the first place.
It's what leads to us pretending there can't be any unpalatable differences between ethnic groups, clutching to presumptions about maybe the health care system being too racist, and avoiding looking deeper into it because we think we've already decided what the answer is (racism). Meanwhile real people are suffering because they aren't getting treated as effectively as they could.
We don't know what we don't know, so it's impossible to tell how many people are suffering or for what reasons.
So, I say again, the the priorities, as you've described them, are out of wack. We are much more at risk of jumping to damaging conclusions than we are of missing helpful breakthroughs. Our vigilance should be tuned in regard to this.
Exactly. Whether "conservative" or "progressive" in one's politics, the last decade or so has shown that when faced with a value contrary to one's own, or a bias that contradicts one's own currently held biases, one will simply dismiss the contradiction and likely hunker down on those preexisting values or biases.
In the current era that is manifesting itself as who can shout the loudest and most outrageously on social media or new-media, facts and reason be damned.
Well data doesn't lie. Because data doesn't _say_ anything. People interpret data and they do so according to their own inherent biases. And if the data is already biased (i.e. gathered according to peoples' biases) its interpretations end up far, far away from any objective truth.
You can always train a model to reject a class just as easily as you train it to accept a class. So train it to reject a common class and accept a mutant and it will function more like an evolutionary algorithm that protects against bad luck like bacteria with antibiotics.
I am way more optimistic for AI I guess.
>Machine learning systems are good at identifying cars that are similar to the cars they already know about. They’re also good at identifying faces that are similar to the faces they know about, which is why faces that are white and male are more reliably recognized by these systems — the systems are trained by the people who made them and the people in their circles.
But this seems like an absurd claim. Surely it's more likely that darker skin reflects less light than paler skin and so features are harder to detect on darker skin people. This explains why facial recognition works better on white people than on black people. Likewise women are more likely to wear makeup than men are and therefore are more likely to have a visually different face which could cause problems for facial recognition.
To assume that dark skinned people aren't involved in creating or testing facial recognition, or that women aren't, seems bizarre and vaguely racist/sexist. It's also a strange and I think poor assumption to think that the different accuracy rates for facial recognition on different demographics are due to a bunch of white dudes being too dumb to train their models on the faces of anyone but themselves.
The author also writes about how recommendation systems are "conservative" but to me it actually seems the opposite. Recommendation systems get you to stuff that you haven't tried before. They take what they know about you and progress you to new content or products.
For example, I used to listen to listen to a lot of country music on YouTube. Over time my recommendations evolved to include Irish folk music - which I quite like but would've never discovered without YouTube gradually testing and expanding my musical tastes.
It's unlikely anyone making a facial recognition system would choose to release a version that didn't work, at the very least, for everyone on the team.
If the team is only white men then you can see why they might accidently release something that only works on white men's faces. If there was a woman or a PoC on the team they'd have someone saying "Hang on, it doesn't work on me!"
The important thing to learn here is that any system is only ever as good as the data used to test it. If your test data sucks then the system you build is also going to suck.
It's just objectively not correct and not reasonable to think that the reason facial recognition works better on some demographics than others is because nobody thought of training models on those demographics. If you believe that is the case, then do you think people working on facial recognition reading this blog post are slapping themselves on the forehead saying "Of course. Train the models on someone besides the dev team!"?
The different accuracy of facial recognition by demographic group is common across multiple facial recognition systems. It's highly unlikely that all of the affected systems owe their inaccuracy to a post-training configuration issue.
To me it seems more like you are arguing for the sake of arguing rather than that you actually believe in your point. You are just trying to conceive of imaginative possibilities that preserve the "dumb facial recognition developers forgot non-white people exist" interpretation. The problem is that your imagined possible alternatives are not plausible explanations.
I also didn't say anything about ML training data in the post you replied to. My point was about facial recognition app developers not testing with a diverse range of inputs, which is why we see apps failing in public when people try to use them with inputs the app hasn't been tested against. If the apps were well tested those cases would have been caught before the public ever saw them.
My original point stands - if your testing sucks then it's very likely that your app will suck too. There will be problems with it that you don't know about, and the users will find them very quickly.
They should be, yes, but those training sets can have biases too, and if you use them your app will have the same biases. There are moves to make things like diverse data sets for this exact problem - https://www.theverge.com/2018/6/27/17509400/facial-recogniti...
A sample of on average five people seems to be a dumb way to train facial recognition - whether diverse or not.
I was talking about testing rather than training. You can train your facial recognition model on a diverse range of faces, but if you only test it on a few faces you don't know if it works for diverse people.
Teams of developers who test using their own data (eg faces) and nothing else is annoyingly common.
Also the argument that the group of people gathering, tagging etc the datasets, feeding them to algorithms and tweaking the algorithms is by definition truly diverse is .. well, have you looked at your average tech campus lately?
Or explained otherwise: then don't use techniques that depends on lightning reflection. The fact you use techniques that needs high lightning reflection means that you only care about the result with white faces and/or only tested fitness with white faces.... which is the piece main argument.
If I was trying to recognize a huge number of different people, who I had only ever known from still images, and all I had to recognize them were still images then yes, of course, it would be easier to recognize white men than black women for the reasons I just described. I actually don't understand how you can doubt that.
Imagine two images of a black woman. In one she has makeup and in the other she doesn't. Do you doubt that these two images are more different, and therefore harder to identify as the same person, than two pictures of a man, neither of which have makeup?
Do you genuinely think that the difference in facial recognition accuracy is because models aren't trained on dark skinned or female people?
I'm also at a loss as to how you could use any kind of camera or image that didn't depend on the reflection of light.
You're conflating "high lighting reflection" with "lighting reflection." Lighter faces reflect more light, but if you calibrate your instruments properly you can get just as much information out of a picture with less reflected light.
Also, obviously the difference between the dynamic range of cameras and our eyes play a major part in technical difficulties in multi-racial facial recognition as well?
(b) A mod changed the title after it was submitted.
(c) The site changed the article title after it was submitted.
I suspect (b), given the baity-ness of the article title: "Our Neophobic, Conservative AI Overlords Want Everything to Stay the Same", in accordance with the guidelines:
> "please use the original title, unless it is misleading or linkbait; don't editorialize."
If you think a title has been changed inappropriately (or should be changed), you can let the mods know directly using the Contact link in the footer.
Mate, it doesn't hate anything. It's math.
This is a moral / philosophical problem not technological one.
We have lots of things opposing each other:
- democracy and majority vote vs minorities
- solving 80% of problems efficiently vs ideal utopia
The world itself is biased, we are all biased, we all have different moral values. There is no definitive moral rule with which we could compare anything and easily say it's good or bad.
Conservatism isn't inherently bad just like openness to new ideas isn't. We usually need the right proportion of both to go forward.
Attaching emotional value to technology is also fine. "Machine learning is fundamentally conservative" gets the point across better than "machine learning doesn't handle change robustly," which is vague and doesn't give a person looking for something to read any reason why that would matter.
Writing alarmist articles for clicks being one of those reasons.
> Attaching emotional value to technology is also fine.
Why??? Doing this with other things would be ridiculous. For example, instead of "The poor are disproportionately affected by climate change" a journalist could write "The climate hates the poor". Which subtly plays down the human cause in all of this.
> "Machine learning is fundamentally conservative" gets the point across better
It demonstrably doesn't. If you just read this thread, you'll find enough people discussing whether "conservative" has a political meaning or just a "deviation from the norm" meaning.
a good TL;DR; of this article would be:
Computers are not yet sentient life forms capable of understanding the real wants and needs of human beings. Current day recommender systems, and predictive text systems are just very souped up statistics, that don't really have a lot of context or understanding of the world and their role in it.