Wish I had time for a longer comment, but this advice rings very true for me. Reflecting back on grad school and the transition to professional life, you have to realize that your role changes every couple of years and that the things that got you to one stage won’t get you to the next. Many people end up stuck in a local maximum (lacking vision) which partly explains the Peter principle.
As a mathematician who’s only recently started to get into computation and programming, I think the difference between my thought patterns when switching hats is so fascinating.
I was so accustomed to hearing that mathematics is nothing if not rigorous, but the more I reflect, mathematics is much more dependent on social convention and agreement amongst a community. While an outsider might think that proofs rigorously establish theorems, the purpose of a proof might be better seen as having enough detail to convince a substantial portion of the prominent mathematicians in a field that the proof is correct. In fact, there are theorems (e.g. the ABC conjecture) where a “proof” has been proposed, but not enough mathematicians have expertise with the techniques used to prove it in order to agree whether the proof is sufficient or not (though I’ve heard that the general opinion is that the proof does not suffice). William Thurston wrote one of my favorite essays related to this topic: https://www.math.toronto.edu/mccann/199/thurston.pdf
Reflecting on my own experience in mathematics, a better way to think of proofs is as being composed of “thought patterns” which many mathematicians agree are likely to be correct - when I scan a proof, I don’t look through every detail to verify that it is correct, but rather run it through a series of high level tests to see if it fails in any way, then if it passes all of those I look more closely at the argument and analyze the structure and mathematical power of each statement (e.g. one is unlikely to establish a hard analytic result through purely algebraic means, so where is the magic going on?) and so on until I’ve convinced myself that the argument is probably true. Other times, the result may be “visually apparent” (e.g. in geometry) at which point it might be sufficient for me to just to connect certain canonical arguments with the pictures as I read through the proof. For an excellent overview of this process, read Terry Tao’s blog on identifying errors in proofs : https://terrytao.wordpress.com/advice-on-writing-papers/on-l....
I don’t feel as confident commenting on the programming/computational perspective, as I’ve probably developed a very idiosyncratic way of thinking from approaching the topic so late in my education, but my feeling is that they are much different, and that the types of things a mathematician wants to convey to another mathematician rely much more on “trust” rather than the kind of rigor that might be needed by a computer.
I think this would be an interesting topic to explore in longer form.
I was just about to say, I was slightly disappointed to find that this wasn’t an article about a novel technique to identify fast moving water (akin to the fact that we can identify the temperature water by the sound it makes when it is poured).
It's one of those interesting examples where Hackernews's dual identities conflict... it's half "interesting engineering stuff" and half "startup culture stuff".
In this case, the article's business stuff made my eyes glaze over, lol, but that random thumbnail and metaphor they used was really interesting. I wish I could find the source of that image... unclear if it's just a mockup that NFX made for the blog, or if there's an actual machine vision water flow classification project somewhere.
My Googling is really failing me this time. I can find articles (scholarly or not) on kayaking, water physics, machine vision... but not at the intersection of the above, and not any other instances of that image or lookalikes. Most of the AI projects out there are classifying aerial images from a bird's eye view, at ocean scale, not "how fast is this water that I'm looking at from shore" scale.
As a mathematician who works with many engineers and computer scientists, I wanted to expand on one of the points under the “Getting a Job” section. While it is certainly true that a mathematics education provides a great background for understanding other STEM fields, I would caution math Ph.D students who expect these jobs to be open to them because of their STEM connection: the onus is completely on you to bridge the gap between what you do and the field you want to work in. While it may be true that someone will hire you for your critical thinking skills (though I will personally say that I have never seen this happen), it is more likely that your deep specialization in a tangentially connected field (coupled with not being involved in the culture/conferences of the community you wish to enter) will be an impediment to entering a new field: you expect to be paid like a Ph.D., but will potentially require years of training to get up to speed.
As an example, I remember the advice of “just go into data science” being handed out like candy to students interested in industry around the time I was in grad school (10 years ago). To be sure, there was a period where a STEM background + interest could get you in the door, but that time is over. These days you are competing with many equally brilliant students who have taken multiple courses and done research in this area, and it is highly unlikely that an employer will take a chance retraining an e.g. algebraic geometer with no precious data science experience to suit their needs.
All this to say, if you have an interest in another area, you must know the players and their work in that area while simultaneously knowing your area in math. It is not easy by any means, you are essentially signing up for twice as much work learning your field and theirs, but the rewards are great - as a connector between two fields, you have precious expertise that is very employable across a broad range of industries (my first job out of grad school was essentially providing advice on research programs helping connect different STEM communities to government funding agencies, but I was able to use my connections from that job to get back into research).
Yes, the onus is on you to learn the skills you need to show up and produce on day 1. That's 100% true and it is a lot of work. Having a math phd alone and interest is not enough.
However, there are still plenty of paths from math phd into data science and adjacent fields. Those courses you mention are things that someone with a math phd can 100% self teach, almost certainly to a level of understanding that is stronger than someone coming out of a DS masters. Learning how to interview on those concepts is important but ultimately they are very easy for a math phd once they know what the rules of the game are.
Personally, I spent the last year or so of my postdoc obsessively leetcoding and doing side projects in DS and landed a non-entry level data science position as my first job out of academia at a FANG. This is with a pure math research background totally unrelated to the position.
So it is still very possible. I think painting it like you do is a bit pessimistic and will discourage the wrong people. From personal experience, I saw comments like yours over the past year while I was job hunting and found them very discouraging.
The most important things are:
- have a network of similar people who have also made the transition (recently!) to get good advice and maybe also some referrals. These are the people you met in grad school a few years ahead of you.
- know exactly what type of position you want (or converge quickly) and focus on it relentlessly.
- understand the value you bring and the value others perceive someone like you to bring. Talking to people in hiring positions for different roles is the fastest way to learn what you have that is valuable. Do that as much as possible. Then you can line up how you value yourself with how a hiring manager values you, which will be the happiest result.
- take as many interview opportunities as possible to get that interview experience.
- Work relentlessly to interview better than those people from DS masters or whatever other sources they might come from.
It aggravates me too, however, considering that the definition of a tensor according to mathematicians is an element of the tensor product of two vector spaces (or whatever other objects you can tensor together), and according to physicists would be an object which transforms like a tensor, I’m somewhat sympathetic. Neither definition sheds any light on what a tensor is to anyone who doesn’t already understand what a tensor is, and I’m convinced that the moment one understands what tensors are they lose the ability to explain what tensors are.
I will say IMO (and experience) in professional math that while there is perhaps more of a chance for an outsider to have an impact, Mathematics is hardly free from bias towards insiders: it can manifest itself as subtly as using notation as a shibboleth (e.g. it’s somewhat easy to tell which community an author comes from through their notation and terminology, and equally easy to harbor resentment towards those outside your field) all the way to active “prove I’m the most clever in the room” syndrome during seminars. I’d like to think that a more collaborative atmosphere is prevailing now due to the rise of interdisciplinary and applied math, but people are people everywhere and as Sayre stated “Academic politics is the most vicious and bitter form of politics, because the stakes are so low.”
To your point, how can one measure 1% improvement in a meaningful way for abstract activities (e.g., learning)? The premise is that 1% improvement each day will yield large gains through the power of compounding, but if you aren’t sure you are making a quantified 1% improvement over your previous improved state things go awry quickly: if you only make 1% improvement over your original state each day then by the end of the year you have only improved by a factor of 4.65x, rather than your projected 37.8x improvement. Point being, if you want to take advantage of the sensitivity and power of exponential growth, you better be sure that you have a good way of quantifying your growth.
I know this is probably meant to be more inspirational than quantifiable, but then why even insert numbers unless to mislead people about the amount of effort it takes to improve? Either way you have to put in the same amount of effort to achieve the improvement, whether you break it up into 1% chunks or not there’s no free lunch.