Hacker News new | past | comments | ask | show | jobs | submit login
Ask HN: What do data scientists need to know about LLMs to be competitive?
5 points by m_kos 8 months ago | hide | past | favorite | 9 comments
I have noticed job postings asking for experience with LLMs, but they are a bit vague. The most specific requirements I found were RAG, fine-tuning, and PyTorch.

I'd like to understand what I need to learn about LLMs to be competitive for data science positions. I have no experience with deep learning, generative models, and NLP, but I have a Ph.D. in a CS-adjacent area.




One thing to know about data scientists are (1) there's the job posting and (2) there's the work. Often you'll be hired for one thing and end up doing entirely different work.

To improve your chops in this field, (a) learn the basics of NLP, (b) build yourself a RAG using llamaindex or langchain (or other but build it).

As for fine tuning and deep learning, without experience in that field it's tough. It's like any complex thing (fine tuning not so much but deep learning), knowledge comes with time and exposure. So go find a reason to fine tune or build and train a neural network and go friggin do it.


Thank you for the advice! Regarding building a RAG, do mean making one based on, e.g., the this example? https://python.langchain.com/docs/expression_language/cookbo...


I wouldn’t waste my time learning about RAG. It’s quite likely GPT5 will make all current methods obsolete.


That’s not quite correct.

Private data especially in the enterprise cannot use public LLM’s like GPT-4 or 5 or N. Use cases needing data privacy have to use an internally implemented LLM application. In Currently, RAG is a concrete and pragmatic enterprise use of LLM’s aside from summarization, which is not amenable to using GPT-4.

GPT-5 may very well be amazing. But unless it runs on-prem it can’t be used in many scenarios because of data privacy.

To the OP - learning how to run LLM’s locally via say Ollama (see ollama.ai) will get you started in a hands on manner. See the /r/LocaLlama subreddit for a very active community around running LLM’s locally.


I think you’re missing my point. RAG is something that you currently implement yourself, or pay someone else who already implemented it. With GPT5 (or GPT6 at the latest), you just give it the same access you would give to a RAG system, and describe what you want it to do. It will do the rest.

Edit: I’m assuming the scenario where you do want to use the best model.


What would you recommend learning instead?


Learn whatever interests you the most.


What it is considered not to be a waste of time?


You say you have no experience with deep learning. Start by learning about "traditional" deep learning models, then move on to LLMs. Learning LLMs before deep learning is like learning algebra before addition.

"Deep Learning with Python, Second Edition" by Google's François Chollet is a pretty good intro, and also touches on LLMs / generative models towards the end.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: