Hacker News new | comments | show | ask | jobs | submit login
Practical Deep Learning for Coders 2018 (fast.ai)
389 points by jph00 23 days ago | hide | past | web | favorite | 61 comments



Jeremy from fast.ai here. Happy to answer any questions about the course, fastai, or anything else relevant!

BTW the 2018 version of the course is being discussed in this forum, for those interested: http://forums.fast.ai/c/part1-v2


I have no questions, but feel compelled to say thank you.

I've been a ML/DL practitioner for the past five-plus years, and first watched one of your lectures a bit over a year ago. All I remember thinking is that a wealth of practical knowledge that had taken me years to acquire was there for the taking, for free, for anyone who cared to look.

Since then, I have been recommending these courses to anyone who asks me for advice for learning about deep learning. You -- and Rachel Thomas -- have created by far the easiest and fastest path for a wide range of people to gain deep learning expertise.

In fact, I'm so sure the new lectures will contain valuable nuggets of know-how that even though I consider myself pretty knowledgeable about deep learning (and an expert in my narrow domain of interest), I will make it a point to find time to watch all the updated lectures.

Kudos!


Hi jeremy,

I am a web dev, JS being my first lang of choice :-), and i have been trying to get into machine learning/deep learning/AI, but I am having a information overload.

I have 0 knowledge about this field, so my question is should i just start with machine learning, instead of jumping right into deep learning ? Or is it ok to jump right into deep learning and is it possible to do all the stuff done with machine learning with deep learning ?

Just like when a person who shows an interest on learning to code I refer them to Python instead of C++, what do you recommend ?


Just jump in with this course. Your background in web dev will be really helpful, but there will be plenty of new concepts to learn - don't try to understand everything perfectly the first time through, but just try to complete each week's assignments as best as you can. You'll need to learn python and numpy along the way, but there's plenty of free online resources you can refer to when you see something you don't understand. And use http://forums.fast.ai of course.

You can then go back through them a 2nd time and do a deeper dive. By that time, our Intro to Machine Learning course will be out too :)


So. Many. Resources. When I have the time I'm watching the first course with Keras and Theano, haven't gotten to the second course with Tensorflow, and now this. My question is, is it worth my while to continue watching the Keras/Theano videos or should I skip to the Keras/Tensorflow videos or even to the fastai/PyTorch videos? Are the same concepts covered or does each series cover different things?


Definitely skip to this new course. Sorry I should have provided a link to the FAQ where this and other questions are answered: http://forums.fast.ai/t/part-1-faq/10330

The concepts and skills you learned from the 2017 are entirely transferable - learning the software packages is the easiest part, and the various libraries are similar enough that switching between them doesn't take much time.


Jeremy, I don't have a question, but just wanted to thank you guys for fastai. It is just the right mix of "throw you in the pool" and "I'm still here, you aren't going to drown"!


Hi Jeremy,

I too want to say thank you, even though I have only started with the material. Your philosophy on how to teach the subject I feel will be much more useful to me than the classes I was signed up for recently (and lost interest in rather quickly). It's a lot easier to get excited about the practical things surrounding the topic which is why I'm looking forward to diving into the new content.

On a related note, one thing I do see a heavy emphasis on with the material is on the Computer Vision / Image Processing side of things which is certainly understandable considering how popular that area is specifically right now.

Something not really computer vision related though that I'm curious about (and I'm not sure if it's covered in the new/existing lessons) would be on how to craft a data set using data I might have accessible to me, but which isn't necessarily image-based, and to apply these techniques to that sort of data set to come up with predictions (I bring this up, because one of the goals I have for learning about this topic specifically is to see how I might be able to apply it back to my job at a community college and if I can pull historical data related to our students and use that for forecasting / recommendation purposes and create some useful applications our students can utilize...as a simple idea one example would be using historical data about the current student, and maybe data from other similar students, to predict success in a student's upcoming courses).

Thank you and keep up the awesome work (and for sharing it freely :-)!


Jeremy,

Just wanted to say thanks for all your work. I'm taking some time off from my regular job as an iOS developer to pursue an ML project of mine. I wouldn't have had the confidence to do that with out your online course. I wrote up a project I did after last years class on Medium, also because of your guidance.

https://towardsdatascience.com/25-lights-part-ii-e021b66e449...


Thanks for sharing! FYI you may find resnet18 or (better still) densenet a good option for that dataset - you shouldn't need so many training images then.


Jeremy, thanks for putting this out. I had issues with part 1, v1 with the AWS set up. But I had none of those issues with this course. As a beginner programmer the template provided via Paperspace is just what I needed. I've been waiting for this course to be released for a really long time now, so I am incredibly excited to have successfully set up the environment and am now FINALLY ready to learn. Thank you!


Two questions:

1. You've made an off-hand comment on one of your videos that a sequential dense network is just a generalization of any other type of neural network architecture. In theory you could re-create an RNN or CNN through just Dense layers. But obviously it's not practical.

Why isn't it practical? Is it because the network would have to be too deep, or too wide? Would the optimizer just get stuck in a local minima or would overfitting be inevitable? Or perhaps some combination of issues?

What do you think is the best hope for a generalized network architecture, most similar to our brain?

2. On a somewhat related note, do you have a strong enough faith in the current machine learning algorithms and architectures being used (RNNs, CNNs, capsule networks) that given infinite resources (time training and network size), that we would be able to create a meaningful general AI? Or do you think that our current approach is just incremental and a truly different approach would be required to achieve meaningful AI?


> Why isn't it practical? Is it because the network would have to be too deep, or too wide? Would the optimizer just get stuck in a local minima or would overfitting be inevitable? Or perhaps some combination of issues?

Schmidhuber did a paper a few years ago showing near SoTA performance on computer vision using just a fully connected net. One of our students showed how a convolution is just a weight-tied matrix multiply here: https://medium.com/impactai/cnns-from-different-viewpoints-f...

So the issue is that without the weight-tying, you've got more parameters to regularize (which can decrease performance) and train (which takes longer). So you should use weight tying where you can - e.g. by using convolutions.

In general, domain-specific architectures try to find structure in the underlying data and problem, and use that to decrease the number of parameters we need. The use of implicit factorizations in the inception and xception architectures is a good example.


Really looking forward to getting stuck into this course. I’ve a background in fluid mechanics, in particular writing code (in Matlab but more recently with python) to identify structures in experimental and numerical data. In essence this is very similar to classical image analysis. Looking to machine learning seems a logical next step. I’ve been working through Andrew Ng’s course and while it is very good it suffers from fundamentalitis to a degree. However I do plan on completing Ng’s course first.


Is the 'fastai' library on Pytorch something you'd recommend only for beginning/learning/this course, or is it intended to be used in production beyond the course? Is it being positioned more as a learning aid or as an open source library with a life beyond this course?

Also, any chance of https on the forum?


Definitely well beyond the course. It's designed to be the easiest way to create world-class models. I'll be providing a lot more information on how we're doing this in the next week or two.

PS: The focus of fastai is training, not production. The models you end up with are largely standard pytorch models, so standard pytorch approaches to production work fine. For most people, a simple flask endpoint with CPU inference is generally the best approach for this (e.g. all the deep learning web apps at http://allenai.org/ use this approach AFAIK.)


Do you envision people using the fast.ai library without ever taking the course?


I do, but at the moment the course is the documentation, on the whole - so it won't really be standalone until we're written proper docs! (There's basic docstrings for most functions, but they don't tell you how it all fits together.)


I thought the first version of the course was amazing but this version unlocked a new level of brilliance. Specially, the fastai library. I'm counting the days to part 2. Thank you so much! You're making a huge difference around the world! This is education at its finest


Started this weekend and finished the first two chapters. Loving it so far. This is the first course that makes me so excited and I keep thinking about when I can start doing the next chapter. The forums and other resources well documented and extremely useful.

Thanks Jeremy for putting this together.


Hi Jeremy, thank you so much for your great effort, it helps me a lot.

Would you release a new version for the part2 course this year? I haven't started the part2 yet, but I heard that the part2 course is using the new Pytorch based library already.


Very excited about the pytorch version of this course. Will you offer more NLP focused materials in future courses. I've found that the majority of the deep learning materials gravitate towards vision and image classification.


What's the recommended cloud runtime for v2? Is it still AWS, or is it Crestle or similar?


We suggest Paperspace now. Lesson 1 walks you through getting set up. It's very easy - there's a fast.ai template there ready to go. They have the best price/performance ratio of anyone at the moment for GPUs.


Can I still use AWS though? Does the setup scripts remain the same? I had already set up an AWS server (my company has provided us some cash for personal learning) from the old lessons, but haven't started the lessons yet.

It'd be great if I can switch to the new lessons but with AWS.


Yup AWS is fine too. We have a new AMI available if you want to use it, and also a predefined conda environment to add fastai/pytorch to an existing server. Lesson 2 covers AWS usage, and there's also step by step instructions linked from the forum.


How much should I budget for paperspace for completing the course?


If you remember to shut down your machine when you're done, and you do the suggested 10 hours per week, plus maybe an extra 5 hours of model training per week, at $0.45/hour for Paperspace's cheapest machine (still works great though): $47.25


Jeremy,

I'm having trouble finding the (old) 2017 material. I would like to see it for my reference. Do you know where this is located?


Here you go: http://course17.fast.ai . (It's also linked from the 2nd paragraph of the new course FYI.)

The old forum is still at the same place http://forums.fast.ai/c/part1


This is so uncool!

This is again going to take up my weekends just like their previous course. :D

Congrats on the new course! Thank you. :)


> This is so uncool!

Why thank you :) Being uncool is our specialty.


Hottest course in AI town!!!

(I was trying to be sarcastic. Thankful for the courses :)).


Do you have good references about how to setup the environment with a Nvidia GPU in a Windows machine?


Dan from Paperspace here. Our fast.ai template is Linux based but you can spinup a GPU backed Windows instance on our cloud if you’re interested in a remote option. Our streaming tech is GPU accelerated so it feels snappy.


Yes, I've provided a step by step walk-through here: http://forums.fast.ai/t/howto-installation-on-windows/10439

This includes setup of fastai, pytorch, cuda90, cudnn, opencv, bcolz, and much more!


Hi Jeremy, I just recently "upgraded" from a Macbook air to a Windoze Gaming Laptop with the GTX1050 GPU just for the reason that one day I could hopefully do all the Fast.ai assignments on my laptop. I think this announcement right here will spur me in taking up this course and Finishing it. I have an actual need for automating some of our IQC/OQC( Incoming and outgoing Quality control) of sub-assemblies and parts that we purchase from our vendors and I hope to leverage what I learn from this course. Thanks a Ton for democratizing Deeplearning and ML for a large population. Ananth


My pleasure! You may need to reduce the batch size in a few lessons to fit on that card, FYI.


I just finished the "soft launch" version of this. It's a truly excellent way to get started with deep learning. The course is very hands on and practical. This version has been updated to use pytorch and a bunch of new techniques. Highly recommended if you are looking to get started with deep learning.


Time to shill my little tool to help with rewatching lecture videos. Search the caption by keyword and jump to that timestamp of the video. https://github.com/mingrui/youtube_caption_search


Thank you for creating that! :)

BTW I just had a thought. What if instead of a txt file like you have:

    hi everybody welcome to practical deep
    00:00:00.000
    https://www.youtube.com/watch?v=IPBSB1HLNLo#t=00h00m00s
...you instead created a markdown file that had:

    [hi everybody welcome to practical deep]
    (https://www.youtube.com/watch?v=IPBSB1HLNLo#t=00h00m00s)
...then you'd have a readable transcript, where every phrase is hyperlinked! For bonus points, run it through the bidir RNN punctuator to add punctuation: https://github.com/ottokart/punctuator2 . (I plan to do this for the transcripts in the future BTW).


Good idea on markdown, it should be a quick fix and quality of life improvement.

Didn’t know punctuator existed, I’ll look into that!


Is this course free? What's the catch? I can't seem to find pricing anywhere.

(this is a genuine question, and not a meta-comment)


It's free. We're a self-funded non-profit. We do this because we actually want to help people!


Sweet! Thank you!


Hi Jeremy, recently google have free gpu with colab notebook product. Student can use it when they need free gpu option. https://colab.research.google.com/notebook


fast.ai international fellowship recipient here - The second iteration of the fast.ai course is totally worth. If you can code in Python, you can master deep learning. Join our fantastic community!


Hi Jeremy, i have an computer running ubuntu 17.10 with 16gb ram and 4gb nvidia gpu. Is it possible to setup my system to run the exercises locally without having to use paperspace? Thank you.


Yup that should be fine. There's resources on http://forums.fast.ai to help you get set up.

(I'd still suggest doing the first few lessons on Paperspace so you can focus on the deep learning, rather than the setup. It's only $0.45/hour and 20 hours is plenty enough to get going. Sometimes getting your computer set up can be distracting and frustrating at first!)


Can anyone reccomend an "AI for Dummies"? Every resource I've yet to find starts with assuming I have any idea what "Eigenvector" or "Gradient descent" means.


That would be this course. It assumes neither of those things :)

(It's for dummies that are prepared to work hard over a 7 week period and that have been coding for at least a year.)


Did anyone do Andrew Ng's deeplearning.ai and know how it compares?


Yeah there's a discussion of this on http://forums.fast.ai if you want the details. Basically, Andrew's course is bottom-up (i.e. start with the theory and math and simple examples) and works up from there. Ours is top-down (start with code to train world-class models and dig deeper and deeper in to how they work). fast.ai also has more examples of more recent techniques, and covers a wider variety of applications. Students who have done both tell us that they found both helpful to their understanding of the field and practice.


I'm currently doing Andrew Ng's course and wrote a short review of course 1 here [1]. If you have the mathematics and statistics background and want to go through that rigorously enough then it's exceptional. If you are not the mathematically inclined it is still accessible but I can imagine fast.ai being more appropriate, although I have no personal experience.

[1] https://dandermotj.github.io/post/review-deeplearning-ai-cou...


What are the benefits of the in-person "Part 2" in SF? Would it be feasible to fly in weekly for the course?


There are certainly folks that fly in weekly. One student flew in from Australia for the duration! But if you're far away you'd probably be better off applying to be an international fellow, which means you watch the lesson live over Youtube Live and can ask questions and interact with me and the class in real time: http://www.fast.ai/2018/01/17/international-spring-2018/

If you're considering the investment of weekly flights, you should probably first ask the students at http://forums.fast.ai whether they would recommend it.


I had a good conversation with the student who flew in from Australia. She said she got a lot out of it, and found the discussion groups very useful.

OTOH, I've done it remotely just watching the videos and I've found it great too.


Thanks for the information.

I'm only down the coast in SD so it wouldn't be too bad... but certainly something to consider.


Can I run through this course on a no-GPU laptop or would that be too slow?


You setup your environment in a cloud environment that gives you GPU access. Details in the course and forums.


Enjoyed the first iteration of the course, will definetly check this one!




Applications are open for YC Summer 2018

Guidelines | FAQ | Support | API | Security | Lists | Bookmarklet | Legal | Apply to YC | Contact

Search: