Hacker News new | past | comments | ask | show | jobs | submit login
SpaCy PyTorch Transformers (explosion.ai)
106 points by syllogism 78 days ago | hide | past | web | favorite | 11 comments

Spacy also has one of the best documentations that I have come across. I really like the use of Tufte-style notes for code examples. https://spacy.io/api

Also the recent SpaCy IRL talks are well worth watching @ https://www.youtube.com/watch?v=hNPwRPg9BrQ&list=PLBmcuObd5A...

I'm surprised transfer learning via fine-tuning large transformer models hasn't taken off more in the public consciousness a la image recognition models. In my experience, the results can be staggering with very small amounts of training data.

It has taken off. Pretty much everyone and their grandma has fine tuned gpt-2 to generate all kinds of stuff, even poetry: https://www.gwern.net/GPT-2

It's a very new thing. Pretrained ImageNet models were first released around 2011. Pretrained transformer models have only been released recently.

Also, transformers are only useful for short text, not full documents (AFAIK).

are transformer models are the text version of imagenet model? The first time I am hearing this term.

Google (and I assume most ML providers) offers transfer learning on image models: https://cloud.google.com/vision/automl/docs/

Awesome! Cant wait to try this out. Are there any plans in the future to incorporate sentence/word generation?

That would be cool, yeah. We haven't had anything like that in spaCy so far, but the results from these models can be very good. I think with a good API it could be very useful.

This stuff is great! Will example training files be uploaded for GLUE/SQAUD/etc like on the HuggingFace's implementation?

Very cool!

Guidelines | FAQ | Support | API | Security | Lists | Bookmarklet | Legal | Apply to YC | Contact