
jalammar.github.io/illustrated-bert
Preview meta tags from the jalammar.github.io website.
Linked Hostnames
32- 12 links totwitter.com
- 10 links togithub.com
- 7 links toarxiv.org
- 3 links tojalammar.github.io
- 3 links towww.youtube.com
- 1 link toa-coles.github.io
- 1 link toai.googleblog.com
- 1 link toallenai.org
Search Engine Appearance
The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning)
Discussions: Hacker News (98 points, 19 comments), Reddit r/MachineLearning (164 points, 20 comments) Translations: Chinese (Simplified), French 1, French 2, Japanese, Korean, Persian, Russian, Spanish 2021 Update: I created this brief and highly accessible video intro to BERT The year 2018 has been an inflection point for machine learning models handling text (or more accurately, Natural Language Processing or NLP for short). Our conceptual understanding of how best to represent words and sentences in a way that best captures underlying meanings and relationships is rapidly evolving. Moreover, the NLP community has been putting forward incredibly powerful components that you can freely download and use in your own models and pipelines (It’s been referred to as NLP’s ImageNet moment, referencing how years ago similar developments accelerated the development of machine learning in Computer Vision tasks).
Bing
The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning)
Discussions: Hacker News (98 points, 19 comments), Reddit r/MachineLearning (164 points, 20 comments) Translations: Chinese (Simplified), French 1, French 2, Japanese, Korean, Persian, Russian, Spanish 2021 Update: I created this brief and highly accessible video intro to BERT The year 2018 has been an inflection point for machine learning models handling text (or more accurately, Natural Language Processing or NLP for short). Our conceptual understanding of how best to represent words and sentences in a way that best captures underlying meanings and relationships is rapidly evolving. Moreover, the NLP community has been putting forward incredibly powerful components that you can freely download and use in your own models and pipelines (It’s been referred to as NLP’s ImageNet moment, referencing how years ago similar developments accelerated the development of machine learning in Computer Vision tasks).
DuckDuckGo
The Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning)
Discussions: Hacker News (98 points, 19 comments), Reddit r/MachineLearning (164 points, 20 comments) Translations: Chinese (Simplified), French 1, French 2, Japanese, Korean, Persian, Russian, Spanish 2021 Update: I created this brief and highly accessible video intro to BERT The year 2018 has been an inflection point for machine learning models handling text (or more accurately, Natural Language Processing or NLP for short). Our conceptual understanding of how best to represent words and sentences in a way that best captures underlying meanings and relationships is rapidly evolving. Moreover, the NLP community has been putting forward incredibly powerful components that you can freely download and use in your own models and pipelines (It’s been referred to as NLP’s ImageNet moment, referencing how years ago similar developments accelerated the development of machine learning in Computer Vision tasks).
General Meta Tags
9- titleThe Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning) – Jay Alammar – Visualizing machine learning one concept at a time.
- charsetutf-8
- Content-Typetext/html; charset=utf-8
- X-UA-CompatibleIE=edge
- viewportwidth=device-width, initial-scale=1.0, maximum-scale=1.0
Open Graph Meta Tags
2- og:descriptionDiscussions: Hacker News (98 points, 19 comments), Reddit r/MachineLearning (164 points, 20 comments) Translations: Chinese (Simplified), French 1, French 2, Japanese, Korean, Persian, Russian, Spanish 2021 Update: I created this brief and highly accessible video intro to BERT The year 2018 has been an inflection point for machine learning models handling text (or more accurately, Natural Language Processing or NLP for short). Our conceptual understanding of how best to represent words and sentences in a way that best captures underlying meanings and relationships is rapidly evolving. Moreover, the NLP community has been putting forward incredibly powerful components that you can freely download and use in your own models and pipelines (It’s been referred to as NLP’s ImageNet moment, referencing how years ago similar developments accelerated the development of machine learning in Computer Vision tasks).
- og:titleThe Illustrated BERT, ELMo, and co. (How NLP Cracked Transfer Learning)
Link Tags
6- alternate/feed.xml
- stylesheet/css/bootstrap.min.css
- stylesheet/css/bootstrap-theme.min.css
- stylesheet/bower_components/jquery.gifplayer/dist/gifplayer.css
- stylesheethttps://cdnjs.cloudflare.com/ajax/libs/KaTeX/0.6.0/katex.min.css
Links
62- http://blog.class.vision/1397/09/bert-in-nlp
- http://creativecommons.org/licenses/by-nc-sa/4.0
- http://ruder.io/nlp-imagenet
- https://a-coles.github.io/2020/11/15/bert-illustre.html
- https://ai.googleblog.com/2018/11/open-sourcing-bert-state-of-art-pre.html