- For full Data Science tasks, materials, etc. please check Data Science repository.
- For Machine Learning algorithms please check Machine Learning repository.
- For Deep Learning algorithms please check Deep Learning repository.
- For Computer Vision please check Computer Vision repository.
Folders with all materials for specific task/domain
[Natural Language Processing - Stanford University | Dan Jurafsky, Christopher](https://www.youtube.com/playlist?list=PLLssT5z_DsK8HbD2sPcUIDfQ7zmBarMYv) |
[Stanford CS224N: Natural Language Processing with Deep Learning | Winter 2021](https://www.youtube.com/playlist?list=PLoROMvodv4rOSH4v6133s9LFPRHjEmbmJ) |
Modern NLP techniques from recurrent neural networks and word embeddings to transformers and self-attention. Covers applied topics like questions answering and text generation.
[Stanford CS224U: Natural Language Understanding | Spring 2019](https://www.youtube.com/playlist?list=PLoROMvodv4rObpMCir6rNNUlFAn56Js20) |
[Natural Language Processing | University of Michigan](https://www.youtube.com/playlist?list=PLLssT5z_DsK8BdawOVCCaTCO99Ya58ryR) |
This is an extension to the (ML for) Natural Language Processing course teached at the Yandex School of Data Analysis (YSDA). For now, only part of the topics is likely to be covered here.
This course will cover search engine technologies, which play an important role in any data mining applications involving text data for two reasons. First, while the raw data may be large for any particular problem, it is often a relatively small subset of the data that are relevant, and a search engine is an essential tool for quickly discovering a small subset of relevant text data in a large text collection. Second, search engines are needed to help analysts interpret any patterns discovered in the data by allowing them to examine the relevant original text data to make sense of any discovered pattern. You will learn the basic concepts, principles, and the major techniques in text retrieval, which is the underlying science of search engines.
[Text Mining and Analytics [FULL COURSE] | UIUC](https://www.youtube.com/playlist?list=PLLssT5z_DsK8Xwnh_0bjN4KNT81bekvtt) |
Official Repository for ‘Practical Natural Language Processing’ by O’Reilly Media
Available as a Book: NLP in Python - Quickstart Guide
Title | Description | |
---|---|---|
ACL | at Vimeo | |
[Stanford CS224N: Natural Language Processing with Deep Learning | Winter 2021](https://www.youtube.com/playlist?list=PLoROMvodv4rOSH4v6133s9LFPRHjEmbmJ) | |
Natural Language Processing (NLP) Zero to Hero | by TensorFlow | |
Zero to Hero: NLP with Tensorflow and Keras (GDG Sofia meetup) | ||
Natural Language Processing | This content is based on Machine Learning University (MLU) Accelerated Natural Language Processing class. Slides, notebooks and datasets are available on GitHub |
I’m a research scientist at Google. I blog about natural language processing and machine learning.
A curated list of engineering blogs
The Gradient is an organization with the missions of making it easier for anyone to learn about AI and of facilitating discussion within the AI community. We were founded in 2017 by a group of students and researchers at the Stanford AI Lab.
Title | Description |
---|---|
NVIDIA Deep Learning Examples for Tensor Cores - Natural Language Processing | Deep Learning Examples |
Natural Language Processing with Transformers | Notebooks and materials for the O’Reilly book “Natural Language Processing with Transformers” |
Awesome NLP References | A curated list of resources dedicated to Knowledge Distillation, Recommendation System, especially Natural Language Processing (NLP) |
NLP - Tutorial | |
Natural Language-Process Tutorials | |
NLP with Python | Scikit-Learn, NLTK, Spacy, Gensim, Textblob and more… |
NLP and Data Science GitHub Repository Spotlight | Daily spotlights of some underrated NLP and Data Science GitHub repositories. |
NLP 101: a Resource Repository for Deep Learning and Natural Language Processing | This document is drafted for those who have enthusiasm for Deep Learning in natural language processing. If there are any good recommendations or suggestions, I will try to add more. |
NLP-progress | Repository to track the progress in Natural Language Processing (NLP), including the datasets and the current state-of-the-art for the most common NLP tasks. |
Hugging Face | Public repo for HF blog posts |
AllenNLP | An open-source NLP research library, built on PyTorch. Allenai.org |
NLP libraries, frameworks, modules
Title | Description |
---|---|
Natural Language Toolkit (NLTK) | NLTK - the Natural Language Toolkit - is a suite of open source Python modules, data sets, and tutorials supporting research and development in Natural Language Processing. |
flair | <ul><p>A very simple framework for state-of-the-art Natural Language Processing (NLP).</p><p>Flair is:</p><li>A powerful NLP library. Flair allows you to apply our state-of-the-art natural language processing (NLP) models to your text, such as named entity recognition (NER), part-of-speech tagging (PoS), special support for biomedical data, sense disambiguation and classification, with support for a rapidly growing number of languages.</li><li>A text embedding library. Flair has simple interfaces that allow you to use and combine different word and document embeddings, including our proposed Flair embeddings, BERT embeddings and ELMo embeddings.</li><li>A PyTorch NLP framework. Our framework builds directly on PyTorch, making it easy to train your own models and experiment with new approaches using Flair embeddings and classes.</li></ul> |
textacy | textacy is a Python library for performing a variety of natural language processing (NLP) tasks, built on the high-performance spaCy library. With the fundamentals — tokenization, part-of-speech tagging, dependency parsing, etc. — delegated to another library, textacy focuses primarily on the tasks that come before and follow after. |
AllenNLP | NLP research library, built on PyTorch, for developing state-of-the-art deep learning models on a wide variety of linguistic tasks. |
NLPGym | NLPGym is a toolkit to bridge the gap between applications of RL and NLP. This aims at facilitating research and benchmarking of DRL application on natural language processing tasks. The toolkit provides interactive environments for standard NLP tasks such as sequence tagging, question answering, and sequence classification. |
Gensim |
Title | Description |
---|---|
nlp-tutorial | <ul><p>Natural Language Processing Tutorial for Deep Learning Researchers</p><p>nlp-tutorial is a tutorial for who is studying NLP(Natural Language Processing) using Pytorch. Most of the models in NLP were implemented with less than 100 lines of code.(except comments or blank lines)</p></ul> |
Natural Language Processing in Python Tutorial | comparing stand up comedians using natural language processing |
🔹 Stanford, NLP Seminar Schedule
🔹 CS224n: Natural Language Processing with Deep Learning
🔹 CIS 700-008 - Interactive Fiction and Text Generation
Implemented a Bidirectional Attention Flow neural network as a baseline on SQuAD, improving Chris Chute’s model implementation, adding word-character inputs as described in the original paper and improving GauthierDmns’ code.
# Useful Articles
# Some Concepts