Highlights of the Month: March 2020

4 minute read

Published:

Key words Sub-Word Tokenization; Ligand-Based Virtual Screening; Meta Learning; ELECTRA;

Research Papers 🎓

NLP:

Pre-trained Models for Natural Language Processing: A Survey

A comprehensive review of Pre-trained Models for NLP

A Primer in BERTology: What we know about how BERT works

ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS. [Github]. [Blog]

Sub-Word Tokenization:

Ligand-Based Virtual Screening:

Ligand-based virtual screening utilizes information about molecules with known activity to predict the activity of new molecules. It assumes similar molecules tend to have similar activities/properties. Searching for active molecules can be viewed as similarity-based querying of a database storing molecules with unknown activity. Thus the use of molecular similarity measure is the cornerstone of the success of virtual screening. The following are papers of benchmarking different methods.

Meta Learning:

Meta-Learning: Learning to Learn Fast

Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks

Investigating Meta-Learning Algorithms for Low-Resource Natural Language Understanding Tasks

META-LEARNING INITIALIZATIONS FOR LOW-RESOURCE DRUG DISCOVERY

Others

Molecule Attention Transformer. [Code]

architecture

Breaking Down Structural Diversity for Comprehensive Prediction of Ion-Neutral Collision Cross Sections

  • A diverse set of 7405 molecues with CCS values.
  • Machine learning models to predict CCS.

A Deep Generative Model for Fragment-Based Molecule Generation

Software and Tools 💻

flair

A very simple framework for state-of-the-art NLP.

A text embedding library. Flair has simple interfaces that allow you to use and combine different word and document embeddings, including our proposed Flair embeddings, BERT embeddings and ELMo embeddings.

PyTorch implementation of SimCLR

textualheatmap

Create interactive textual heatmaps for Jupiter notebooks.

TensorFlow Quantum

An Open Source Library for Quantum Machine Learning

Generative Teaching Networks:

AI Generative Teaching Networks: Accelerating Neural Architecture Search by Learning to Generate Synthetic Training Data

Implementation of Generative Teaching Networks for PyTorch.

Articles and Blog Posts 📃

Introducing BART

Transformers are Graph Neural Networks

Graph Transformer tutorial

From PyTorch to PyTorch Lightning — A gentle introduction

PyTorch Lightning was created for professional researchers and PhD students working on AI research. PyTorch Lightning

WHAT IS TORCH.NN REALLY?

The Ultimate Guide to using the Python regex module

Tokenizers: How machines read

A Deep Dive into the Wonderful World of Preprocessing in NLP

Machine Learning for Everyone

Explain machine learning concepts with (1) simple words and (2) real-world examples.

From PyTorch to JAX: towards neural net frameworks that purify stateful code

Introducing DIET: state-of-the-art architecture that outperforms fine-tuning BERT and is 6X faster to train

Dual Intent and Entity Transformer (DIET) is a multi-task transformer architecture that handles both intent classification and entity recognition together.

TRAINING ROBERTA FROM SCRATCH - THE MISSING GUIDE

How to generate text: using different decoding methods for language generation with Transformers

FROM Pre-trained Word Embeddings TO Pre-trained Language Models — Focus on BERT

Meta learning

In meta-learning, there is a meta-learner and a learner. The meta-learner (or the agent) trains the learner (or the model) on a training set that contains a large number of different tasks. In this stage of meta-learning, the model will acquire a prior experience from training and will learn the common features representations of all the tasks. Then, whenever, there is a new task to learn, the model with its prior experience will be fine-tuned using the small amount of the new training data brought by that task.

Notable Mentions ✨

AI Curriculum. Open Deep Learning and Reinforcement Learning lectures from top Universities like Stanford University, MIT, UC Berkeley.

Huggingface’s official notebook tutorials

Limitations of Graph Neural Networks (Stanford University)

Leave a Comment