- mlsys
- transformer
- paper-summaries
- MLSys
- LLMs
- PPML
•
•
•
•
•
-
Paper Summary #7 - Efficient Transformers: A Survey
A survey paper of improvements over the original Transformer architecture in terms of memory-efficiency.
-
Deploying Machine Learning models using GCP's Google AI Platform - A Detailed Tutorial
A step-wise tutorial to demonstrate the steps required to deploy a ML model using GCP, specifically the Google AI Platform and use Streamlit to access the model through a UI.
-
Deploying Machine Learning models using AWS Lambda and Github Actions - A Detailed Tutorial
A step-wise tutorial to demonstrate the steps required to deploy a ML model using AWS Lambda, Github Actions, API Gateway and use Streamlit to access the model API through a UI.
-
PPML Series #3 - Federated Learning for Mobile Keyboard Prediction
Understanding how your mobile keyboard (Gboard, specifically) performs the next word prediction task and performs model training and updates
-
PPML Series #2 - Federated Optimization Algorithms - FedSGD and FedAvg
A mathematical deep dive on a Federated Optimization algorithm - FedAvg and comparing it with a standard approach - FedSGD.
-
PPML Series #1 - An introduction to Federated Learning
A short general introduction to Federated Learning (FL) for folks interested in privacy-preserving machine learning (PPML).
-
Paper Summary #6 - Language Models are Unsupervised Multitask Learners
The GPT2 model which aimed to perform complex NLP tasks while relying only on a language model trained in a completely unsupervised fashion.
-
Paper Summary #5 - XLNet: Generalized Autoregressive Pretraining for Language Understanding
XLNet tries to overcome the limitations of BERT by having a autoregressive component while also capturing the bidirectional context.
-
Paper Summary #4 - BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
The ground breaking paper that introduced the famous BERT model. This started the inflow of a large number of BERT-based language understanding models.
-
Paper Summary #3 - Improving Language Understanding by Generative Pre-Training
The first paper in the GPT set of models. This is OpenAI's GPT-1.