This repository contains everything you need to become proficient in ML/AI Research and Research Papers.
How to Make Best Use of ML/DL Research Papers?
Pic credits: ResearchGate
Youtube for all the implemented projects and tech interview resources - Ignito Youtube Channel
Complete Cheat Sheet for Tech Interviews - How to prepare efficiently
I took theses Projects Based Courses to Build Industry aligned Data Science and ML skills
Part 1 - How to solve Any ML System Design Problem
Link - Complete ML Research Papers Summarized Series
We will covering each and every Research Paper using 10 step framework —
Model Name | Link |
---|---|
Transformer | Link |
TransformerXL | Link |
VGG | Link |
Mask RCNN | Link |
Masked Autoencoder | Link |
BEiT | Link |
BERT | Link |
ColD Fusion | Link |
ConvMixer | Link |
Deep and Cross Network | Link |
DenseNet | Link |
DistilBERT | Link |
DiT | Link |
DocFormer | Link |
Donut | Link |
EfficientNet | Link |
ELMo | Link |
Entity Embeddings | Link |
ERNIE-Layout | Link |
FastBERT | Link |
Fast RCNN | Link |
Faster RCNN | Link |
MobileBERT | Link |
MobileNetV1 | Link |
MobileNetV2 | Link |
MobileNetV3 | Link |
RCNN | Link |
ResNet | Link |
ResNext | Link |
SentenceBERT | Link |
Single Shot MultiBox Detector (SSD) | Link |
StructuralLM | Link |
Swin Transformer | Link |
TableNet | Link |
TabTransformer | Link |
Tabular ResNet | Link |
TinyBERT | Link |
Vision Transformer | Link |
Wide and Deep Learning | Link |
Xception | Link |
XLNet | Link |
AlexNet | Link |
BART | Link |
InceptionNetV2 and InceptionNetV3 | Link |
InceptionNetV4 and InceptionResNet | Link |
Layout LM | Link |
Layout LM v2 | Link |
Layout LM v3 | Link |
Lenet | Link |
LiLT | Link |
Feature Pyramid Network | Link |
Feature Tokenizer Transformer | Link |
Focal Loss (RetinaNet) | Link |
Paper Name | Simplified/Summarized Version |
---|---|
Fine-mixing: Mitigating Backdoors in Fine-tuned Language Models | Link |
Bag of Tricks for Efficient Text Classification | Link |
Visualizing Linguistic Diversity of Text Datasets Synthesized by Large Language Models | Link |
(QA)²: Question Answering with Questionable Assumptions | Link |
QueryForm: A Simple Zero-shot Form Entity Query Framework | Link |
Semi-supervised Sequence Learning | Link |
Universal Language Model Fine-tuning for Text Classification | Link |
DARTS: Differentiable Architecture Search | Link |
RoBERTa: A Robustly Optimized BERT Pretraining Approach | Link |
Generating Sequences With Recurrent Neural Networks | Link |
Deep contextualized word representations | Link |
Regularizing and Optimizing LSTM Language Models | Link |
End-To-End Memory Networks | Link |
Listen, Attend and Spell | Link |
Well-Read Students Learn Better: On the Importance of Pre-training Compact Models | Link |
Language Models are Few-Shot Learners | Link |
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context | Link |
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter | Link |
Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond | Link |
LIMA: Less Is More for Alignment | Link |
Efficient Neural Architecture Search via Parameter Sharing | Link |
Tree of Thoughts: Deliberate Problem Solving with Large Language Models | Link |
AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head | Link |
FrugalGPT: How to Use Large Language Models While Reducing Cost and Improving Performance | Link |
CodeT5+: Open Code Large Language Models for Code Understanding and Generation | Link |
Unlimiformer: Long-Range Transformers with Unlimited Length Input | Link |
Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4 | Link |
PaLM: Scaling Language Modeling with Pathways | Link |
Attention Is All You Need | Link |
Denoising Diffusion Probabilistic Models | Link |
ZeRO: Memory Optimizations Toward Training Trillion Parameter Models | Link |
Wide Residual Networks | Link |
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness | Link |
STaR: Bootstrapping Reasoning With Reasoning | Link |
Meta-Gradient Reinforcement Learning | Link |
Distilling the Knowledge in a Neural Network | Link |
How to Fine-Tune BERT for Text Classification? | Link |
Primer: Searching for Efficient Transformers for Language Modeling | Link |
Training Compute-Optimal Large Language Models | Link |
Learning Transferable Visual Models From Natural Language Supervision | Link |
More Coming soon |
Paper Name | Summarized and Simplified Version |
---|---|
NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis | Link |
More Coming Soon |