Welcome to the AI repository! This repository is dedicated to implementing machine learning models from cutting-edge research papers across various domains of Artificial Intelligence. Whether you're looking to replicate the latest breakthroughs or understand how to bring theory into practice, this repo will provide clean, well-documented implementations to help you get started.
This repository features implementations of important machine learning models, accompanied by their original research papers. Each project contains:
- Model implementations with clean, well-documented code
- Links to research papers for reference
- Instructions for running, training, and testing models
- Key insights and discussions about the model architectures
The goal is to provide a comprehensive learning resource for those who want to explore the underlying techniques of modern AI research.
Here’s the ordered list of models that will be implemented, each based on influential papers. This list spans from foundational models to cutting-edge architectures.
- Paper: The Perceptron: A Probabilistic Model for Information Storage and Organization in the Brain (1958) by Frank Rosenblatt
- Paper: Learning representations by back-propagating errors (1986) by David E. Rumelhart et al.
- Link: Paper1, Paper2
- Paper: Backpropagation Applied to Handwritten Zip Code Recognition (1989) by Y. LeCun et al.
- Paper: Gradient-Based Learning Applied to Document Recognition (1998) by Yann LeCun et al.
- Link: Paper
- Paper: Finding Structure in Time by Jeffrey L. Elman (1990)
- Link: Paper1
- Paper: Long Short-Term Memory (1997) by Sepp Hochreiter and Jürgen Schmidhuber
- Link: Paper
- Paper: Reducing the Dimensionality of Data with Neural Networks (2006) by Geoffrey Hinton and Ruslan Salakhutdinov
- Link: Paper
- Paper: ImageNet Classification with Deep Convolutional Neural Networks (2012) by Alex Krizhevsky et al.
- Link: Paper
- Paper: Playing Atari with Deep Reinforcement Learning (2013) by Mnih et al.
- Link: Paper
- Paper: Auto-Encoding Variational Bayes (2013) by Kingma and Welling
- Link: Paper
- Paper: Neural Machine Translation by Jointly Learning to Align and Translate (2014) by Dzmitry Bahdanau et al.
- Link: Paper
- Paper: Generative Adversarial Nets (2014) by Ian Goodfellow et al.
- Link: Paper
- Paper: Deep Residual Learning for Image Recognition (2105) by Kaiming He et al.
- Link: Paper
- Paper: Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation (2014) by Cho et al.
- Link: Paper
- Significance: GRUs are a simpler and more efficient variant of LSTMs, with fewer gates. They retain the ability to model long-term dependencies but are computationally less expensive.
- Paper: You Only Look Once: Unified, Real-Time Object Detection (2016) by Joseph Redmon et al.
- Link: Paper
- Paper: Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network (2016) by Shi et al.
- Link: Paper
- Paper: Attention Is All You Need (2017) by Vaswani et al.
- Link: Paper
- Paper: Proximal Policy Optimization Algorithms (2017) by Schulman et al.
- Link: Paper
- Paper: Wide Residual Networks (2017) by Zagoruyko and Komodakis
- Link: Paper
- Paper: MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications (2017) by Howard et al.
- Link: Paper
- Paper: Densely Connected Convolutional Networks (2017) by Huang et al.
- Link: Paper
- Paper: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2019) by Devlin et al.
- Link: Paper
- Paper: Inverted Residuals and Linear Bottlenecks: Mobile Networks for Classification, Detection, and Segmentation (2018) by Sandler et al.
- Link: Paper
- Paper: EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks (2019) by Tan and Le
- Link: Paper
- Paper: A Simple Framework for Contrastive Learning of Visual Representations (2020) by Chen et al.
- Link: Paper
- Significance: Key development in self-supervised learning, influencing many computer vision applications.
- Paper: MobileNeXt: Enhanced Inverted Residuals for Efficient Mobile Vision Applications (2020) by Zhou et al.
- Link: Paper
- Paper: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale (2020) by Dosovitskiy et al.
- Link: Paper
- Paper: Denoising Diffusion Probabilistic Models (2020) by Ho et al.
- Link: Paper
- Paper: Learning Transferable Visual Models From Natural Language Supervision (2021) by Radford et al.
- Link: Paper
- Significance: Integrated image and text representations, showing significant progress in multimodal learning.
- Paper: High-Resolution Image Synthesis with Latent Diffusion Models (2022) by Rombach et al.
- Link: Paper
- Significance: An evolution of generative models, making high-quality image generation practical.
This repository is licensed under the Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0) - see the LICENSE file for details.