All Day AI
Confernce
All Day AI Virtual Conference
2023-12-12 08:00:00
Register Now!
  • Home
  • Speakers
  • Schedule
  • News
  • Topics
    • Types of machine learning Algorithms
  • Registration
  • Call For Papers
  • Home
  • Speakers
  • Schedule
  • News
  • Topics
    • Types of machine learning Algorithms
  • Registration
  • Call For Papers
NewsNewsNews

Latest News

05/202023

ColossalChat: An Open-Source Solution for Cloning ChatGPT With a Complete RLHF Pipeline

footer

Thanks to Yang You

Large AI models and applications like ChatGPT and GPT-4 have become extremely popular worldwide, serving as a foundation for the technological industrial revolution and the development of AGI (Artificial General Intelligence). Not only are technology giants racing to release new products, but many AI experts from academia and industry are also joining the related entrepreneurial wave. Generative AI is rapidly iterating on a daily basis, continuously improving!

However, OpenAI has not made its models open source, leaving many curious about the technical details behind them.

  • How can we stay current and participate in this wave of technology development?
  • How can we lower the high cost of building and applying large AI models?
  • How can we protect core data and IP from being leaked through third-party large model APIs?

As the leading open-source large AI model solution today, Colossal-AI is the first to open-source a complete RLHF pipeline that includes supervised data collection, supervised fine-tuning, reward model training, and reinforcement learning fine-tuning, based on the LLaMA pre-trained model, and shares ColossalChat, the most practical open-source project that closely resembles the original ChatGPT technical solution!

Open source address: https://github.com/hpcaitech/ColossalAI

It includes the following contents:

  • Demo: an interactive demo to try it online without registration or joining the waiting list.
  • Training code: Open-source and complete RLHF training code, including 7B and 13B models.
  • Dataset: Open-source 104K bilingual dataset of Chinese and English.
  • Inference: 4-bit quantized inference for 7 billion-parameter models that only require 4GB GPU memory.
  • Model weights: Achieve quick reproduction with only a small amount of computing power on a single server.
  • Additional larger models, datasets, and other optimizations will be rapidly updated and added.

Affordable models, powerful capabilities

ColossalChat only requires less than 10 billion parameters to attain bilingual proficiency in English and Chinese, through RLHF fine-tuning on the basis of large language models, achieving comparable results to ChatGPT and GPT-3.5.

For example, a general knowledge quiz

Answering in Chinese

Write an email

Write an algorithm

Complete ChatGPT cloning solution

Although models in the GPT series, such as ChatGPT and GPT-4, are highly powerful, they are unlikely to be fully open-sourced. Fortunately, the open-source community has been working hard to address this, especially in the most widespread and easy-to-use PyTorch community.

For example, Meta has open-sourced the LLaMA model, which offers parameter sizes ranging from 7 billion to 65 billion. A 13 billion parameter model can outperform the 175 billion GPT-3 model on most benchmark tests. However, since it doesn’t have an instruct tuning stage, its actual generated results are not satisfactory.

Stanford’s Alpaca generates training data in a self-instructed manner by calling OpenAI’s API. With only 7 billion parameters, this lightweight model can be fine-tuned at a fraction of the cost to achieve conversational performance similar to a very large language model like GPT-3.5 with 175 billion parameters.

However, existing open-source solutions can only be considered as supervised fine-tuned models in the first stage of RLHF (Reinforcement Learning from Human Feedback), with subsequent alignment and fine-tuning stages not performed. Additionally, Alpaca’s training dataset is limited to English, which to some extent restricts the model’s performance.

Yet, the impressive effects of ChatGPT and GPT-4 are due to the introduction of RLHF into the training process, which increases the consistency of the generated content with human values.

Three Stages of RLHF [6]

Based on the LLaMA model and the widespread AI framework PyTorch, ColossalChat is the first practical open-source project that includes a complete RLHF process for replicating ChatGPT-like models, and is the closest project to the original technical route of ChatGPT!

Utilizing PyTorch in the development of ColossalChat is crucial, as it provides a flexible and efficient deep-learning framework. This allows for easier experimentation, rapid prototyping, and seamless integration with other libraries, ultimately enabling ColossalChat to deliver a high-performance, user-friendly conversational AI experience.

Training Dataset Open Source

ColossalChat releases a bilingual dataset comprising approximately 100,000 Q&A pairs in both English and Chinese. The dataset was collected and cleaned from real-life question scenarios on social media platforms, serving as the seed dataset, and was expanded using self-instruct technology, and annotation costs were approximately $900. Compared to datasets generated by other self-instruct methods, this dataset contains more realistic and diverse seed data and encompasses a wider range of topics. The dataset is suitable for both fine-tuning and RLHF training. With the provision of high-quality data, ColossalChat can achieve better dialogue interactions and also support Chinese.

ColossalChat Dataset Collection Process

RLHF Algorithm Replication

The RLHF algorithm replication involves three stages:

In RLHF-Stage1, supervised instruct fine-tuning is performed using the datasets mentioned earlier to fine-tune the model.

In RLHF-Stage2, a reward model is trained to assign corresponding scores by manually ranking different outputs for the same prompt, which then supervises the training of the reward model.

In RLHF-Stage3, the reinforcement learning algorithm is being used, which is the most complex part of the training process:

RLHF-Stage3 Algorithm Flowchart

In the PPO part, ColossalChat follows a two-stage process: first, the make experience stage, which uses SFT (Supervised Fine-Tuning), Actor, RM (Reward Model), and Critic models to calculate generated experience and store it in the buffer. Then comes the parameter update stage, which calculates the policy loss and value loss using the experience.

In the PTX part, ColossalChat calculates the cross-entropy loss between the Actor’s output response and the response part of the input corpus. This loss is used to add pre-training gradients to the PPO gradient to maintain the language model’s original performance and prevent forgetting. Finally, the policy loss, value loss, and PTX loss are summed up for backpropagation and parameter update.

Quick Start

ColossalChat has open-sourced the complete code for replicating ChatGPT in three stages based on the LLaMA model at a low cost.

In stage 1, the SFT model is trained:

# Training with a 4-GPU servers
colossalai run — nproc_per_node=4 train_sft.py \
— pretrain “/path/to/LLaMa-7B/” \
— model ‘llama’ \
— strategy colossalai_zero2 \
— log_interval 10 \
— save_path /path/to/Coati-7B \
— dataset /path/to/data.json \
— batch_size 4 \
— accimulation_steps 8 \
— lr 2e-5

In stage 2, the RM is trained:

# Training with a 4-GPU servers
colossalai run — nproc_per_node=4 train_reward_model.py \
— pretrain “/path/to/LLaMa-7B/” \
— model ‘llama’ \
— strategy colossalai_zero2 \
— dataset /path/to/datasets

In stage 3, training with RL algorithm:

# Training with a 8-GPU servers
colossalai run — nproc_per_node=8 train_prompts.py prompts.csv \
— strategy colossalai_zero2 \
— pretrain “/path/to/Coati-7B” \
— model ‘llama’ \
— pretrain_dataset /path/to/dataset

Once the fine-tuned model weights have been obtained, hardware cost for inference can be reduced through quantization, and online inference services can be launched, requiring only a single GPU with approximately 4GB memory to deploy the 7 billion parameter model inference service.

python server.py /path/to/pretrained — quant 4bit — gptq_checkpoint /path/to/coati-7b-4bit-128g.pt — gptq_group_size 128

System Performance Optimization and Development Acceleration

ColossalChat’s ability to quickly follow the complete RLHF process of ChatGPT replication is largely due to the underlying support from the infrastructure of Colossal-AI and related optimization technologies. Under the same conditions, ColossalChat’s training speed can be improved by almost three times compared with FSDP (Fully Sharded Data Parallel) used by Alpaca.

System Infrastructure Colossal-AI

The AI large model development system Colossal-AI provides the foundational support for this project. It can efficiently and quickly deploy AI large model training and inference based on default PyTorch functionality, reducing the cost of large AI model applications. Colossal-AI is developed based on the expertise of Prof. James Demmel, the Distinguished Professor at UC Berkeley, and Prof. You Yang, the President Young Professor at the National University of Singapore. Since its open source release, Colossal-AI has ranked first on the GitHub Trending multiple times with about 20,000 GitHub stars, and has successfully been accepted as the official tutorial for international AI and HPC top conferences such as SC, AAAI, PPoPP, CVPR, and ISC.

Zero+Gemini to Reduce Memory Redundancy

Colossal-AI supports ZeRO (Zero Redundancy Optimizer) to improve memory usage efficiency, enabling larger models to be accommodated at a lower cost, without affecting computing granularity and communication efficiency. The automatic chunk mechanism can further improve ZeRO’s performance by increasing memory usage efficiency, reducing communication frequency, and avoiding memory fragmentation. The heterogeneous memory space manager, Gemini, supports unloading optimizer states from GPU memory to CPU memory or hard disk space to overcome the limitations of GPU memory capacity, expand the scale of trainable models, and reduce the cost of large AI model applications.

Low-cost Fine-tuning of LoRA

Colossal-AI includes the Low-Rank Adaptation (LoRA) method for low-cost fine-tuning of large models. The LoRA method assumes that large language models are over-parameterized and that the parameter change during fine-tuning is a low-rank matrix. Therefore, this matrix can be decomposed into the product of two smaller matrices. During fine-tuning, the parameters of the large model are fixed, and only the parameters of the low-rank matrix are adjusted, significantly reducing the number of parameters required for training and lowering the cost.

Low-cost Quantized Inference

GPTQ quantization

To reduce the cost of inference deployment, Colossal-AI uses GPTQ 4-bit quantized inference. On GPT/OPT/BLOOM models, it can achieve better Perplexity results than traditional RTN (round-to-nearest) quantization techniques. Compared to common FP16 inference, it can reduce memory consumption by 75% while only sacrificing a small amount of throughput speed and Perplexity performance.

For instance, with ColossalChat-7B, using 4-bit quantized inference, the 7 billion parameter model only requires about 4GB of GPU memory to complete short sequence (128-length generation) inference, which can be done on a common consumer-grade GPU like the RTX 3060 with just one line of code.

if args.quant == ‘4bit’:
model = load_quant(args.pretrained, args.gptq_checkpoint, 4, args.gptq_group_size)

If efficient asynchronous offloading technology is used, the memory requirements can be further reduced, enabling larger models to be inferred on lower-cost hardware.

ColossalChat vs. Alpaca

  1. ColossalChat is the first to open source a complete RLHF pipeline, while Stanford’s Alpaca has not implemented RLHF, which means they didn’t include Stage 2 and Stage 3.
  2. ColossalChat demonstrates superior performance and broader conversational coverage. Its significant improvements are due to the utilization of a larger and higher quality dataset, along with the implementation of reinforcement learning to align responses more closely with human-like answers.
  3. ColossalChat’s training process incorporates various system optimizations from Colossal-AI, resulting in faster training times of about three times compared to Alpaca when using the same dataset and model size. This enables researchers and small to medium-sized enterprises to independently train and deploy their own chatbots.
  4. The ColossalChat team has collected a larger dataset for training, consisting of approximately 24 million tokens for English and 30 million tokens for Chinese, resulting in a total of around 54 million tokens. Notably, ColossalChat collected 6 million tokens for English and 18 million tokens for Chinese independently.

The following are some of the performance comparisons between ColossalChat and Alpaca in language dialogues.

Write a Quicksort in Python
Write an email to a professor for a recommendation letter

Limitation

Although RLHF has been further introduced, there is still room to improve the actual performance in some scenarios due to the limited computing power and data set.

Collaboration

Luckily, unlike previous large AI models and cutting-edge technologies that were monopolized by only a few tech giants, open-source communities and startups such as PyTorch, Hugging Face, and OpenAI have also played a key role in this wave. Drawing on the successful experience of the open-source community, Colossal-AI welcomes all parties to participate in building together and embracing the era of large models!

  • You can post an issue or submit a pull request (PR).
  • Join the Colossal-AI WeChat or Slack group to communicate with the team and other users.
  • Send your official proposal to email youy@comp.nus.edu.sg

Acknowledgments

ColossalChat owes a great deal of gratitude to many existing works and outstanding organizations. The incredible Stanford Alpaca project has been a source of inspiration. The Self-Instruct research paper provides the foundation for the powerful capabilities of small datasets. Accurate post-training quantization comes from GPTQ. Thanks to Meta AI Research for releasing the LLaMA models, Meta’s PyTorch, and OpenAI for paving the way for the most powerful AI.

Disclaimer

Similar to Stanford Alpaca, we emphasize that ColossalChat is a contribution to the open-source community, which is intended solely for academic research purposes and any commercial use is prohibited:

  1. ColossalChat is built upon LLaMA, which is licensed for non-commercial use only.
  2. The instruction data derived from OpenAI’s model API, and the terms of use for this data prohibit the development of competing models.
  3. ColossalChat, like other large language models, may exhibit several common deficiencies, including hallucination, toxicity, and bias.

Reference

[1] Wang, Yizhong, et al. “Self-Instruct: Aligning Language Model with Self Generated Instructions.” arXiv preprint arXiv:2212.10560 (2022).

[2] Touvron, Hugo, et al. “LLaMA: Open and efficient foundation language models.” arXiv preprint arXiv:2302.13971 (2023).

[3] Rohan, Taori, et al. “Stanford Alpaca: An Instruction-following LLaMA model.” arXiv preprint arXiv:2302.13971 (2023).

[4] Hu, Edward J., et al. “Lora: Low-rank adaptation of large language models.” arXiv preprint arXiv:2106.09685 (2021).

[5] Frantar, Elias, et al. “GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers.” arXiv preprint arXiv:2210.17323 (2022).

[6] OpenAI. 2022. ChatGPT. https://openai.com/blog/chatgpt

[7] Rajbhandari, Samyam, et al. “Zero: Memory optimizations toward training trillion parameter models.” SC20: International Conference for High Performance Computing, Networking, Storage and Analysis. IEEE, 2020.

Read more
ml-in-6-months
03/152023

6-Month Roadmap to Becoming a Machine Learning Engineer for Free

footer #ml

6-Month Roadmap to Becoming a Machine Learning Engineer for Free

(Thanks to Brij kishore Pandey)

19 free lessons to get you interview-ready and move ahead of 90% of people.

Follow these steps in the specified order to ensure success:

𝗠𝗼𝗻𝘁𝗵 𝟭: 𝗠𝗮𝘁𝗵𝗲𝗺𝗮𝘁𝗶𝗰𝘀 & 𝗦𝘁𝗮𝘁𝗶𝘀𝘁𝗶𝗰𝘀

Weeks 1-2: Study Linear Algebra concepts – https://lnkd.in/eabKGp_p

Weeks 3-4: Continue with Calculus and Probability & Statistics.
Practice problems to solidify your understanding – https://lnkd.in/ea2DmZ2d

𝗠𝗼𝗻𝘁𝗵 𝟮: 𝗦𝗤𝗟 & 𝗗𝗮𝘁𝗮𝗯𝗮𝘀𝗲𝘀 𝗮𝗻𝗱 𝗠𝗼𝗿𝗲 𝗦𝘁𝗮𝘁𝗶𝘀𝘁𝗶𝗰𝘀

Weeks 1-2: Learn SQL basics – https://lnkd.in/ea2DmZ2d

Weeks 3-4: Continue studying Probability & Statistics. Apply statistical concepts in SQL where possible.

𝗠𝗼𝗻𝘁𝗵 𝟯: 𝗖𝗼𝗿𝗲 𝗖𝗼𝗻𝗰𝗲𝗽𝘁𝘀 𝗼𝗳 𝗠𝗮𝗰𝗵𝗶𝗻𝗲 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴

Weeks 1-2: Go through Google’s ML Crash Course – https://lnkd.in/eT7NiGp6

Weeks 3-4: Go through Andrew Ng’s ML Course – https://lnkd.in/e964AiC7

𝗠𝗼𝗻𝘁𝗵 𝟰: 𝗣𝗿𝗼𝗴𝗿𝗮𝗺𝗺𝗶𝗻𝗴 𝗦𝗸𝗶𝗹𝗹𝘀 (𝗣𝘆𝘁𝗵𝗼𝗻 & 𝗟𝗶𝗯𝗿𝗮𝗿𝗶𝗲𝘀)

Weeks 1-2: Learn Python basics. https://lnkd.in/euyfHHxa

Weeks 3-4: Start with Python libraries for Machine Learning.

• Scikit-learn – https://lnkd.in/eqFhCwXt
• TensorFlow – https://lnkd.in/e6RWbe9h
• PyTorch – https://lnkd.in/efhPxZPM

𝗠𝗼𝗻𝘁𝗵 𝟱: 𝗠𝗼𝗱𝗲𝗹 𝗧𝗿𝗮𝗶𝗻𝗶𝗻𝗴 & 𝗧𝘂𝗻𝗶𝗻𝗴 𝗧𝗲𝗰𝗵𝗻𝗶𝗾𝘂𝗲𝘀, 𝗔𝗱𝘃𝗮𝗻𝗰𝗲𝗱 𝗗𝗲𝗲𝗽 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴 𝗠𝗼𝗱𝗲𝗹𝘀

Weeks 1-2: Learn about model training and tuning techniques.

• Intermediate ML – https://lnkd.in/e89AmkzE
• Hyperparameter Tuning – https://lnkd.in/ezEnqeG2

Weeks 3-4: Start with Advanced Deep Learning Models.

• Stanford’s CS231n (CNNs) – http://cs231n.github.io/
• Deep Learning Book – https://lnkd.in/e_utEgZM

𝗠𝗼𝗻𝘁𝗵 𝟲: 𝗗𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁, 𝗠𝗼𝗻𝗶𝘁𝗼𝗿𝗶𝗻𝗴, & 𝗠𝗮𝗶𝗻𝘁𝗲𝗻𝗮𝗻𝗰𝗲 𝗮𝗻𝗱 𝗥𝗲𝘀𝘂𝗺𝗲 𝗣𝗿𝗲𝗽𝗮𝗿𝗮𝘁𝗶𝗼𝗻, 𝗦𝗼𝗳𝘁 𝗦𝗸𝗶𝗹𝗹𝘀 & 𝗧𝗶𝗽𝘀

Weeks 1-2: Learn about deployment, monitoring, and maintenance.

• Docker – https://lnkd.in/esXHzx9k
• Git – https://lnkd.in/esQ8FMxS
• AWS ML – https://lnkd.in/eZcdQPee
• Azure ML – https://lnkd.in/e5fvmvtk

Weeks 3-4: Prepare your resume and improve your soft skills Resume and Soft Skills & Tips, and work on projects.

• 217 Machine Learning Projects – https://lnkd.in/e5kyv3Tv

Set realistic goals.

Practice is key — so work on projects and apply your knowledge to real-world problems for the best learning experience.

Don’t try to learn everything about machine learning in 6 months.

Focus on learning the basics and then start working on your own projects.

Machine learning is a fascinating field with endless possibilities.

Read more
the_incredible_pytorch
03/132023

The Incredible Pytorch

footer pytorch

Pytorch is an open source machine learning framework that accelerates the path from research prototyping to production deployment.
Here are some curated lists of tutorials, projects, libraries, videos, papers, books, and valuable resources to help sharpen your tools.

Thanks to Charafeddine Mouzouni for this list


This is a curated list of tutorials, projects, libraries, videos, papers, books and anything related to the incredible PyTorch. Feel free to make a pull request to contribute to this list.

Table Of Contents

  • Tabular Data
  • Tutorials
  • Visualization
  • Explainability
  • Object Detection
  • Long-Tailed / Out-of-Distribution Recognition
  • Activation Functions
  • Energy-Based Learning
  • Missing Data
  • Architecture Search
  • Continual Learning
  • Optimization
  • Quantization
  • Quantum Machine Learning
  • Neural Network Compression
  • Facial, Action and Pose Recognition
  • Super resolution
  • Synthetesizing Views
  • Voice
  • Medical
  • 3D Segmentation, Classification and Regression
  • Video Recognition
  • Recurrent Neural Networks (RNNs)
  • Convolutional Neural Networks (CNNs)
  • Segmentation
  • Geometric Deep Learning: Graph & Irregular Structures
  • Sorting
  • Ordinary Differential Equations Networks
  • Multi-task Learning
  • GANs, VAEs, and AEs
  • Unsupervised Learning
  • Adversarial Attacks
  • Style Transfer
  • Image Captioning
  • Transformers
  • Similarity Networks and Functions
  • Reasoning
  • General NLP
  • Question and Answering
  • Speech Generation and Recognition
  • Document and Text Classification
  • Text Generation
  • Text to Image
  • Translation
  • Sentiment Analysis
  • Deep Reinforcement Learning
  • Deep Bayesian Learning and Probabilistic Programmming
  • Spiking Neural Networks
  • Anomaly Detection
  • Regression Types
  • Time Series
  • Synthetic Datasets
  • Neural Network General Improvements
  • DNN Applications in Chemistry and Physics
  • New Thinking on General Neural Network Architecture
  • Linear Algebra
  • API Abstraction
  • Low Level Utilities
  • PyTorch Utilities
  • PyTorch Video Tutorials
  • Datasets
  • Community
  • Links to This Repository
  • To be Classified
  • Contributions

Tabular Data

  • PyTorch-TabNet: Attentive Interpretable Tabular Learning
  • carefree-learn: A minimal Automatic Machine Learning (AutoML) solution for tabular datasets based on PyTorch

Tutorials

  • Official PyTorch Tutorials
  • Official PyTorch Examples
  • Dive Into Deep Learning with PyTorch
  • Minicourse in Deep Learning with PyTorch (Multi-language)
  • Practical Deep Learning with PyTorch
  • Deep Learning Models
  • C++ Implementation of PyTorch Tutorial
  • Simple Examples to Introduce PyTorch
  • Mini Tutorials in PyTorch
  • Deep Learning for NLP
  • Deep Learning Tutorial for Researchers
  • Fully Convolutional Networks implemented with PyTorch
  • Simple PyTorch Tutorials Zero to ALL
  • DeepNLP-models-Pytorch
  • MILA PyTorch Welcome Tutorials
  • Effective PyTorch, Optimizing Runtime with TorchScript and Numerical Stability Optimization
  • Practical PyTorch
  • PyTorch Project Template
  • Semantic Search with PyTorch

Visualization

  • Loss Visualization
  • Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization
  • Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps
  • SmoothGrad: removing noise by adding noise
  • DeepDream: dream-like hallucinogenic visuals
  • FlashTorch: Visualization toolkit for neural networks in PyTorch
  • Lucent: Lucid adapted for PyTorch
  • DreamCreator: Training GoogleNet models for DeepDream with custom datasets made simple
  • CNN Feature Map Visualisation

Explainability

  • Neural-Backed Decision Trees
  • Efficient Covariance Estimation from Temporal Data
  • Hierarchical interpretations for neural network predictions
  • Shap, a unified approach to explain the output of any machine learning model
  • VIsualizing PyTorch saved .pth deep learning models with netron
  • Distilling a Neural Network Into a Soft Decision Tree
  • Captum, A unified model interpretability library for PyTorch

Object Detection

  • MMDetection Object Detection Toolbox
  • Mask R-CNN Benchmark: Faster R-CNN and Mask R-CNN in PyTorch 1.0
  • YOLOS
  • YOLOF
  • YOLOX
  • Yolov7
  • YOLOv6
  • Yolov5
  • Yolov4
  • YOLOv3
  • YOLOv2: Real-Time Object Detection
  • SSD: Single Shot MultiBox Detector
  • Detectron models for Object Detection
  • Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks
  • Whale Detector
  • Catalyst.Detection

Long-Tailed / Out-of-Distribution Recognition

  • Distributionally Robust Neural Networks for Group Shifts: On the Importance of Regularization for Worst-Case Generalization
  • Invariant Risk Minimization
  • Training Confidence-Calibrated Classifier for Detecting Out-of-Distribution Samples
  • Deep Anomaly Detection with Outlier Exposure
  • Large-Scale Long-Tailed Recognition in an Open World
  • Principled Detection of Out-of-Distribution Examples in Neural Networks
  • Learning Confidence for Out-of-Distribution Detection in Neural Networks
  • PyTorch Imbalanced Class Sampler

Activation Functions

  • Rational Activations – Learnable Rational Activation Functions

Energy-Based Learning

  • EBGAN, Energy-Based GANs
  • Maximum Entropy Generators for Energy-based Models

Missing Data

  • BRITS: Bidirectional Recurrent Imputation for Time Series

Architecture Search

  • EfficientNetV2
  • DenseNAS
  • DARTS: Differentiable Architecture Search
  • Efficient Neural Architecture Search (ENAS)
  • EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks

Continual Learning

  • Renate, Automatic Retraining of Neural Networks

Optimization

  • AccSGD, AdaBound, AdaMod, DiffGrad, Lamb, NovoGrad, RAdam, SGDW, Yogi and more
  • Lookahead Optimizer: k steps forward, 1 step back
  • RAdam, On the Variance of the Adaptive Learning Rate and Beyond
  • Over9000, Comparison of RAdam, Lookahead, Novograd, and combinations
  • AdaBound, Train As Fast as Adam As Good as SGD
  • Riemannian Adaptive Optimization Methods
  • L-BFGS
  • OptNet: Differentiable Optimization as a Layer in Neural Networks
  • Learning to learn by gradient descent by gradient descent
  • Surrogate Gradient Learning in Spiking Neural Networks
  • TorchOpt: An Efficient Library for Differentiable Optimization

Quantization

  • Additive Power-of-Two Quantization: An Efficient Non-uniform Discretization For Neural Networks

Quantum Machine Learning

  • Tor10, generic tensor-network library for quantum simulation in PyTorch
  • PennyLane, cross-platform Python library for quantum machine learning with PyTorch interface

Neural Network Compression

  • Bayesian Compression for Deep Learning
  • Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research
  • Learning Sparse Neural Networks through L0 regularization
  • Energy-constrained Compression for Deep Neural Networks via Weighted Sparse Projection and Layer Input Masking
  • EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis
  • Pruning Convolutional Neural Networks for Resource Efficient Inference
  • Pruning neural networks: is it time to nip it in the bud? (showing reduced networks work better)

Facial, Action and Pose Recognition

  • Facenet: Pretrained Pytorch face detection and recognition models
  • DGC-Net: Dense Geometric Correspondence Network
  • High performance facial recognition library on PyTorch
  • FaceBoxes, a CPU real-time face detector with high accuracy
  • How far are we from solving the 2D & 3D Face Alignment problem? (and a dataset of 230,000 3D facial landmarks)
  • Learning Spatio-Temporal Features with 3D Residual Networks for Action Recognition
  • PyTorch Realtime Multi-Person Pose Estimation
  • SphereFace: Deep Hypersphere Embedding for Face Recognition
  • GANimation: Anatomically-aware Facial Animation from a Single Image
  • Shufflenet V2 by Face++ with better results than paper
  • Towards 3D Human Pose Estimation in the Wild: a Weakly-supervised Approach
  • Unsupervised Learning of Depth and Ego-Motion from Video
  • FlowNet 2.0: Evolution of Optical Flow Estimation with Deep Networks
  • FlowNet: Learning Optical Flow with Convolutional Networks
  • Optical Flow Estimation using a Spatial Pyramid Network
  • OpenFace in PyTorch
  • Deep Face Recognition in PyTorch

Super resolution

  • Enhanced Deep Residual Networks for Single Image Super-Resolution
  • Superresolution using an efficient sub-pixel convolutional neural network
  • Perceptual Losses for Real-Time Style Transfer and Super-Resolution

Synthetesizing Views

  • NeRF, Neural Radian Fields, Synthesizing Novels Views of Complex Scenes

Voice

  • Google AI VoiceFilter: Targeted Voice Separatation by Speaker-Conditioned Spectrogram Masking

Medical

  • Medical Zoo, 3D multi-modal medical image segmentation library in PyTorch
  • U-Net for FLAIR Abnormality Segmentation in Brain MRI
  • Genomic Classification via ULMFiT
  • Deep Neural Networks Improve Radiologists’ Performance in Breast Cancer Screening
  • Delira, lightweight framework for medical imaging prototyping
  • V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation
  • Medical Torch, medical imaging framework for PyTorch
  • TorchXRayVision – A library for chest X-ray datasets and models. Including pre-trainined models.

3D Segmentation, Classification and Regression

  • Kaolin, Library for Accelerating 3D Deep Learning Research
  • PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation
  • 3D segmentation with MONAI and Catalyst

Video Recognition

  • Dancing to Music
  • Devil Is in the Edges: Learning Semantic Boundaries from Noisy Annotations
  • Deep Video Analytics
  • PredRNN: Recurrent Neural Networks for Predictive Learning using Spatiotemporal LSTMs

Recurrent Neural Networks (RNNs)

  • SRU: training RNNs as fast as CNNs
  • Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks
  • Averaged Stochastic Gradient Descent with Weight Dropped LSTM
  • Training RNNs as Fast as CNNs
  • Quasi-Recurrent Neural Network (QRNN)
  • ReSeg: A Recurrent Neural Network-based Model for Semantic Segmentation
  • A Recurrent Latent Variable Model for Sequential Data (VRNN)
  • Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks
  • Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling
  • Attentive Recurrent Comparators
  • Collection of Sequence to Sequence Models with PyTorch
    1. Vanilla Sequence to Sequence models
    2. Attention based Sequence to Sequence models
    3. Faster attention mechanisms using dot products between the final encoder and decoder hidden states

Convolutional Neural Networks (CNNs)

  • LegoNet: Efficient Convolutional Neural Networks with Lego Filters
  • MeshCNN, a convolutional neural network designed specifically for triangular meshes
  • Octave Convolution
  • PyTorch Image Models, ResNet/ResNeXT, DPN, MobileNet-V3/V2/V1, MNASNet, Single-Path NAS, FBNet
  • Deep Neural Networks with Box Convolutions
  • Invertible Residual Networks
  • Stochastic Downsampling for Cost-Adjustable Inference and Improved Regularization in Convolutional Networks
  • Faster Faster R-CNN Implementation
    • Faster R-CNN Another Implementation
  • Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
  • Wide ResNet model in PyTorch –DiracNets: Training Very Deep Neural Networks Without Skip-Connections
  • An End-to-End Trainable Neural Network for Image-based Sequence Recognition and Its Application to Scene Text Recognition
  • Efficient Densenet
  • Video Frame Interpolation via Adaptive Separable Convolution
  • Learning local feature descriptors with triplets and shallow convolutional neural networks
  • Densely Connected Convolutional Networks
  • Very Deep Convolutional Networks for Large-Scale Image Recognition
  • SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size
  • Deep Residual Learning for Image Recognition
  • Training Wide ResNets for CIFAR-10 and CIFAR-100 in PyTorch
  • Deformable Convolutional Network
  • Convolutional Neural Fabrics
  • Deformable Convolutional Networks in PyTorch
  • Dilated ResNet combination with Dilated Convolutions
  • Striving for Simplicity: The All Convolutional Net
  • Convolutional LSTM Network
  • Big collection of pretrained classification models
  • PyTorch Image Classification with Kaggle Dogs vs Cats Dataset
  • CIFAR-10 on Pytorch with VGG, ResNet and DenseNet
  • Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
  • NVIDIA/unsupervised-video-interpolation

Segmentation

  • Detectron2 by FAIR
  • Pixel-wise Segmentation on VOC2012 Dataset using PyTorch
  • Pywick – High-level batteries-included neural network training library for Pytorch
  • Improving Semantic Segmentation via Video Propagation and Label Relaxation
  • Super-BPD: Super Boundary-to-Pixel Direction for Fast Image Segmentation
  • Catalyst.Segmentation
  • Segmentation models with pretrained backbones

Geometric Deep Learning: Graph & Irregular Structures

  • PyTorch Geometric, Deep Learning Extension
  • PyTorch Geometric Temporal: A Temporal Extension Library for PyTorch Geometric
  • PyTorch Geometric Signed Directed: A Signed & Directed Extension Library for PyTorch Geometric
  • ChemicalX: A PyTorch Based Deep Learning Library for Drug Pair Scoring
  • Self-Attention Graph Pooling
  • Position-aware Graph Neural Networks
  • Signed Graph Convolutional Neural Network
  • Graph U-Nets
  • Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks
  • MixHop: Higher-Order Graph Convolutional Architectures via Sparsified Neighborhood Mixing
  • Semi-Supervised Graph Classification: A Hierarchical Graph Perspective
  • PyTorch BigGraph by FAIR for Generating Embeddings From Large-scale Graph Data
  • Capsule Graph Neural Network
  • Splitter: Learning Node Representations that Capture Multiple Social Contexts
  • A Higher-Order Graph Convolutional Layer
  • Predict then Propagate: Graph Neural Networks meet Personalized PageRank
  • Lorentz Embeddings: Learn Continuous Hierarchies in Hyperbolic Space
  • Graph Wavelet Neural Network
  • Watch Your Step: Learning Node Embeddings via Graph Attention
  • Signed Graph Convolutional Network
  • Graph Classification Using Structural Attention
  • SimGNN: A Neural Network Approach to Fast Graph Similarity Computation
  • SINE: Scalable Incomplete Network Embedding
  • HypER: Hypernetwork Knowledge Graph Embeddings
  • TuckER: Tensor Factorization for Knowledge Graph Completion
  • PyKEEN: A Python library for learning and evaluating knowledge graph embeddings
  • Pathfinder Discovery Networks for Neural Message Passing
  • SSSNET: Semi-Supervised Signed Network Clustering
  • MagNet: A Neural Network for Directed Graphs

Sorting

  • Stochastic Optimization of Sorting Networks via Continuous Relaxations

Ordinary Differential Equations Networks

  • Latent ODEs for Irregularly-Sampled Time Series
  • GRU-ODE-Bayes: continuous modelling of sporadically-observed time series

Multi-task Learning

  • Hierarchical Multi-Task Learning Model
  • Task-based End-to-end Model Learning

GANs, VAEs, and AEs

  • BigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis
  • High Fidelity Performance Metrics for Generative Models in PyTorch
  • Mimicry, PyTorch Library for Reproducibility of GAN Research
  • Clean Readable CycleGAN
  • StarGAN
  • Block Neural Autoregressive Flow
  • High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
  • A Style-Based Generator Architecture for Generative Adversarial Networks
  • GANDissect, PyTorch Tool for Visualizing Neurons in GANs
  • Learning deep representations by mutual information estimation and maximization
  • Variational Laplace Autoencoders
  • VeGANS, library for easily training GANs
  • Progressive Growing of GANs for Improved Quality, Stability, and Variation
  • Conditional GAN
  • Wasserstein GAN
  • Adversarial Generator-Encoder Network
  • Image-to-Image Translation with Conditional Adversarial Networks
  • Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks
  • On the Effects of Batch and Weight Normalization in Generative Adversarial Networks
  • Improved Training of Wasserstein GANs
  • Collection of Generative Models with PyTorch
    • Generative Adversarial Nets (GAN)
      1. Vanilla GAN
      2. Conditional GAN
      3. InfoGAN
      4. Wasserstein GAN
      5. Mode Regularized GAN
    • Variational Autoencoder (VAE)
      1. Vanilla VAE
      2. Conditional VAE
      3. Denoising VAE
      4. Adversarial Autoencoder
      5. Adversarial Variational Bayes
  • Improved Training of Wasserstein GANs
  • CycleGAN and Semi-Supervised GAN
  • Improving Variational Auto-Encoders using Householder Flow and using convex combination linear Inverse Autoregressive Flow
  • PyTorch GAN Collection
  • Generative Adversarial Networks, focusing on anime face drawing
  • Simple Generative Adversarial Networks
  • Adversarial Auto-encoders
  • torchgan: Framework for modelling Generative Adversarial Networks in Pytorch
  • Evaluating Lossy Compression Rates of Deep Generative Models
  • Catalyst.GAN
    1. Vanilla GAN
    2. Conditional GAN
    3. Wasserstein GAN
    4. Improved Training of Wasserstein GANs

Unsupervised Learning

  • Unsupervised Embedding Learning via Invariant and Spreading Instance Feature
  • AND: Anchor Neighbourhood Discovery

Adversarial Attacks

  • Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images
  • Explaining and Harnessing Adversarial Examples
  • AdverTorch – A Toolbox for Adversarial Robustness Research

Style Transfer

  • Pystiche: Framework for Neural Style Transfer
  • Detecting Adversarial Examples via Neural Fingerprinting
  • A Neural Algorithm of Artistic Style
  • Multi-style Generative Network for Real-time Transfer
  • DeOldify, Coloring Old Images
  • Neural Style Transfer
  • Fast Neural Style Transfer
  • Draw like Bob Ross

Image Captioning

  • CLIP (Contrastive Language-Image Pre-Training)
  • Neuraltalk 2, Image Captioning Model, in PyTorch
  • Generate captions from an image with PyTorch
  • DenseCap: Fully Convolutional Localization Networks for Dense Captioning

Transformers

  • Attention is all you need
  • Spatial Transformer Networks

Similarity Networks and Functions

  • Conditional Similarity Networks

Reasoning

  • Inferring and Executing Programs for Visual Reasoning

General NLP

  • nanoGPT, fastest repository for training/finetuning medium-sized GPTs
  • minGPT, Re-implementation of GPT to be small, clean, interpretable and educational
  • Espresso, Module Neural Automatic Speech Recognition Toolkit
  • Label-aware Document Representation via Hybrid Attention for Extreme Multi-Label Text Classification
  • XLNet
  • Conversing by Reading: Contentful Neural Conversation with On-demand Machine Reading
  • Cross-lingual Language Model Pretraining
  • Libre Office Translate via PyTorch NMT
  • BERT
  • VSE++: Improved Visual-Semantic Embeddings
  • A Structured Self-Attentive Sentence Embedding
  • Neural Sequence labeling model
  • Skip-Thought Vectors
  • Complete Suite for Training Seq2Seq Models in PyTorch
  • MUSE: Multilingual Unsupervised and Supervised Embeddings
  • TorchMoji: PyTorch Implementation of DeepMoji to under Language used to Express Emotions

Question and Answering

  • Visual Question Answering in Pytorch
  • Reading Wikipedia to Answer Open-Domain Questions
  • Deal or No Deal? End-to-End Learning for Negotiation Dialogues
  • Interpretable Counting for Visual Question Answering
  • Open Source Chatbot with PyTorch

Speech Generation and Recognition

  • PyTorch-Kaldi Speech Recognition Toolkit
  • WaveGlow: A Flow-based Generative Network for Speech Synthesis
  • OpenNMT
  • Deep Speech 2: End-to-End Speech Recognition in English and Mandarin
  • WeNet: Production First and Production Ready End-to-End Speech Recognition Toolkit

Document and Text Classification

  • Hierarchical Attention Network for Document Classification
  • Hierarchical Attention Networks for Document Classification
  • CNN Based Text Classification

Text Generation

  • Pytorch Poetry Generation

Text to Image

  • Stable Diffusion
  • Dall-E 2
  • Dall-E

Translation

  • Open-source (MIT) Neural Machine Translation (NMT) System

Sentiment Analysis

  • Recurrent Neural Networks for Sentiment Analysis (Aspect-Based) on SemEval 2014
  • Seq2Seq Intent Parsing
  • Finetuning BERT for Sentiment Analysis

Deep Reinforcement Learning

  • Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels
  • Exploration by Random Network Distillation
  • EGG: Emergence of lanGuage in Games, quickly implement multi-agent games with discrete channel communication
  • Temporal Difference VAE
  • High-performance Atari A3C Agent in 180 Lines PyTorch
  • Learning when to communicate at scale in multiagent cooperative and competitive tasks
  • Actor-Attention-Critic for Multi-Agent Reinforcement Learning
  • PPO in PyTorch C++
  • Reinforcement Learning for Bandit Neural Machine Translation with Simulated Human Feedback
  • Asynchronous Methods for Deep Reinforcement Learning
  • Continuous Deep Q-Learning with Model-based Acceleration
  • Asynchronous Methods for Deep Reinforcement Learning for Atari 2600
  • Trust Region Policy Optimization
  • Neural Combinatorial Optimization with Reinforcement Learning
  • Noisy Networks for Exploration
  • Distributed Proximal Policy Optimization
  • Reinforcement learning models in ViZDoom environment with PyTorch
  • Reinforcement learning models using Gym and Pytorch
  • SLM-Lab: Modular Deep Reinforcement Learning framework in PyTorch
  • Catalyst.RL

Deep Bayesian Learning and Probabilistic Programmming

  • BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning
  • Subspace Inference for Bayesian Deep Learning
  • Bayesian Deep Learning with Variational Inference Package
  • Probabilistic Programming and Statistical Inference in PyTorch
  • Bayesian CNN with Variational Inferece in PyTorch

Spiking Neural Networks

  • Norse, Library for Deep Learning with Spiking Neural Networks

Anomaly Detection

  • Detection of Accounting Anomalies using Deep Autoencoder Neural Networks

Regression Types

  • Quantile Regression DQN

Time Series

  • Dual Self-Attention Network for Multivariate Time Series Forecasting
  • DILATE: DIstortion Loss with shApe and tImE
  • Variational Recurrent Autoencoder for Timeseries Clustering
  • Spatio-Temporal Neural Networks for Space-Time Series Modeling and Relations Discovery
  • Flow Forecast: A deep learning for time series forecasting framework built in PyTorch

Synthetic Datasets

  • Meta-Sim: Learning to Generate Synthetic Datasets

Neural Network General Improvements

  • In-Place Activated BatchNorm for Memory-Optimized Training of DNNs
  • Train longer, generalize better: closing the generalization gap in large batch training of neural networks
  • FreezeOut: Accelerate Training by Progressively Freezing Layers
  • Binary Stochastic Neurons
  • Compact Bilinear Pooling
  • Mixed Precision Training in PyTorch

DNN Applications in Chemistry and Physics

  • Wave Physics as an Analog Recurrent Neural Network
  • Neural Message Passing for Quantum Chemistry
  • Automatic chemical design using a data-driven continuous representation of molecules
  • Deep Learning for Physical Processes: Integrating Prior Scientific Knowledge
  • Differentiable Molecular Simulation for Learning and Control

New Thinking on General Neural Network Architecture

  • Complement Objective Training
  • Decoupled Neural Interfaces using Synthetic Gradients

Linear Algebra

  • Eigenvectors from Eigenvalues

API Abstraction

  • Torch Layers, Shape inference for PyTorch, SOTA Layers
  • Hummingbird, run trained scikit-learn models on GPU with PyTorch

Low Level Utilities

  • TorchSharp, .NET API with access to underlying library powering PyTorch

PyTorch Utilities

  • Functorch: prototype of JAX-like composable Function transformers for PyTorch
  • Poutyne: Simplified Framework for Training Neural Networks
  • PyTorch Metric Learning
  • Kornia: an Open Source Differentiable Computer Vision Library for PyTorch
  • BackPACK to easily Extract Variance, Diagonal of Gauss-Newton, and KFAC
  • PyHessian for Computing Hessian Eigenvalues, trace of matrix, and ESD
  • Hessian in PyTorch
  • Differentiable Convex Layers
  • Albumentations: Fast Image Augmentation Library
  • Higher, obtain higher order gradients over losses spanning training loops
  • Neural Pipeline, Training Pipeline for PyTorch
  • Layer-by-layer PyTorch Model Profiler for Checking Model Time Consumption
  • Sparse Distributions
  • Diffdist, Adds Support for Differentiable Communication allowing distributed model parallelism
  • HessianFlow, Library for Hessian Based Algorithms
  • Texar, PyTorch Toolkit for Text Generation
  • PyTorch FLOPs counter
  • PyTorch Inference on C++ in Windows
  • EuclidesDB, Multi-Model Machine Learning Feature Database
  • Data Augmentation and Sampling for Pytorch
  • PyText, deep learning based NLP modelling framework officially maintained by FAIR
  • Torchstat for Statistics on PyTorch Models
  • Load Audio files directly into PyTorch Tensors
  • Weight Initializations
  • Spatial transformer implemented in PyTorch
  • PyTorch AWS AMI, run PyTorch with GPU support in less than 5 minutes
  • Use tensorboard with PyTorch
  • Simple Fit Module in PyTorch, similar to Keras
  • torchbearer: A model fitting library for PyTorch
  • PyTorch to Keras model converter
  • Gluon to PyTorch model converter with code generation
  • Catalyst: High-level utils for PyTorch DL & RL research
  • PyTorch Lightning: Scalable and lightweight deep learning research framework
  • Determined: Scalable deep learning platform with PyTorch support
  • PyTorch-Ignite: High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently
  • torchvision: A package consisting of popular datasets, model architectures, and common image transformations for computer vision.
  • Poutyne: A Keras-like framework for PyTorch and handles much of the boilerplating code needed to train neural networks.
  • torchensemble: Scikit-Learn like ensemble methods in PyTorch

PyTorch Video Tutorials

  • PyTorch Zero to All Lectures
  • PyTorch For Deep Learning Full Course
  • [PyTorch Lightning 101 with Alfredo Canziani and William Falcon](https://www.you tube.com/playlist?list=PLaMu-SDt_RB5NUm67hU2pdE75j6KaIOv2)
  • Practical Deep Learning with PyTorch

Datasets

  • Worldbank Data

Community

  • PyTorch Discussion Forum
  • StackOverflow PyTorch Tags
  • Catalyst.Slack

Links to This Repository

  • Github Repository
  • Website

Read more
02/182023

AWS Code Whisper overview

footer

CodeWhisperer is trained on billions of lines of code and can generate code suggestions ranging from snippets to full functions in real time based on your comments and existing code. Bypass time-consuming coding tasks and accelerate building with unfamiliar APIs.

 

Code with confidence

CodeWhisperer can flag or filter code suggestions that resemble open-source training data. Get the associated open-source project’s repository URL and license so that you can more easily review them and add attribution.

Enhance code security

Scan your code to detect hard-to-find vulnerabilities, and get code suggestions to remediate them immediately. Align to best practices for tackling security vulnerabilities, such as those outlined by Open Worldwide Application Security Project (OWASP), or those that don’t meet crypto library best practices and other similar security best practices.

Check it out at https://aws.amazon.com/codewhisperer/

Read more
01/152023

AI tools – Jan ’23

footer #ai #aitools

Try these 18 powerful AI tools to increase your productivity. Jan ’23 Edition

Read more
09/042022

Types of machine learning Algorithms

footer

Types of machine learning Algorithms

There some variations of how to define the types of Machine Learning Algorithms but commonly they can be divided into categories according to their purpose and the main categories are the following:

  • Supervised learning
  • Unsupervised Learning
  • Semi-supervised Learning
  • Reinforcement Learning

Supervised Learning

  • I like to think of supervised learning with the concept of function approximation, where basically we train an algorithm and in the end of the process we pick the function that best describes the input data, the one that for a given X makes the best estimation of y (X -> y). Most of the time we are not able to figure out the true function that always make the correct predictions and other reason is that the algorithm rely upon an assumption made by humans about how the computer should learn and this assumptions introduce a bias, Bias is topic I’ll explain in another post.
  • Here the human experts acts as the teacher where we feed the computer with training data containing the input/predictors and we show it the correct answers (output) and from the data the computer should be able to learn the patterns.
  • Supervised learning algorithms try to model relationships and dependencies between the target prediction output and the input features such that we can predict the output values for new data based on those relationships which it learned from the previous data sets.

Draft

  • Predictive Model
  • we have labeled data
  • The main types of supervised learning problems include regression and classification problems

List of Common Algorithms

  • Nearest Neighbor
  • Naive Bayes
  • Decision Trees
  • Linear Regression
  • Support Vector Machines (SVM)
  • Neural Networks

Unsupervised Learning

  • The computer is trained with unlabeled data.
  • Here there’s no teacher at all, actually the computer might be able to teach you new things after it learns patterns in data, these algorithms a particularly useful in cases where the human expert doesn’t know what to look for in the data.
  • are the family of machine learning algorithms which are mainly used in pattern detection and descriptive modeling. However, there are no output categories or labels here based on which the algorithm can try to model relationships. These algorithms try to use techniques on the input data to mine for rules, detect patterns, and summarize and group the data points which help in deriving meaningful insights and describe the data better to the users.

Draft

  • Descriptive Model
  • The main types of unsupervised learning algorithms include Clustering algorithms and Association rule learning algorithms.

List of Common Algorithms

  • k-means clustering, Association Rules

Semi-supervised Learning

In the previous two types, either there are no labels for all the observation in the dataset or labels are present for all the observations. Semi-supervised learning falls in between these two. In many practical situations, the cost to label is quite high, since it requires skilled human experts to do that. So, in the absence of labels in the majority of the observations but present in few, semi-supervised algorithms are the best candidates for the model building. These methods exploit the idea that even though the group memberships of the unlabeled data are unknown, this data carries important information about the group parameters.

Reinforcement Learning

method aims at using observations gathered from the interaction with the environment to take actions that would maximize the reward or minimize the risk. Reinforcement learning algorithm (called the agent) continuously learns from the environment in an iterative fashion. In the process, the agent learns from its experiences of the environment until it explores the full range of possible states.

Reinforcement Learning is a type of Machine Learning, and thereby also a branch of Artificial Intelligence. It allows machines and software agents to automatically determine the ideal behavior within a specific context, in order to maximize its performance. Simple reward feedback is required for the agent to learn its behavior; this is known as the reinforcement signal.

There are many different algorithms that tackle this issue. As a matter of fact, Reinforcement Learning is defined by a specific type of problem, and all its solutions are classed as Reinforcement Learning algorithms. In the problem, an agent is supposed decide the best action to select based on his current state. When this step is repeated, the problem is known as a Markov Decision Process.

In order to produce intelligent programs (also called agents), reinforcement learning goes through the following steps:

  1. Input state is observed by the agent.
  2. Decision making function is used to make the agent perform an action.
  3. After the action is performed, the agent receives reward or reinforcement from the environment.
  4. The state-action pair information about the reward is stored.

List of Common Algorithms

  • Q-Learning
  • Temporal Difference (TD)
  • Deep Adversarial Networks

Use cases:

Some applications of the reinforcement learning algorithms are computer played board games (Chess, Go), robotic hands, and self-driving cars.

Final Notes

There is possible to use different criteria to classify types of machine learning algorithms but I think using the learning task is great to visualize the big picture of ML and I believe according to your problem and the data you have in hand you can easily decide if you will use Supervised, unsupervised or reinforcement learning. In the upcoming posts I’ll give more examples about each type of machine learning algorithms.

This image from en.proft.me below might help you.

Read more
photo-29
03/232016

Final Preparations

footer, News preparations 3

It’s amazing how fast this year has gone. After the initial launch of All Day AI last year in it’s pilot form with many speakers presenting in their first online conference, there’s been a lot of discussions about how to improve the conference.

Read more
jeffwilcox-11
03/102016

We are preparing something special

footer, News new, special

New products from vendors, new algorithms, new services from cloud vendors – it’s an every-changing world in AI/ML. This conference will bring the latest techniques that industry leaders have applied to solving complex problems – register today to be able to see what they have done.

Read more
boedker-1
03/022016

What’s new this year?

footer, News new, preparations, registration 1

The latest in Natural Language Processing and Image Recognition are just some of the latest presentations in this years AllDayAI conference

Read more
photo-40
02/252016

Registration is now open! Are you attending?

footer, News attendance, registration

Uniquely productize reconceptualize existing “outside the box” resources globally web-readiness grow experiences. Evolve engineer vortals productivate re-engineer resource-leveling extensible base reinvent base time multimedia other’s. “organic” plug-and-play collaboratively extensible mesh cross-platform stand-alone 2.0 invested potentialities extend deliver theme results.

Read more
Older Entries

  • Code of Conduct
  • Privacy Policy
  • Terms of Service
Latest News
  • ColossalChat: An Open-Source Solution for Cloning ChatGPT With a Complete RLHF Pipeline May 20, 2023
  • 6-Month Roadmap to Becoming a Machine Learning Engineer for Free March 15, 2023
  • The Incredible Pytorch March 13, 2023
  • AWS Code Whisper overview February 18, 2023
Twitter Hightlights

An error has occured.

Copyright Technology Resolution Group, All Rights Reserved