PythonML
Supervised, Unsupervised and Reinforcement Learning
- Python Automation and Machine Learning for ICs -
- An Online Book -
Chapter/Index: Introduction | A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z | Appendix
http://www.globalsino.com/ICs/  


=================================================================================

Table 2412. Comparison between machine learning algorithms (supervised Learning, unsupervised Learning and reinforcement Learning).
  Supervised Learning Unsupervised Learning Reinforcement Learning
Main difference Relies on labelled input and output training data Processes unlabelled or raw data  
  Known Unknown  
  Classification Clustering Clustering
 
Categorical target variable
Customer segmentation
   
 
Keyword analysis [1]
Probabilistic clustering
 
 
K-nearest neighbors
Mean-shift clustering
Association
 
Decision tree
CURE
Robotics
 
Discernment analysis
Deep learning
Business
 
Image classification
K-means
Chemistry
 
Random decision forest
Self-organizing map
Control
 
Hidden Markov model
K-medoids Fuzzy C-means (FCM)
   
 
Multi-layer perception (MLP)
KNN
Classification
 
Routing optimization
Density-based clustering
Optimized marketing
 
Intrusion detection
Biology
   
 
SVM
BIRCH
Control
 
Support vector machines
Chameleon
Driving a vehicle
 
Fraud detection
City planning
 
 
Email spam detection
Targetted marketing
Decision making
 
Diagnostics
Gaussian mixture model
Manufacturing
 
Naïve-Bayes
Hierarchical clustering
Network reconfiguration
 
Traffic classfication
Apriori
Resource management
 
Machines
Gaussian mixture
Q-learning
 
Matching the doctor prescriptions
Hidden Markov
R learning
 
Neural networks
Neural networks
TD learning
 
Discriminant analysis
DBSCAN clustering
Deep adversarial networks
 
Medical imaging
Agglomerative Hierarchical clustering
Multi-armed bandits
   
ANN (artificial neural networks)
Markov decision process
  Regression  
Inventory management
 
Continuous target variable
 
Artificial neural networks
 
Ridge regression
Dimensionality reduction
Robot navigation
 
Housing price prediction
Text mining
Maximizing cumulative survival rates
 
Linear regression
Principal component analysis (PCOA/PCA)
Gaming/game playing
 
Locally weighted R.
Big data visualization
Brute force
 
Logistic regression
Image recognition
Approximate dynamic programming
 
Polynomial/non-linear regression
Factor analysis
Finance sector
 
Generalised
Independent components analysis
 
 
Risk assessment
Face recognition
 
 
Score prediction
SVD
 
 
Stochastic gradient descent
   
 
Random forests
Association  
 
Ensemble methods
Market basket analysis
 
 
Support vector machine (SVM)
   
 
Neural network (NN) regression
   
 
 
Decision trees
   
 
Lasso regression
   
 
Hierarchical
   
 
Ensembles
   
 
 
GLM
   
 
GPR
   
 
SVR
   
   
   
   
 
  Well defined goals Outcome is based only on inputs Start state and end state are defined
  Reverse engineering Outcome-typically clustering or segmentation The agent discovers the path and the relationships on its own
  Example-fraud/non-fraud transactions Machine understands the data (identifies patterns/structures) An approach to AI
  Inventory management Evaluation is qualitative or indirect Reward based learning
  Makes machine learn explicitly Does not predict or find anything specific Learning form reinforcement
  Data with clearly defined output is given Data driven (identify clusters) Machine learns how to act in a certain environment
  Direct feedback is given

Maximize rewards
  Predicts outcome and future
  Learn from mistakes
  Resolves classification and regression problems
   
  Task driven (predict next value)    
Training info Desired (target) output   Evaluations (rewards/penalties)
Output Mapping Classes State/action
Results Desired result (predict, estimate or classify a variable)    
Most frequent applications Some robotics   Less robotics  Robotics*
Research examples [1]    

* The prominence of reinforcement learning (RL) compared to supervised learning and unsupervised learning in the field of robotics can depend on the specific application and the nature of the tasks involved. Howeve, RL has gained significant attention and adoption in the field of robotics because of the reasons below: 

     i) Interactive Learning: Reinforcement learning is well-suited for interactive learning scenarios where an agent (robot) can learn by interacting with its environment, receiving feedback (rewards), and adapting its behavior over time. 

     ii) Adaptability to Dynamic Environments: In robotics, environments can be dynamic and subject to changes. RL excels in scenarios where adaptive decision-making is required to handle dynamic and uncertain conditions. 

     iii) Sequential Decision-Making: Many robotic tasks involve sequential decision-making, where actions taken at one time affect future states and outcomes. RL is designed to handle such sequential decision processes. 

     iv) Ability to Learn From Scratch: RL algorithms can learn from scratch without the need for a large amount of labeled training data. This is particularly beneficial in robotics applications where collecting labeled data may be challenging or impractical. 

viTransfer Learning and Generalization: RL can facilitate transfer learning, allowing knowledge gained in one task or environment to be transferred to another. This is valuable in robotics where generalization across different scenarios is important. Real-Time Adaptation: RL enables real-time adaptation to changing conditions, which is crucial in robotics applications where the robot needs to respond dynamically to its surroundings. While RL has gained traction, supervised learning and unsupervised learning also have their places in robotics: Supervised Learning: It is commonly used in scenarios where labeled data is available and the robot needs to learn specific mappings or associations. Unsupervised Learning: In cases where the robot needs to explore and understand the structure of its environment without explicit supervision. 

     vi) Transfer Learning and Generalization: RL can facilitate transfer learning, allowing knowledge gained in one task or environment to be transferred to another. This is valuable in robotics where generalization across different scenarios is important. 

     vii) Real-Time Adaptation: RL enables real-time adaptation to changing conditions, which is crucial in robotics applications where the robot needs to respond dynamically to its surroundings.

While RL has gained traction, supervised learning and unsupervised learning also have their places in robotics: 

     i) Supervised Learning: It is commonly used in scenarios where labeled data is available and the robot needs to learn specific mappings or associations. 

     ii) Unsupervised Learning: In cases where the robot needs to explore and understand the structure of its environment without explicit supervision. 

Comparison between machine learning algorithms (supervised Learning, unsupervised Learning and reinforcement Learning). Error = target output - actual output

Figure 2412. Comparison between machine learning algorithms (supervised Learning, unsupervised Learning and reinforcement Learning). Error = target output - actual output.

 

 

 

 

       

        

=================================================================================