Get in Touch

Course Outline

1. Grasping Classification via Nearest Neighbors

  • The kNN algorithm
  • Distance calculation
  • Selecting an optimal k value
  • Data preparation for kNN
  • Why is the kNN algorithm considered lazy?

2. Grasping Naive Bayes

  • Fundamental concepts of Bayesian methods
  • Probability
  • Joint probability
  • Conditional probability using Bayes' theorem
  • The Naive Bayes algorithm
  • Naive Bayes classification
  • The Laplace estimator
  • Applying numeric features with Naive Bayes

3. Grasping Decision Trees

  • Divide and conquer strategy
  • The C5.0 decision tree algorithm
  • Selecting the optimal split
  • Pruning the decision tree

4. Grasping Classification Rules

  • Separate and conquer approach
  • The One Rule algorithm
  • The RIPPER algorithm
  • Deriving rules from decision trees

5. Grasping Regression

  • Simple linear regression
  • Ordinary least squares estimation
  • Correlations
  • Multiple linear regression

6. Grasping Regression Trees and Model Trees

  • Integrating regression into trees

7. Grasping Neural Networks

  • From biological to artificial neurons
  • Activation functions
  • Network topology
  • Determining the number of layers
  • The direction of information flow
  • Specifying the number of nodes per layer
  • Training neural networks via backpropagation

8. Grasping Support Vector Machines

  • Classification using hyperplanes
  • Identifying the maximum margin
  • Scenarios with linearly separable data
  • Scenarios with non-linearly separable data
  • Employing kernels for non-linear spaces

9. Grasping Association Rules

  • The Apriori algorithm for association rule learning
  • Evaluating rule interest – support and confidence
  • Constructing a rule set using the Apriori principle

10. Grasping Clustering

  • Clustering as a machine learning task
  • The k-means clustering algorithm
  • Utilizing distance for cluster assignment and updates
  • Determining the appropriate number of clusters

11. Evaluating Performance for Classification

  • Handling classification prediction data
  • Examining confusion matrices in detail
  • Assessing performance using confusion matrices
  • Beyond accuracy – alternative performance metrics
  • The kappa statistic
  • Sensitivity and specificity
  • Precision and recall
  • The F-measure
  • Visualizing performance trade-offs
  • ROC curves
  • Estimating future performance
  • The holdout method
  • Cross-validation
  • Bootstrap sampling

12. Optimizing Base Models for Enhanced Performance

  • Utilizing caret for automated parameter tuning
  • Developing a basic tuned model
  • Customizing the tuning workflow
  • Enhancing model performance through meta-learning
  • Understanding ensemble methods
  • Bagging
  • Boosting
  • Random forests
  • Training random forests
  • Evaluating random forest performance

13. Deep Learning

  • Three Categories of Deep Learning
  • Deep Autoencoders
  • Pre-trained Deep Neural Networks
  • Deep Stacking Networks

14. Discussion of Specific Application Domains

 21 Hours

Number of participants


Price per participant

Testimonials (1)

Upcoming Courses

Related Categories