PowerPoint Content Outline: Unit 2 - Decision Trees and
Computational Learning (Detailed with Figures and Code)
Slide 3: Introduction to Linear Algebra
Concepts: Scalars, Vectors, Matrices, Dot Product, Transpose, Inverse
Applications in ML: Feature transformation, dimensionality reduction
Diagram: Matrix multiplication visual
Slide 4: Linear Regression
Goal: Predict continuous outcomes based on input features
Model: (y = _0 + _1 x + )
Loss Function: Mean Squared Error (MSE)
Python Snippet:
from sklearn.linear_model import LinearRegression
model = LinearRegression()
model.fit(X_train, y_train)
Plot: Regression line fitting on scatter data
Slide 5: Learning with Trees
Tree-based models: Represent decisions hierarchically
Advantages: Easy to interpret, non-linear boundaries, handles
missing data
Diagram: Sample binary decision tree
Slide 6: Decision Trees (ID3, C4.5, CART)
Concepts:
o Nodes = decision points
o Leaves = outcomes
Splitting Criteria:
o Information Gain (ID3)
o Gain Ratio (C4.5)
o Gini Index (CART)
Python Snippet:
from sklearn.tree import DecisionTreeClassifier
clf = DecisionTreeClassifier(criterion='entropy')
clf.fit(X_train, y_train)
Slide 7: Constructing Trees using Weka
Weka GUI: Import dataset → Select classifier → Choose J48 (C4.5)
Steps:
1. Load ARFF/CSV file
2. Choose “Classify” tab
3. Select J48 → Start training
Screenshot: Tree output in Weka (insert manually)
Slide 8: Classification and Regression Trees (CART)
Handles both classification and regression tasks
Splits: Binary (yes/no)
Pruning: Reduces overfitting
Diagram: CART tree with split conditions
Slide 9: Support Vector Machines (SVM)
Goal: Find hyperplane with maximum margin
Kernels: Linear, Polynomial, RBF
Advantages: Works well for high-dimensional data
Diagram: SVM margin and support vectors
Python Snippet:
from sklearn.svm import SVC
model = SVC(kernel='linear')
model.fit(X_train, y_train)
Slide 10: Principal Component Analysis (PCA)
Purpose: Dimensionality reduction while preserving variance
Steps:
1. Standardize data
2. Compute covariance matrix
3. Find eigenvectors/eigenvalues
Python Snippet:
from sklearn.decomposition import PCA
pca = PCA(n_components=2)
X_reduced = pca.fit_transform(X)
Diagram: Projection of high-D data to 2D plane
Slide 11: Deep Choice Model
Extension of logistic regression for complex decision-making
Can incorporate deep learning layers to model choices
Example Use Case: Predicting user selections in recommender
systems
Note: Research-oriented concept, often implemented using custom
neural nets
Slide 12: Summary
Linear models are foundational for understanding complex learners
Decision trees and SVMs offer interpretable and powerful classifiers
PCA aids in feature reduction for visualization and speed
Tools like Weka simplify implementation of tree models
Let me know if you’d like me to generate diagrams, export to PowerPoint, or
add animations.