Deep Learning isa part of Artificial Intelligence (AI) and
Machine Learning (ML).
• It teaches computers to learn from large amounts of data
(like images, text, voice, numbers) just like humans learn
from experience.
• Called “deep” because it uses many layers of processing to
learn patterns.
Example:
• Facebook uses deep learning to recognize faces in photos.
• Netflix uses it to recommend movies based on what you
watch.
• In simple terms: Deep Learning = Teaching computers to
think like the human brain using layers of learning.
3.
• Input Data→ Raw data (images, voice, text).
• Layers of Neurons → The data passes through many
"layers," each layer learning something new.
• First layer → learns simple features (edges in an image,
keywords in text).
• Middle layers → learn combinations (shapes, grammar).
• Last layer → gives final prediction (cat vs. dog, positive vs.
negative review).
• Output → The answer/result.
• Deep Learning = A smart AI technique that learns from big
data using neural networks, works like the brain, and is used
in HR, Marketing, Manufacturing, Finance, and Healthcare.
4.
Common Applications (BusinessRelated)
HR
• Resume screening using AI.
• Predicting employee turnover using deep learning models.
Marketing
• Customer sentiment analysis from social media.
• Personalized product recommendations (Amazon, Netflix).
Manufacturing
• Predictive maintenance (machines failing early).
• Quality control using image recognition in production lines.
5.
Healthcare
• Diagnosing diseasesfrom X-rays, MRI
scans.
• Drug discovery.
Finance
• Fraud detection in online transactions.
• Algorithmic trading.
6.
Advantages
• Learns complexpatterns better than normal machine
learning.
• Works very well with images, videos, speech, and text.
• Can make very accurate predictions if enough data is
available.
Disadvantages
• Needs huge data and powerful computers.
• Takes longer training time.
• Works like a black box → hard to understand how
exactly it makes decisions.
7.
• Suppose youwant to build a system to recognize cats vs.
dogs in photos:
• Traditional ML → You would program rules like “If ears are
pointed and tail is straight, maybe it’s a cat.”
• Deep Learning → You give the system 1,00,000 cat & dog
photos. The system automatically learns features (ears,
tails, face shapes) without you writing any rule.
8.
• A companywants to predict if employees will leave
(attrition) or stay.
• How Deep Learning helps:
• Input data: employee records (age, salary, job role,
performance, promotions, leave patterns, etc.).
• Deep learning model: Neural networks learn hidden
patterns like “employees with low promotion chances and
long working hours are more likely to quit.”
• Output: Prediction → “Stay” or “Leave.”
• Benefit: HR can take action early (salary hike, promotion,
training) to retain employees.
9.
• Example 2:Marketing (Customer Sentiment Analysis)
• Question: A brand wants to know whether customers
are happy or unhappy from social media comments.
• How Deep Learning helps:
• Input data: Tweets, Facebook posts, Google reviews.
• Neural network reads words and tone (like “love,”
“worst,” “amazing”).
• Model predicts whether the review is Positive, Neutral,
or Negative.
• ✅ Benefit: Marketing team can improve customer
service and brand reputation.
10.
Example 3:
Manufacturing (DefectDetection in Products)
Question: A car company wants to detect defective parts in its
production line.
How Deep Learning helps:
Input data: Thousands of images of car parts.
• Neural network learns to recognize cracks, scratches, or
misaligned parts.
• Output: Flags a product as “Defective” or “Good.”
• Benefit: Saves cost, improves quality, and avoids faulty
products reaching customers.
• What isa Neural Network?
• A Neural Network is the basic building block of Deep
Learning.
• It is designed to work like the human brain – with
“neurons” (nodes) connected by “links.”
• Each neuron takes input, processes it, and passes it to
the next neuron.
13.
Layers:
Input Layer –takes data (like exam marks, customer details, product features).
Hidden Layers – process the data step by step (like the brain thinking).
Output Layer – gives the final result (like “Hire candidate” or “Not
hire”).
• Example:
• In HR: A neural network can take inputs like skills, experience, and
interview scores → predict whether the candidate will perform well.
• In Marketing: It can learn customer purchase history → predict who will
buy a new product.
14.
• Example 1:Neural Network in Banking (Loan Approval
Prediction)
Problem:
• A bank wants to predict whether a customer’s loan
application should be Approved or Rejected.
Step 1: Input Data (Input Layer)
• Customer details are given as input:
• Age = 30
• Income = 60,000/month
₹
• Credit Score = 750
• Existing Loans = 1
15.
• Step 2:Hidden Layers (Learning Patterns)
• The network learns important combinations:
• Neuron 1: “High income + Good credit score → More chance of
approval.”
• Neuron 2: “Low income + Too many loans → High risk.”
• Neuron 3: “Young age + Good repayment history → Safe
customer.”
• The weights adjust during training based on past loan approvals
and defaults.
16.
• Step 3:Output Layer (Prediction)
• The system predicts:
• Approval = 85%
• Rejection = 15%
• ✅ Final Decision: Loan Approved
• This helps banks reduce risk and speed up loan processing.
17.
• Example 2:Neural Network in Healthcare (Disease
Prediction)
Problem:
• Doctors want to predict if a patient has Diabetes.
• Step 1: Input Data (Input Layer)
• Patient details are given as input:
• Age = 45
• BMI (Body Mass Index) = 32
• Blood Sugar Level = 180 mg/dL
• Family History = Yes
18.
• Step 2:Hidden Layers (Learning Patterns)
• The network analyzes the data:
• Neuron 1: “High BMI + High blood sugar → Strong
indicator of diabetes.”
• Neuron 2: “Family history + High age → Risk
increases.”
• Neuron 3: “Normal sugar + Healthy BMI → Lower risk.”
19.
• Step 3:Output Layer (Prediction)
• The neural network predicts:
• Diabetes = 78%
• No Diabetes = 22%
• ✅ Final Prediction: Diabetes Likely
Doctors can take preventive action early.
• What isa Deep Networks (DFN)?
• A Deep Forward Network is a type of Artificial Neural
Network where information flows only in one direction:
from input → hidden layers → output.
• There are no loops or feedback connections.
• It is widely used for prediction and classification tasks in
HR, marketing, finance, healthcare, and manufacturing.
22.
• Neural Networks= All types of transport (cars, buses,
trains, flights).
• Deep Forward Network = Only cars (a type of transport).
Scope:
• Neural Network = General concept (umbrella term).
• DFN = A specific kind of neural network (the simplest
one).
23.
• Flow ofInformation:
• Neural Network = Can include loops, feedback (like
RNN, CNN).
• DFN = Only forward direction, no feedback loops.
• Use Case:
• Neural Network = Any problem (images, text, speech,
prediction).
• DFN = Mostly used for simple classification or
prediction problems.
24.
• Deep ForwardNetwork (DFN) – A Type of Neural
Network
A DFN (or Feedforward Neural Network) is one specific
type of neural network.
• In DFN, information flows only in one direction:
• Input → Hidden Layers → Output
• There are no loops or backward connections.
• Example: Predicting whether a customer will buy a
product.
25.
• Example 1– HR (Employee Attrition Prediction)
• Problem: Predict whether an employee will stay or leave.
• Input Layer:
• Age, Salary, Years of Service, Job Satisfaction.
• Hidden Layers:
• Neuron 1 learns: “Low salary + low satisfaction → high chance of
leaving.”
• Neuron 2 learns: “High salary + long service → high chance of staying.”
• Output Layer:
• Stay = 0
• Leave = 1
• Prediction = Leave (with 80% probability).
• ✅ Helps HR take action before the employee leaves.
26.
• Example 2– Marketing (Customer Purchase Prediction)
• Problem: Predict whether a customer will buy a product after seeing an
online ad.
• Input Layer:
• Age, Gender, Browsing History, Time spent on Ad.
• Hidden Layers:
• Neuron 1 learns: “Young age + more time spent on ad → higher purchase
probability.”
• Neuron 2 learns: “Past buying history → higher chance of buying again.”
• Output Layer:
• Buy = 1
• Not Buy = 0
• Prediction = Buy (with 65% probability).
• ✅ Helps marketing teams target the right customers.
• 1. Whatis XOR?
• XOR (Exclusive OR) is a logical operation.
• Rule: Output is 1 if inputs are different, and 0 if inputs are
the same.
• Example truth table:
HR Domain
• Problem:Predict whether an employee will leave the company
(Attrition).
• Inputs (A, B):
• A = Job Satisfaction (High=1, Low=0)
• B = Salary Satisfaction (High=1, Low=0)
• Output (XOR):
• If both are satisfied (0,0 or 1,1) → Employee stays (0)
• If one is satisfied but not the other (0,1 or 1,0) →
Employee leaves (1)
• ✅ Deep Learning can learn this XOR-like relation by adjusting
hidden neurons.
31.
• Example 2:Marketing Domain
Problem: Predict whether a customer will purchase.
• Inputs (A, B):
• A = Discount Offered (Yes=1, No=0)
• B = Brand Loyalty (Yes=1, No=0)
• Output (XOR):
32.
• If customergets discount but no loyalty → Buy (1)
• If loyal but no discount → Buy (1)
• If both are loyal + discount (too much saturation) → No
buy (0)
• If neither loyalty nor discount → No buy (0)
• ✅ Neural network hidden layers learn this non-linear XOR
pattern for purchase prediction.
• What isGradient-Based Learning?
• Gradient-Based Learning is a method used to train
machine learning and deep learning models.
• It works by minimizing errors (difference between actual
and predicted results).
• The key technique used is Gradient Descent.
35.
• 2. Whatis Gradient Descent?
• A mathematical optimization algorithm.
• It updates model parameters (weights) step by step to
reduce the loss function (error).
• Imagine climbing down a hill → each step moves toward
the lowest point (minimum error).
36.
Steps in GradientDescent
• Initialize Weights (random values).
• Forward Pass – Predict output using current weights.
• Calculate Loss – Find difference between prediction and actual value.
• Compute Gradient – Slope/derivative shows the direction of steepest change.
• Update Weights – Adjust weights opposite to gradient direction.
• Repeat steps until loss is very small.
37.
• Importance inDeep Learning
• Without gradient-based learning, neural networks cannot
be trained.
• It allows models to learn from data and improve
predictions.
• Used in HR analytics, marketing prediction,
manufacturing defect detection, finance forecasting, etc.
38.
4. Types ofGradient Descent
• Batch Gradient Descent – Uses the whole dataset for each
update (accurate but slow).
• Stochastic Gradient Descent (SGD) – Updates weights after each
sample (faster, but noisy).
• Mini-Batch Gradient Descent – Uses small groups of data at a
time (balance of speed and accuracy).
39.
6. Example (SimpleUnderstanding)
• HR Example: Predict employee attrition.
• The model first guesses wrongly (say, predicts 80% chance
employee will stay but actually they leave).
• Error = difference between prediction and reality.
• Gradient descent adjusts the weights (e.g., importance given to
salary, job satisfaction) so the next prediction is better.
40.
• HR Example:Predicting Employee Attrition (No Calculations)
• Suppose HR is building a model to predict if an employee will
leave.
• The model looks at factors like salary satisfaction and work–life
balance.
• 👉 Step 1 – First Guess (Wrong):
• The model predicts that the employee will stay (because it gives
more importance to salary), but in reality the employee leaves.
• 👉 Step 2 – Error:
• There is a gap between prediction and reality → this gap is called
the error.
41.
• Step 3– Gradient Descent Adjustment:
• The model checks where it went wrong and realizes:
• It did not give enough importance to “work–life balance.”
• It slightly overvalued salary satisfaction.
• 👉 Step 4 – Learning:
• The model adjusts its internal settings (weights). Now, it gives more
importance to work–life balance as a reason for leaving.
• 👉 Step 5 – Next Prediction (Better):
• The next time a similar employee profile comes, the model is more likely to
correctly predict that they will leave.
42.
• Meaning forHR:
• Over time, the model learns patterns — like “poor work–
life balance” and “low salary satisfaction” → higher
chance of leaving.
• This helps HR plan retention strategies like better
engagement programs or flexible work policies.
• What areHidden Units?
• Hidden units (or hidden neurons) are nodes inside the
hidden layers of a neural network.
• They are called hidden because we don’t directly see
them — they work between input and output.
45.
• Role ofHidden Units:
• They combine and transform inputs to detect patterns.
• Each hidden unit applies a mathematical function to the
input (like adding weights and applying an activation).
• They allow the model to capture non-linear relationships
(things that are not straight-line simple).
46.
Why Important?
• Withouthidden units, a neural network is just a simple
linear model.
• With hidden units, networks can learn complex tasks (like
speech recognition, image recognition, or predicting
employee attrition).
• Too Few Hidden Units → Underfitting:
• The model is too simple, misses patterns.
• Too Many Hidden Units → Overfitting:
• The model memorizes training data but fails on new data.
• Balance Needed.
47.
• Examples ofHidden Units
• 1. HR Example – Employee Attrition Prediction
• Input: Salary, Age, Job Satisfaction, Workload.
• Hidden Units: Learn patterns like:
• “Low salary + High workload → risk of attrition.”
• “Young age + Low satisfaction → risk of attrition.”
• Output: Predicts whether the employee will Stay or Leave.
• 👉 Here, hidden units are capturing relationships HR managers might not
see directly.
48.
• 2. MarketingExample – Customer Buying Behavior
• Input: Ad clicks, Browsing time, Discounts offered.
• Hidden Units: Learn patterns like:
• “More ad clicks + discount = higher chance of buying.”
• “Browsing but no discount = low chance of buying.”
• Output: Predicts whether the customer buys or does not buy.
• What isArchitecture Design?
• It is about how we design the structure of a neural
network — how many layers, how many hidden units,
what type of activation functions, etc.
• Just like a building needs a blueprint, a neural network
needs an architecture.
52.
• Key Elementsof Architecture Design:
• Input Layer – Where data enters the model (e.g., employee details, customer
data, machine signals).
• Hidden Layers – Layers in between that learn patterns using hidden units.
• Output Layer – Gives the final prediction (e.g., “leave or stay,” “buy or not,”
“failure or safe”).
• Activation Functions – Decide how signals flow (ReLU, Sigmoid, Tanh).
• Connections – How neurons are linked (fully connected, convolution,
recurrent).
53.
• Simple Typesof Architectures:
• Shallow Network: Few hidden layers → simple patterns.
• Deep Network: Many hidden layers → complex patterns.
• Specialized Architectures: CNN (images), RNN
(sequences like time series, text).
54.
• Why isit Important?
• The right architecture helps the model learn effectively.
• Wrong design can lead to overfitting (too complex) or
underfitting (too simple
55.
• Common activationfunctions:
• Sigmoid → outputs between 0 and 1 (good for binary
classification).
• ReLU (Rectified Linear Unit) → outputs positive values only, helps
deep networks train faster.
• Tanh → outputs between -1 and +1, centered around 0.
• Softmax → used for multi-class classification (e.g., predicting 3+
categories).
56.
• 1. HRExample – Employee Attrition Prediction
• Problem: The HR department wants to predict whether an employee will stay or leave.
• Input Layer: Data features such as age, years of experience, salary level, work–life balance
score, and job satisfaction rating.
• Hidden Layers:
• First hidden layer combines inputs → e.g., it learns that younger employees with low salary
are more likely to leave.
• Second hidden layer refines the pattern → e.g., low work–life balance + high workload
increases attrition risk.
• Output Layer: Gives final prediction: Stay (0) or Leave (1).
57.
• Importance ofArchitecture Design:
• If the network is too shallow (only 1 hidden layer), it might miss
complex patterns.
• If it is too deep (many unnecessary hidden layers), it may
“memorize” instead of “learning” (overfitting).
• 👉 Benefit to HR: Helps HR predict attrition early, so they can
design retention strategies (salary revision, flexible hours,
training).
58.
• 2. MarketingExample – Customer Churn Prediction
• Problem: A retail company wants to know which
customers are likely to stop buying (churn).
• Input Layer: Customer data such as frequency of
purchases, loyalty points, response to promotions, and
social media engagement.
59.
• Hidden Layers:
•First hidden layer combines behavior → e.g., low loyalty
points + few recent purchases means customer may
churn.
• Second hidden layer adds more detail → e.g., did not
respond to latest promotion + reduced engagement =
stronger chance of churn.
60.
• Output Layer:Predicts Churn (Yes/No).
• Importance of Architecture Design:
• Too simple → only sees 1–2 factors (e.g., just purchase
history).
• Proper design → sees combined patterns like “low
engagement + low purchase frequency” together.
61.
• 3. ManufacturingExample – Predictive Maintenance
• Problem: A factory wants to predict if a machine will fail soon.
• Input Layer: Sensor readings (temperature, vibration, noise level,
pressure).
• Hidden Layers:
• First hidden layer learns basic relations → e.g., high temperature
+ high vibration indicates potential wear.
• Second hidden layer refines → e.g., temperature rise + vibration +
unusual noise pattern = very high failure risk.
62.
• Output Layer:Predicts Failure soon (1) or Safe (0).
• Importance of Architecture Design:
• Wrong design → model might focus only on temperature and miss
combined signals.
• Good design → captures multi-factor interactions (temperature +
vibration + noise).
• 👉 Benefit to Manufacturing: Prevents sudden breakdowns,
reduces downtime, and saves costs.
63.
• HR Example– Employee Attrition
1 ️
1️⃣
• Input Features: Age, Salary, Work-life balance, Job satisfaction.
• Hidden Layer Activation (ReLU):
• Example: If salary satisfaction is low, ReLU activates strongly (positive
signal).
• If not relevant (e.g., salary already high), ReLU returns 0 → ignores it.
• Output Layer Activation (Sigmoid):
• Predicts probability of Stay (0) or Leave (1).
• Example: Sigmoid outputs 0.85 → 85% chance employee leaves.
• 👉 ReLU captures hidden patterns, Sigmoid gives a final “yes/no” probability.
64.
• 3️Manufacturing Example– Predictive Maintenance
• Input Features: Temperature, Vibration, Noise level.
• Hidden Layer Activation (ReLU or Tanh):
• Example: ReLU activates strongly when temperature + vibration exceed safe
limits.
• Tanh may be used if the signal needs to capture positive or negative
deviations.
• Output Layer Activation (Sigmoid):
• Example: Machine failure probability = 0.92 → 92% risk of breakdown soon.
• 👉 Maintenance team can schedule repair before failure happens.
• Regularization inDeep Learning
• Regularization is a set of techniques used in deep
learning to prevent overfitting.
• 👉 Overfitting = when the model learns the training data
too well (including noise and irrelevant details) and fails to
perform well on new/unseen data.
• 👉 Regularization helps the model to generalize better
(work well on new data).
67.
• Types ofRegularization
• L1 Regularization (Lasso):
• Adds the absolute value of weights to the loss function.
• Encourages sparsity (some weights become zero → irrelevant
features are removed).
• Example: In HR attrition, if "height" is an input but not useful, L1
can reduce its weight to zero.
68.
• L2 Regularization(Ridge):
• Adds the square of weights to the loss function.
• Keeps all weights small but not zero.
• Example: In Marketing churn prediction, prevents any single factor
(like discount usage) from dominating the model.
69.
• Dropout:
• Randomlydrops some neurons during training.
• Prevents the network from depending too much on
specific neurons.
• Example: In Manufacturing fault prediction, dropout
ensures the model doesn’t overly rely only on
"temperature" but considers vibration and noise too.
70.
• Early Stopping:
•Stop training when performance on validation data stops
improving.
• Example: In HR attrition, if the model starts memorizing
employee data after 50 epochs, training is stopped early.
71.
• Data Augmentation:
•Increase training data artificially (e.g., rotating images,
adding noise).
• Mostly used in image/text domains.
• Batch Normalization (indirect regularization):
• Normalizes data inside the network layers, improves
stability and prevents overfitting.
72.
• 1. HRExample – Employee Attrition Prediction
• Scenario: An HR department wants to build a model to
predict whether an employee will leave the company or
stay.
• Problem: The model starts overfitting → it learns from
small details like “age = 29” or “employee ID pattern”
(irrelevant for attrition).
73.
• Regularization TechniqueUsed:
• L1 Regularization: The model drops useless features like "employee ID."
• Dropout: During training, the model ignores some hidden neurons, forcing it
to focus on important factors like salary, job satisfaction, and career growth.
• Outcome: The model generalizes better and accurately predicts attrition
across all employees, not just the training data.
• ✅ Result for HR: HR can now identify employees at risk of leaving and design
retention programs.
74.
• 3. ManufacturingExample – Machine Breakdown
Prediction
• Scenario: A factory uses sensors to predict when
machines may fail.
• Problem: The model overfits by focusing only on
temperature data, ignoring other signals like vibration and
noise levels.
75.
• Regularization TechniqueUsed:
• Dropout: Forces the model not to rely only on temperature
but also consider vibration and sound.
• Batch Normalization: Stabilizes learning and prevents
extreme weight values.
• Outcome: The model now uses a combination of signals
→ predictions become more reliable.
• ✅ Result for Manufacturing: Maintenance team gets
accurate alerts, reducing downtime and costs.