- Data Science
- Data Analysis
- Data Visualization
- Machine Learning
- Deep Learning
- Computer Vision
- Artificial Intelligence
- AI ML DS Interview Series
- AI ML DS Projects series
- Data Engineering
- Web Scrapping
Inductive Learning Algorithm
In this article, we will learn about Inductive Learning Algorithm which generally comes under the domain of Machine Learning.
What is Inductive Learning Algorithm?
Inductive Learning Algorithm (ILA) is an iterative and inductive machine learning algorithm that is used for generating a set of classification rules, which produces rules of the form “IF-THEN”, for a set of examples, producing rules at each iteration and appending to the set of rules.
There are basically two methods for knowledge extraction firstly from domain experts and then with machine learning. For a very large amount of data, the domain experts are not very useful and reliable. So we move towards the machine learning approach for this work. To use machine learning One method is to replicate the expert’s logic in the form of algorithms but this work is very tedious, time taking, and expensive. So we move towards the inductive algorithms which generate the strategy for performing a task and need not instruct separately at each step.
Why you should use Inductive Learning?
The ILA is a new algorithm that was needed even when other reinforcement learnings like ID3 and AQ were available.
- The need was due to the pitfalls which were present in the previous algorithms, one of the major pitfalls was the lack of generalization of rules.
- The ID3 and AQ used the decision tree production method which was too specific which were difficult to analyze and very slow to perform for basic short classification problems.
- The decision tree-based algorithm was unable to work for a new problem if some attributes are missing.
- The ILA uses the method of production of a general set of rules instead of decision trees , which overcomes the above problems
Basic Requirements to Apply Inductive Learning Algorithm
- List the examples in the form of a table ‘T’ where each row corresponds to an example and each column contains an attribute value.
- Create a set of m training examples, each example composed of k attributes and a class attribute with n possible decisions.
- Create a rule set, R, having the initial value false.
- Initially, all rows in the table are unmarked.
Necessary Steps for Implementation
- Step 1: divide the table ‘T’ containing m examples into n sub-tables (t1, t2,…..tn). One table for each possible value of the class attribute. (repeat steps 2-8 for each sub-table)
- Step 2: Initialize the attribute combination count ‘ j ‘ = 1.
- Step 3: For the sub-table on which work is going on, divide the attribute list into distinct combinations, each combination with ‘j ‘ distinct attributes.
- Step 4: For each combination of attributes, count the number of occurrences of attribute values that appear under the same combination of attributes in unmarked rows of the sub-table under consideration, and at the same time, not appears under the same combination of attributes of other sub-tables. Call the first combination with the maximum number of occurrences the max-combination ‘ MAX’.
- Step 5: If ‘MAX’ == null, increase ‘ j ‘ by 1 and go to Step 3.
- Step 6: Mark all rows of the sub-table where working, in which the values of ‘MAX’ appear, as classified.
- Step 7: Add a rule (IF attribute = “XYZ” –> THEN decision is YES/ NO) to R whose left-hand side will have attribute names of the ‘MAX’ with their values separated by AND, and its right-hand side contains the decision attribute value associated with the sub-table.
- Step 8: If all rows are marked as classified, then move on to process another sub-table and go to Step 2. Else, go to Step 4. If no sub-tables are available, exit with the set of rules obtained till then.
An example showing the use of ILA suppose an example set having attributes Place type, weather, location, decision, and seven examples, our task is to generate a set of rules that under what condition is the decision.
Subset – 1
Subset – 2
- At iteration 1 rows 3 & 4 column weather is selected and rows 3 & 4 are marked. the rule is added to R IF the weather is warm then a decision is yes.
- At iteration 2 row 1 column place type is selected and row 1 is marked. the rule is added to R IF the place type is hilly then the decision is yes.
- At iteration 3 row 2 column location is selected and row 2 is marked. the rule is added to R IF the location is Shimla then the decision is yes.
- At iteration 4 row 5&6 column location is selected and row 5&6 are marked. the rule is added to R IF the location is Mumbai then a decision is no.
- At iteration 5 row 7 column place type & the weather is selected and row 7 is marked. the rule is added to R IF the place type is beach AND the weather is windy then the decision is no.
Finally, we get the rule set:- Rule Set
- Rule 1: IF the weather is warm THEN the decision is yes.
- Rule 2: IF the place type is hilly THEN the decision is yes.
- Rule 3: IF the location is Shimla THEN the decision is yes.
- Rule 4: IF the location is Mumbai THEN the decision is no.
- Rule 5: IF the place type is beach AND the weather is windy THEN the decision is no.
Similar Reads
Improve your coding skills with practice.
What kind of Experience do you want to share?
Machine Learning
- Machine Learning Tutorial
- Machine Learning Applications
- Life cycle of Machine Learning
- Install Anaconda & Python
- AI vs Machine Learning
- How to Get Datasets
- Data Preprocessing
- Supervised Machine Learning
- Unsupervised Machine Learning
- Supervised vs Unsupervised Learning
Supervised Learning
- Regression Analysis
- Linear Regression
- Simple Linear Regression
- Multiple Linear Regression
- Backward Elimination
- Polynomial Regression
Classification
- Classification Algorithm
- Logistic Regression
- K-NN Algorithm
- Support Vector Machine Algorithm
- Na�ve Bayes Classifier
Miscellaneous
- Classification vs Regression
- Linear Regression vs Logistic Regression
- Decision Tree Classification Algorithm
- Random Forest Algorithm
- Clustering in Machine Learning
- Hierarchical Clustering in Machine Learning
- K-Means Clustering Algorithm
- Apriori Algorithm in Machine Learning
- Association Rule Learning
- Confusion Matrix
- Cross-Validation
- Data Science vs Machine Learning
- Machine Learning vs Deep Learning
- Dimensionality Reduction Technique
- Machine Learning Algorithms
- Overfitting & Underfitting
- Principal Component Analysis
- What is P-Value
- Regularization in Machine Learning
- Examples of Machine Learning
- Semi-Supervised Learning
- Essential Mathematics for Machine Learning
- Overfitting in Machine Learning
- Types of Encoding Techniques
- Feature Selection Techniques in Machine Learning
- Bias and Variance in Machine Learning
- Machine Learning Tools
- Prerequisites for Machine Learning
- Gradient Descent in Machine Learning
- Machine Learning Experts Salary in India
- Machine Learning Models
- Machine Learning Books
- Linear Algebra for Machine learning
- Types of Machine Learning
- Feature Engineering for Machine Learning
- Top 10 Machine Learning Courses in 2021
- Epoch in Machine Learning
- Machine Learning with Anomaly Detection
- What is Epoch
- Cost Function in Machine Learning
- Bayes Theorem in Machine learning
- Perceptron in Machine Learning
- Entropy in Machine Learning
- Issues in Machine Learning
- Precision and Recall in Machine Learning
- Genetic Algorithm in Machine Learning
- Normalization in Machine Learning
- Adversarial Machine Learning
- Basic Concepts in Machine Learning
- Machine Learning Techniques
- Demystifying Machine Learning
- Challenges of Machine Learning
- Model Parameter vs Hyperparameter
- Hyperparameters in Machine Learning
- Importance of Machine Learning
- Machine Learning and Cloud Computing
- Anti-Money Laundering using Machine Learning
- Data Science Vs. Machine Learning Vs. Big Data
- Popular Machine Learning Platforms
- Deep learning vs. Machine learning vs. Artificial Intelligence
- Machine Learning Application in Defense/Military
- Machine Learning Applications in Media
- How can Machine Learning be used with Blockchain
- Prerequisites to Learn Artificial Intelligence and Machine Learning
- List of Machine Learning Companies in India
- Mathematics Courses for Machine Learning
- Probability and Statistics Books for Machine Learning
- Risks of Machine Learning
- Best Laptops for Machine Learning
- Machine Learning in Finance
- Lead Generation using Machine Learning
- Machine Learning and Data Science Certification
- What is Big Data and Machine Learning
- How to Save a Machine Learning Model
- Machine Learning Model with Teachable Machine
- Data Structure for Machine Learning
- Hypothesis in Machine Learning
- Gaussian Discriminant Analysis
- How Machine Learning is used by Famous Companies
- Introduction to Transfer Learning in ML
- LDA in Machine Learning
- Stacking in Machine Learning
- CNB Algorithm
- Deploy a Machine Learning Model using Streamlit Library
- Different Types of Methods for Clustering Algorithms in ML
- EM Algorithm in Machine Learning
- Machine Learning Pipeline
- Exploitation and Exploration in Machine Learning
- Machine Learning for Trading
- Data Augmentation: A Tactic to Improve the Performance of ML
- Difference Between Coding in Data Science and Machine Learning
- Data Labelling in Machine Learning
- Impact of Deep Learning on Personalization
- Major Business Applications of Convolutional Neural Network
- Mini Batch K-means clustering algorithm
- What is Multilevel Modelling
- GBM in Machine Learning
- Back Propagation through time - RNN
- Data Preparation in Machine Learning
- Predictive Maintenance Using Machine Learning
- NLP Analysis of Restaurant Reviews
- What are LSTM Networks
- Performance Metrics in Machine Learning
- Optimization using Hopfield Network
- Data Leakage in Machine Learning
- Generative Adversarial Network
- Machine Learning for Data Management
- Tensor Processing Units
- Train and Test datasets in Machine Learning
- How to Start with Machine Learning
- AUC-ROC Curve in Machine Learning
- Targeted Advertising using Machine Learning
- Top 10 Machine Learning Projects for Beginners using Python
- What is Human-in-the-Loop Machine Learning
- What is MLOps
- K-Medoids clustering-Theoretical Explanation
- Machine Learning Or Software Development: Which is Better
- How does Machine Learning Work
- How to learn Machine Learning from Scratch
- Is Machine Learning Hard
- Face Recognition in Machine Learning
- Product Recommendation Machine Learning
- Designing a Learning System in Machine Learning
- Recommendation System - Machine Learning
- Customer Segmentation Using Machine Learning
- Detecting Phishing Websites using Machine Learning
- Hidden Markov Model in Machine Learning
- Sales Prediction Using Machine Learning
- Crop Yield Prediction Using Machine Learning
- Data Visualization in Machine Learning
- ELM in Machine Learning
- Probabilistic Model in Machine Learning
- Survival Analysis Using Machine Learning
- Traffic Prediction Using Machine Learning
- t-SNE in Machine Learning
- BERT Language Model
- Federated Learning in Machine Learning
- Deep Parametric Continuous Convolutional Neural Network
- Depth-wise Separable Convolutional Neural Networks
- Need for Data Structures and Algorithms for Deep Learning and Machine Learning
- Geometric Model in Machine Learning
- Machine Learning Prediction
- Scalable Machine Learning
- Credit Score Prediction using Machine Learning
- Extrapolation in Machine Learning
- Image Forgery Detection Using Machine Learning
- Insurance Fraud Detection -Machine Learning
- NPS in Machine Learning
- Sequence Classification- Machine Learning
- EfficientNet: A Breakthrough in Machine Learning Model Architecture
- focl algorithm in Machine Learning
- Gini Index in Machine Learning
- Rainfall Prediction using ML
- Major Kernel Functions in Support Vector Machine
- Bagging Machine Learning
- BERT Applications
- Xtreme: MultiLingual Neural Network
- History of Machine Learning
- Multimodal Transformer Models
- Pruning in Machine Learning
- ResNet: Residual Network
- Gold Price Prediction using Machine Learning
- Dog Breed Classification using Transfer Learning
- Cataract Detection Using Machine Learning
- Placement Prediction Using Machine Learning
- Stock Market prediction using Machine Learning
- How to Check the Accuracy of your Machine Learning Model
- Interpretability and Explainability: Transformer Models
- Pattern Recognition in Machine Learning
- Zillow Home Value (Zestimate) Prediction in ML
- Fake News Detection Using Machine Learning
- Genetic Programming VS Machine Learning
- IPL Prediction Using Machine Learning
- Document Classification Using Machine Learning
- Heart Disease Prediction Using Machine Learning
- OCR with Machine Learning
- Air Pollution Prediction Using Machine Learning
- Customer Churn Prediction Using Machine Learning
- Earthquake Prediction Using Machine Learning
- Factor Analysis in Machine Learning
- Locally Weighted Linear Regression
- Machine Learning in Restaurant Industry
- Machine Learning Methods for Data-Driven Turbulence Modeling
- Predicting Student Dropout Using Machine Learning
- Image Processing Using Machine Learning
- Machine Learning in Banking
- Machine Learning in Education
- Machine Learning in Healthcare
- Machine Learning in Robotics
- Cloud Computing for Machine Learning and Cognitive Applications
- Credit Card Approval Using Machine Learning
- Liver Disease Prediction Using Machine Learning
- Majority Voting Algorithm in Machine Learning
- Data Augmentation in Machine Learning
- Decision Tree Classifier in Machine Learning
- Machine Learning in Design
- Digit Recognition Using Machine Learning
- Electricity Consumption Prediction Using Machine Learning
- Data Analytics vs. Machine Learning
- Injury Prediction in Competitive Runners Using Machine Learning
- Protein Folding Using Machine Learning
- Sentiment Analysis Using Machine Learning
- Network Intrusion Detection System Using Machine Learning
- Titanic- Machine Learning From Disaster
- Adenovirus Disease Prediction for Child Healthcare Using Machine Learning
- RNN for Sequence Labelling
- CatBoost in Machine Learning
- Cloud Computing Future Trends
- Histogram of Oriented Gradients (HOG)
- Implementation of neural network from scratch using NumPy
- Introduction to SIFT( Scale Invariant Feature Transform)
- Introduction to SURF (Speeded-Up Robust Features)
- Kubernetes - load balancing service
- Kubernetes Resource Model (KRM) and How to Make Use of YAML
- Are Robots Self-Learning
- Variational Autoencoders
- What are the Security and Privacy Risks of VR and AR
- What is a Large Language Model (LLM)
- Privacy-preserving Machine Learning
- Continual Learning in Machine Learning
- Quantum Machine Learning (QML)
- Split Single Column into Multiple Columns in PySpark DataFrame
- Why should we use AutoML
- Evaluation Metrics for Object Detection and Recognition
- Mean Intersection over Union (mIoU) for image segmentation
- YOLOV5-Object-Tracker-In-Videos
- Predicting Salaries with Machine Learning
- Fine-tuning Large Language Models
- AutoML Workflow
- Build Chatbot Webapp with LangChain
- Building a Machine Learning Classification Model with PyCaret
- Continuous Bag of Words (CBOW) in NLP
- Deploying Scrapy Spider on ScrapingHub
- Dynamic Pricing Using Machine Learning
- How to Improve Neural Networks by Using Complex Numbers
- Introduction to Bayesian Deep Learning
- LiDAR: Light Detection and Ranging for 3D Reconstruction
- Meta-Learning in Machine Learning
- Object Recognition in Medical Imaging
- Region-level Evaluation Metrics for Image Segmentation
- Sarcasm Detection Using Neural Networks
- SARSA Reinforcement Learning
- Single Shot MultiBox Detector (SSD) using Neural Networking Approach
- Stepwise Predictive Analysis in Machine Learning
- Vision Transformers vs. Convolutional Neural Networks
- V-Net in Image Segmentation
- Forest Cover Type Prediction Using Machine Learning
- Ada Boost algorithm in Machine Learning
- Continuous Value Prediction
- Bayesian Regression
- Least Angle Regression
- Linear Models
- DNN Machine Learning
- Why do we need to learn Machine Learning
- Roles in Machine Learning
- Clustering Performance Evaluation
- Spectral Co-clustering
- 7 Best R Packages for Machine Learning
- Calculate Kurtosis
- Machine Learning for Data Analysis
- What are the benefits of 5G Technology for the Internet of Things
- What is the Role of Machine Learning in IoT
- Human Activity Recognition Using Machine Learning
- Components of GIS
- Attention Mechanism
- Backpropagation- Algorithm
- VGGNet-16 Architecture
- Independent Component Analysis
- Nonnegative Matrix Factorization
- Sparse Inverse Covariance
- Accuracy, Precision, Recall or F1
- L1 and L2 Regularization
- Maximum Likelihood Estimation
- Kernel Principal Component Analysis (KPCA)
- Latent Semantic Analysis
- Overview of outlier detection methods
- Robust Covariance Estimation
- Spectral Bi-Clustering
- Drift in Machine Learning
- Credit Card Fraud Detection Using Machine Learning
- KL-Divergence
- Transformers Architecture
- Novelty Detection with Local Outlier Factor
- Novelty Detection
- Introduction to Bayesian Linear Regression
- Firefly Algorithm
- Keras: Attention and Seq2Seq
- A Guide Towards a Successful Machine Learning Project
- ACF and PCF
- Bayesian Hyperparameter Optimization for Machine Learning
- Random Forest Hyperparameter tuning in python
- Simulated Annealing
- Top Benefits of Machine Learning in FinTech
- Weight Initialisation
- Density Estimation
- Overlay Network
- Micro, Macro Weighted Averages of F1 Score
- Assumptions of Linear Regression
- Evaluation Metrics for Clustering Algorithms
- Frog Leap Algorithm
- Isolation Forest
- McNemar Test
- Stochastic Optimization
- Geomagnetic Field Using Machine Learning
- Image Generation Using Machine Learning
- Confidence Intervals
- Facebook Prophet
- Understanding Optimization Algorithms in Machine Learning
- What Are Probabilistic Models in Machine Learning
- How to choose the best Linear Regression model
- How to Remove Non-Stationarity From Time Series
- AutoEncoders
- Cat Classification Using Machine Learning
- AIC and BIC
- Inception Model
- Architecture of Machine Learning
- Business Intelligence Vs Machine Learning
- Guide to Cluster Analysis: Applications, Best Practices
- Linear Regression using Gradient Descent
- Text Clustering with K-Means
- The Significance and Applications of Covariance Matrix
- Stationarity Tests in Time Series
- Graph Machine Learning
- Introduction to XGBoost Algorithm in Machine Learning
- Bahdanau Attention
- Greedy Layer Wise Pre-Training
- OneVsRestClassifier
- Best Program for Machine Learning
- Deep Boltzmann machines (DBMs) in machine learning
- Find Patterns in Data Using Machine Learning
- Generalized Linear Models
- How to Implement Gradient Descent Optimization from Scratch
- Interpreting Correlation Coefficients
- Image Captioning Using Machine Learning
- fit() vs predict() vs fit_predict() in Python scikit-learn
- CNN Filters
- Shannon Entropy
- Time Series -Exponential Smoothing
- AUC ROC Curve in Machine Learning
- Vector Norms in Machine Learning
- Swarm Intelligence
- L1 and L2 Regularization Methods in Machine Learning
- ML Approaches for Time Series
- MSE and Bias-Variance Decomposition
- Simple Exponential Smoothing
- How to Optimise Machine Learning Model
- Multiclass logistic regression from scratch
- Lightbm Multilabel Classification
- Monte Carlo Methods
- What is Inverse Reinforcement learning
- Content-Based Recommender System
- Context-Awareness Recommender System
- Predicting Flights Using Machine Learning
- NTLK Corpus
- Traditional Feature Engineering Models
- Concept Drift and Model Decay in Machine Learning
- Hierarchical Reinforcement Learning
- What is Feature Scaling and Why is it Important in Machine Learning
- Difference between Statistical Model and Machine Learning
- Introduction to Ranking Algorithms in Machine Learning
- Multicollinearity: Causes, Effects and Detection
- Bag of N-Grams Model
- TF-IDF Model
Related Tutorials
- Tensorflow Tutorial
- PyTorch Tutorial
- Data Science Tutorial
- AI Tutorial
- NLP Tutorial
- Reinforcement Learning
Interview Questions
- Machine learning Interview
Latest Courses
We provides tutorials and interview questions of all technology like java tutorial, android, java frameworks
Contact info
G-13, 2nd Floor, Sec-3, Noida, UP, 201301, India
[email protected] .
Latest Post
PRIVACY POLICY
Online Compiler
Chapter 2: Concept Learning and the General-to-Specific Ordering
- Concept Learning: Inferring a boolean valued function from training examples of its input and output.
- X: set of instances
- x: one instance
- c: target concept, c:X → {0, 1}
- < x, c(x) >, training instance, can be a positive example or a negative example
- D: set of training instances
- H: set of possible hypotheses
- h: one hypothesis, h: X → { 0, 1 }, the goal is to find h such that h(x) = c(x) for all x in X
Inductive Learning Hypothesis
Any hypothesis found to approximate the target function well over a sufficiently large set of training examples will also approximate the target function well over other unobserved examples.
Let h j and h k be boolean-valued functions defined over X. h j is more general than or equal to h k (written h j ≥ g h k ) if and only if (∀ x ∈ X) [ (h k (x) = 1) → (h j (x) = 1)]
This is a partial order since it is reflexive, antisymmetric and transitive.
Find-S Algorithm
Outputs a description of the most specific hypothesis consistent with the training examples.
- Initialize h to the most specific hypothesis in H
- If the constraint a i is NOT satisfied by x, then replace a i in h by the next more general constraint that is satisfied by x.
- Output hypothesis h
For this particular algorithm, there is a bias that the target concept can be represented by a conjunction of attribute constraints.
Candidate Elimination Algorithm
Outputs a description of the set of all hypotheses consistent with the training examples.
A hypothesis h is consistent with a set of training examples D if and only if h(x) = c(x) for each example < x, c(x) > in D. Consistent(h, D) ≡ (∀ < x, c(x) > ∈ D) h(x) = c(x)
The version space denoted VS H,D with respect to hypothesis space H and training examples D, is the subset of hypotheses from H consistent with the training examples in D. VS H,D ≡ { h ∈ H | Consistent(h, D) }
The general boundary G, with respect to hypothesis space H and training data D, is the set of maximally general members of H consistent with D.
The specific boundary S, with respect to hypothesis space H and training data D, is the set of maximally specific members of H consistent with D.
Version Space Representation
Let X be an arbitrary set of instances and let H be a set of boolean-valued hypotheses defined over X. Let c:X → {0,1} be an arbitrary target concept defined over X, and let D be an arbitrary set of training examples {<x, c(x)>}. For all X, H, c and D such that S and G are well defined, VS H,D = {h ∈ H | (∃s ∈ S) (∃g ∈ G) (g ≥ g h ≥ g s)}
- Initialize G to the set of maximally general hypotheses in H
- Initialize S to the set of maximally specific hypotheses in H
- Remove from G any hypothesis inconsistent with d
- Remove s from S
- Add to S all minimal generalizations h of s such that h is consistent with d, and some member of G is more general than h
- Remove from S any hypothesis that is more general than another hypothesis in S
- Remove from S any hypothesis inconsistent with d
- Remove g from G
- Add to G all minimal specializations h of g such that h is consistent with d, and some member of S is more specific than h
- Remove from G any hypothesis that is less general than another hypothesis in G
Candidate Elimination Algorithm Issues
- Will it converge to the correct hypothesis? Yes, if (1) the training examples are error free and (2) the correct hypothesis can be represented by a conjunction of attributes.
- If the learner can request a specific training example, which one should it select?
- How can a partially learned concept be used?
Inductive Bias
- Definition: Consider a concept learning algorithm L for the set of instances X. Let c be an arbitrary concept defined over X and let D c = {<x, c(x)>} be an arbitrary set of training examples of c. Let L(x i , D c ) denote the classification assigned to the instance x i by L after training on the data D c . The inductive bias of L is any minimal set of assertions B such that for any target concept c and corresponding training examples D c (∀ x i ∈ X) [ L(x i , D c ) follows deductively from (B ∧ D c ∧ x i ) ]
- Thus, one advantage of an inductive bias is that it gives the learner a rational basis for classifying unseen instances.
- What is another advantage of bias?
- What is one disadvantage of bias?
- What is the inductive bias of the candidate elimination algorithm? Answer: the target concept c is a conjunction of attributes.
- What is meant by a weak bias versus a strong bias?
Sample Exercise
Work exercise 2.4 on page 48.
IMAGES
VIDEO