Machine Learning 1
Machine Learning 1
o The more data provided, the better the model's performance (up to a limit).
2. Generalization:
o The goal of ML is to generalize from training data to unseen data, ensuring accurate
predictions or decisions.
3. Iterative Improvement:
o 1950: Alan Turing introduces the Turing Test, proposing criteria for machine
intelligence.
o 1957: Frank Rosenblatt invents the Perceptron, an early type of artificial neural
network, inspired by biological neurons.
o 1967: The nearest neighbour algorithm is developed, enabling pattern recognition for
tasks like handwriting and image classification.
o 1970s: Interest shifts toward rule-based systems and symbolic AI as neural networks
struggle due to limited computational power.
o 1986: Geoffrey Hinton and others publish the backpropagation algorithm, making
neural networks more practical for training.
o 1989: Yann LeCun demonstrates the first application of backpropagation for digit
recognition, a precursor to modern computer vision.
o IBM’s Deep Blue defeats chess grandmaster Garry Kasparov in 1997, highlighting
machine learning's potential in complex problem-solving.
o 2000: Boosting methods like AdaBoost and ensemble models become widely adopted.
o 2006: Geoffrey Hinton introduces the term deep learning, focusing on multi-layered
neural networks.
o Data availability and computational advances (e.g., GPUs) drive the rise of larger
datasets for training models.
7. 2020s: AI Everywhere
o 2020: OpenAI releases GPT-3, showcasing the power of large language models in
generating human-like text.
o 2022: Stable Diffusion is a text-to-image model that generates images from text
developed by the CompVis group at the Ludwig Maximilian University Munich,
Germany.
o 2022: OpenAI develops and publishes ChatGPT an advanced chat bot. It builds on
previous GPT models, but is much more developed. It is able to remember previous
parts of the conversation and can give complex answers.
• Disease Diagnosis: Predict diseases like cancer, diabetes, and Alzheimer's from medical
imaging or patient data.
• Drug Discovery: Speed up the identification of potential drug candidates using ML algorithms.
• Personalized Medicine: Tailor treatments to individual patients based on genetic, lifestyle, and
clinical data.
• Health Monitoring: Analyse data from wearable devices for early warning of medical
conditions.
2. Finance
• Algorithmic Trading: Optimize trading strategies using real-time data and predictive models.
• Risk Management: Predict and mitigate financial risks through advanced analytics.
• Inventory Management: Predict demand and optimize stock levels using ML models.
• Customer Sentiment Analysis: Analyse customer reviews and feedback for insights.
• Dynamic Pricing: Adjust pricing based on demand, competition, and other market factors.
4. Manufacturing
• Predictive Maintenance: Detect equipment failures before they occur by analysing sensor
data.
• Quality Control: Automate defect detection in production lines using image processing.
5. Transportation
• Autonomous Vehicles: Enable self-driving cars to navigate and make real-time decisions.
• Route Optimization: Suggest optimal routes for delivery and ride-hailing services.
• Traffic Management: Predict and manage congestion using traffic flow data.
6. Energy
• Renewable Energy Optimization: Enhance efficiency in solar and wind energy systems.
• Anomaly Detection: Monitor power grids and pipelines for faults or inefficiencies.
7. Agriculture
• Crop Monitoring: Use satellite imagery and sensor data to assess crop health and yield.
• Livestock Monitoring: Track the health and activity of animals using wearable sensors.
• Game AI: Enhance gameplay with intelligent opponents and dynamic environments.
9. Education
• Virtual Assistants: Provide real-time support for learners using AI-driven chatbots.
10. Security
• Natural Disaster Prediction: Analyse patterns to predict earthquakes, floods, and hurricanes.
• Chatbots and Virtual Assistants: Power intelligent systems like Siri, Alexa, and customer
support bots.
• Language Translation: Enable tools like Google Translate for real-time communication.
• Sentiment Analysis: Analyse public opinion from social media and reviews.
• Satellite Data Analysis: Monitor climate changes, deforestation, and urban growth.
• Wildlife Conservation: Monitor endangered species and illegal activities like poaching.
• Pollution Control: Analyse air and water quality data to identify and mitigate pollution sources.
1. Supervised Learning
Supervised learning algorithms are trained using labelled data, meaning the input data comes with
corresponding correct output labels. These algorithms learn the relationship between input features
and the target labels.
Common Algorithms:
• Decision Trees: Makes decisions based on splitting input data into nodes.
• Support Vector Machines (SVM): Finds the optimal hyperplane to classify data.
• K-Nearest Neighbours (KNN): Classifies based on the majority label of the nearest neighbours.
2. Unsupervised Learning
Unsupervised learning algorithms work with unlabelled data, aiming to identify hidden patterns,
structures, or relationships in the data.
Common Algorithms:
• K-Means Clustering: Groups similar data points into a fixed number of clusters.
• Hierarchical Clustering: Builds a tree of clusters based on similarity.
• Principal Component Analysis (PCA): Reduces the dimensionality of data while retaining the
most important information.
3. Semi-Supervised Learning
Semi-supervised learning is a hybrid approach where the algorithm is trained on a small amount of
labelled data and a large amount of unlabelled data. It is particularly useful when labelled data is scarce
or expensive to obtain.
Common Algorithms:
• Label Propagation: Propagates labels from the labelled instances to the unlabelled instances.
• Semi-supervised SVM: Modified SVM that leverages both labelled and unlabelled data.
4. Reinforcement Learning
Reinforcement learning (RL) algorithms learn by interacting with an environment and receiving
feedback in the form of rewards or punishments. The algorithm aims to maximize the cumulative
reward through trial and error.
Common Algorithms:
• Q-Learning: A model-free RL algorithm that learns the value of actions in a given state.
• Deep Q-Networks (DQN): Combines Q-learning with deep learning to handle large state
spaces.
5. Self-Supervised Learning
Self-supervised learning is a subset of unsupervised learning where the model generates labels from
the input data itself. It creates pseudo-labels based on patterns within the data, often used in NLP and
computer vision tasks.
Common Algorithms:
• Contrastive Learning: Learning to distinguish between similar and dissimilar pairs of data
points.
6. Deep Learning
Deep learning algorithms are a subset of machine learning based on neural networks with many layers
(also known as deep neural networks). These models are particularly powerful for handling large
amounts of unstructured data like images, audio, and text.
Common Algorithms:
• Convolutional Neural Networks (CNNs): Mainly used for image and video recognition tasks.
• Recurrent Neural Networks (RNNs): Used for sequential data like time series or natural
language processing.
• Long Short-Term Memory (LSTM): A type of RNN used to overcome the vanishing gradient
problem in long sequences.
• Transformers: A deep learning model designed for sequential data processing, widely used in
NLP tasks.
7. Evolutionary Algorithms
These algorithms are inspired by biological evolution, such as natural selection and genetic algorithms.
They are used to solve optimization problems.
Common Algorithms:
• Genetic Algorithms: Uses operations like mutation, crossover, and selection to evolve
solutions.
https://colab.research.google.com/drive/1NTsBw7ciS_TPpG61LJjy-qJXMKdsVNdP?usp=sharing