Fire Extinguisher Prediction Using Machine Learning Report
Fire Extinguisher Prediction Using Machine Learning Report
learning
ABSTRACT:
Fire is a natural disaster that can be caused by many different reasons. Recently, more
environmentally friendly and innovative extinguishing methods have started to be tested,
some of which are also used. For this purpose, a sound wave fire-extinguishing system was
created and firefighting tests were performed. With the data obtained, as a result of 17,442
tests, a data set was created. In this study, five different machine learning methods were used
by using the data set created. These are artificial neural network, k-nearest neighbor, random
forest, stacking and deep neural network methods. Stacking method is an ensemble method
created by using artificial neural network, k-nearest neighbor, random forest models together.
Classification of extinction and non-extinction states of the flame was made with the models
created with these methods. The accuracy of models in classification should be analyzed in
detail in order to be used as a decision support system in the sound wave fire-extinguishing
system. Hence, the classification processes were carried out through the 10-fold cross-
validation method. As a result of these tests, the performance analysis of the models was
carried out, and the results showed that the highest classification accuracy was determined
96.58% in SVC . The performance of the methods was compared by analyzing the
performance metrics of machine learning methods.
Introduction:-
Fire is a chemical reaction that breaks out with the combination of heat, fuel,
and oxygen components. The heat, gas and smoke resulting from this oxidation
reaction may significantly harm to human and the environment [1]. Early
intervention to the fire facilitates to extinguish. However, depending on the
scale of the fire and the fuel type, fire-extinguishing agents may vary. These
substances in traditional fire-extinguishing techniques may leave chemical
waste and harm human health . Additionally, it can also cause social and
economic damages In order to eliminate these impacts, researches on fire-
extinguishing with renewable energy sources have been carried out. The sound
waves is one of these sources.
CHAPTER 1
OVERVIEW
The main objective is to predict the fire extinguisher using machine learning.
DOMAIN OVERVIEW:
MACHINE LEARNING
Machine Learning is a system that can learn from example through self-improvement and
without being explicitly coded by programmer. The breakthrough comes with the idea that a
machine can singularly learn from the data (i.e., example) to produce accurate results.
Machine learning combines data with statistical tools to predict an output. This output is then
used by corporate to makes actionable insights. Machine learning is closely related to data
mining and Bayesian predictive modeling. The machine receives data as input, use an algorithm
to formulate answers.
A typical machine learning tasks are to provide a recommendation. For those who have a
Netflix account, all recommendations of movies or series are based on the user's historical data.
Tech companies are using unsupervised learning to improve the user experience with
personalizing recommendation.
Machine learning is also used for a variety of task like fraud detection, predictive maintenance,
portfolio optimization, automatize task and so on.
Machine Learning
Machine learning is the brain where all the learning takes place. The way the machine learns
is similar to the human being. Humans learn from experience. The more we know, the more
easily we can predict. By analogy, when we face an unknown situation, the likelihood of
success is lower than the known situation. Machines are trained the same. To make an accurate
prediction, the machine sees an example. When we give the machine a similar example, it can
figure out the outcome. However, like a human, if its feed a previously unseen example, the
machine has difficulties to predict.
The core objective of machine learning is the learning and inference. First of all, the machine
learns through the discovery of patterns. This discovery is made thanks to the data. One crucial
part of the data scientist is to choose carefully which data to provide to the machine. The list
of attributes used to solve a problem is called a feature vector. You can think of a feature
vector as a subset of data that is used to tackle a problem.
The machine uses some fancy algorithms to simplify the reality and transform this discovery
into a model. Therefore, the learning stage is used to describe the data and summarize it into a
model.
For instance, the machine is trying to understand the relationship between the wage of an
individual and the likelihood to go to a fancy restaurant. It turns out the machine finds a positive
relationship between wage and going to a high-end restaurant: This is the model
Inferring
When the model is built, it is possible to test how powerful it is on never-seen-before data. The
new data are transformed into a features vector, go through the model and give a prediction.
This is all the beautiful part of machine learning. There is no need to update the rules or train
again the model. You can use the model previously trained to make inference on new data.
The life of Machine Learning programs is straightforward and can be summarized in the
following points:
1. Define a question
2. Collect data
3. Visualize data
4. Train algorithm
5. Test the Algorithm
6. Collect feedback
7. Refine the algorithm
8. Loop 4-7 until the results are satisfying
9. Use the model to make a prediction
Once the algorithm gets good at drawing the right conclusions, it applies that knowledge to
new sets of data.
Machine learning Algorithms and where they are used?
Machine learning can be grouped into two broad learning tasks: Supervised and Unsupervised.
There are many other algorithms
Supervised learning
An algorithm uses training data and feedback from humans to learn the relationship of given
inputs to a given output. For instance, a practitioner can use marketing expense and weather
forecast as input data to predict the sales of cans.
You can use supervised learning when the output data is known. The algorithm will predict
new data.
Linear Finds a way to correlate each feature to the output to help predict future Regression
regression values.
Logistic Extension of linear regression that's used for classification tasks. The output Classification
regression variable 3is binary (e.g., only black or white) rather than continuous (e.g.,
an infinite list of potential colors)
Decision tree Highly interpretable classification or regression model that splits data- Regression
feature values into branches at decision nodes (e.g., if a feature is a color, Classification
each possible color becomes a new branch) until a final decision output is
made
Naive Bayes The Bayesian method is a classification method that makes use of the Regression
Bayesian theorem. The theorem updates the prior knowledge of an event Classification
with the independent probability of each feature that can affect the event.
Support Support Vector Machine, or SVM, is typically used for the classification Regression
vector task. SVM algorithm finds a hyperplane that optimally divided the classes. very comm
machine It is best used with a non-linear solver. Classification
Random The algorithm is built upon a decision tree to improve the accuracy Regression
forest drastically. Random forest generates many times simple decision trees and Classification
uses the 'majority vote' method to decide on which label to return. For the
classification task, the final prediction will be the one with the most vote;
while for the regression task, the average prediction of all the trees is the
final prediction.
● Classification task
● Regression task
Classification
Imagine you want to predict the gender of a customer for a commercial. You will start gathering
data on the height, weight, job, salary, purchasing basket, etc. from your customer database.
You know the gender of each of your customer, it can only be male or female. The objective
of the classifier will be to assign a probability of being a male or a female (i.e., the label) based
on the information (i.e., features you have collected). When the model learned how to recognize
male or female, you can use new data to make a prediction. For instance, you just got new
information from an unknown customer, and you want to know if it is a male or female. If the
classifier predicts male = 70%, it means the algorithm is sure at 70% that this customer is a
male, and 30% it is a female.
The label can be of two or more classes. The above example has only two classes, but if a
classifier needs to predict object, it has dozens of classes (e.g., glass, table, shoes, etc. each
object represents a class)
Regression
When the output is a continuous value, the task is a regression. For instance, a financial analyst
may need to forecast the value of a stock based on a range of feature like equity, previous stock
performances, macroeconomics index. The system will be trained to estimate the price of the
stocks with the lowest possible error.
Unsupervised learning
In unsupervised learning, an algorithm explores input data without being given an explicit
output variable (e.g., explores customer demographic data to identify patterns)
You can use it when you do not know how to classify the data, and you want the algorithm to
find patterns and classify the data for you
K-means Puts data into some groups (k) that each contains data with similar Clustering
clustering characteristics (as determined by the model, not in advance by humans)
Gaussian mixture A generalization of k-means clustering that provides more flexibility in Clustering
model the size and shape of groups (clusters
Hierarchical Splits clusters along a hierarchical tree to form a classification system. Clustering
clustering
Can be used for Cluster loyalty-card customer
Recommender Help to define the relevant data for making a recommendation. Clustering
system
PCA/T-SNE Mostly used to decrease the dimensionality of the data. The algorithms Dimension
reduce the number of features to 3 or 4 vectors with the highest variances. Reduction
Augmentation:
● Machine learning, which assists humans with their day-to-day tasks, personally or
commercially without having complete control of the output. Such machine learning is
used in different ways such as Virtual Assistant, Data analysis, software solutions. The
primary user is to reduce errors due to human bias.
Automation:
● Machine learning, which works entirely autonomously in any field without the need for
any human intervention. For example, robots performing the essential process steps in
manufacturing plants.
Finance Industry
● Machine learning is growing in popularity in the finance industry. Banks are mainly
using ML to find patterns inside the data but also to prevent fraud.
Government organization
● The government makes use of ML to manage public safety and utilities. Take the
example of China with the massive face recognition. The government uses Artificial
intelligence to prevent jaywalker.
Healthcare industry
● Healthcare was one of the first industry to use machine learning with image detection.
Marketing
● Broad use of AI is done in marketing thanks to abundant access to data. Before the age
of mass data, researchers develop advanced mathematical tools like Bayesian analysis
to estimate the value of a customer. With the boom of data, marketing department relies
on AI to optimize the customer relationship and marketing campaign.
Machine learning gives terrific results for visual pattern recognition, opening up many potential
applications in physical inspection and maintenance across the entire supply chain network.
Unsupervised learning can quickly search for comparable patterns in the diverse dataset. In
turn, the machine can perform quality inspection throughout the logistics hub, shipment with
damage and wear.
For instance, IBM's Watson platform can determine shipping container damage. Watson
combines visual and systems-based data to track, report and make recommendations in real-
time.
In past year stock manager relies extensively on the primary method to evaluate and forecast
the inventory. When combining big data and machine learning, better forecasting techniques
have been implemented (an improvement of 20 to 30 % over traditional forecasting tools). In
term of sales, it means an increase of 2 to 3 % due to the potential reduction in inventory costs.
For example, everybody knows the Google car. The car is full of lasers on the roof which are
telling it where it is regarding the surrounding area. It has radar in the front, which is informing
the car of the speed and motion of all the cars around it. It uses all of that data to figure out not
only how to drive the car but also to figure out and predict what potential drivers around the
car are going to do. What's impressive is that the car is processing almost a gigabyte a second
of data.
Deep Learning
Deep learning is a computer software that mimics the network of neurons in a brain. It is a
subset of machine learning and is called deep learning because it makes use of deep neural
networks. The machine uses different layers to learn from the data. The depth of the model is
represented by the number of layers in the model. Deep learning is the new state of the art in
term of AI. In deep learning, the learning phase is done through a neural network.
Reinforcement Learning
● Q-learning
● Deep Q network
● State-Action-Reward-State-Action (SARSA)
● Deep Deterministic Policy Gradient (DDPG)
Applications/ Examples of deep learning applications
AI in Finance: The financial technology sector has already started using AI to save time,
reduce costs, and add value. Deep learning is changing the lending industry by using more
robust credit scoring. Credit decision-makers can use AI for robust credit lending applications
to achieve faster, more accurate risk assessment, using machine intelligence to factor in the
character and capacity of applicants.
AI in HR: Under Armour, a sportswear company revolutionizes hiring and modernizes the
candidate experience with the help of AI. In fact, Under Armour Reduces hiring time for its
retail stores by 35%. Under Armour faced a growing popularity interest back in 2012. They
had, on average, 30000 resumes a month. Reading all of those applications and begin to start
the screening and interview process was taking too long. The lengthy process to get people
hired and on-boarded impacted Under Armour's ability to have their retail stores fully staffed,
ramped and ready to operate.
At that time, Under Armour had all of the 'must have' HR technology in place such as
transactional solutions for sourcing, applying, tracking and onboarding but those tools weren't
useful enough. Under armour choose HireVue, an AI provider for HR solution, for both on-
demand and live interviews. The results were bluffing; they managed to decrease by 35% the
time to fill. In return, the hired higher quality staffs.
For example, deep-learning analysis of audio allows systems to assess a customer's emotional
tone. If the customer is responding poorly to the AI chatbot, the system can be rerouted the
conversation to real, human operators that take over the issue.
Apart from the three examples above, AI is widely used in other sectors/industries.
Artificial Intelligence
Machine Learning
ML
Deep Learning
Feature Need to understand the features that No need to understand the best feature
engineering represent the data represents the data
Execution time From few minutes to hours Up to weeks. Neural Network needs
compute a significant number of weights
Interpretability Some algorithms are easy to interpret Difficult to impossible
(logistic, decision tree), some are almost
impossible (SVM, XGBoost)
In the table below, we summarize the difference between machine learning and deep learning.
TensorFlow
the most famous deep learning library in the world is Google's TensorFlow. Google product
uses machine learning in all of its products to improve the search engine, translation, image
captioning or recommendations.
To give a concrete example, Google users can experience a faster and more refined the search
with AI. If the user types a keyword a the search bar, Google provides a recommendation about
what could be the next word.
Google wants to use machine learning to take advantage of their massive datasets to give users
the best experience. Three different groups use machine learning:
● Researchers
● Data scientists
● Programmers.
They can all use the same toolset to collaborate with each other and improve their efficiency.
Google does not just have any data; they have the world's most massive computer, so
TensorFlow was built to scale. TensorFlow is a library developed by the Google Brain Team
to accelerate machine learning and deep neural network research.
It was built to run on multiple CPUs or GPUs and even mobile operating systems, and it has
several wrappers in several languages like Python, C++ or Java.
TensorFlow Architecture
It is called Tensor flow because it takes input as a multi-dimensional array, also known
as tensors. You can construct a sort of flowchart of operations (called a Graph) that you want
to perform on that input. The input goes in at one end, and then it flows through this system of
multiple operations and comes out the other end as output.
This is why it is called TensorFlow because the tensor goes in it flows through a list of
operations, and then it comes out the other side.
Development Phase: This is when you train the mode. Training is usually done on your Desktop
or laptop.
Run Phase or Inference Phase: Once training is done Tensorflow can be run on many different
platforms. You can run it on
You can train it on multiple machines then you can run it on a different machine, once you
have the trained model.
The model can be trained and used on GPUs as well as CPUs. GPUs were initially designed
for video games. In late 2010, Stanford researchers found that GPU was also very good at
matrix operations and algebra so that it makes them very fast for doing these kinds of
calculations. Deep learning relies on a lot of matrix multiplication. TensorFlow is very fast at
computing the matrix multiplication because it is written in C++. Although it is implemented
in C++, TensorFlow can be accessed and controlled by other languages mainly, Python.
Finally, a significant feature of Tensor Flow is the Tensor Board. The Tensor Board enables to
monitor graphically and visually what TensorFlow is doing.
LITERATURE SURVEY:
AUTHOR:
J. Mei; K. W. E. Cheng
YEAR:2020
3. Experimental Study on Fire Extinguishing Effect of Water-based Fixed Fire
Extinguishing System in full-Scale Bus Cabin
Abstract:
In recent years, the frequent fire accidents of public transport vehicles have resulted in the
scrapping of vehicles, catastrophic property losses and casualties, which had drawn
widespread concern of people and society for public safety. The passenger cabin space of bus
is relatively small. Once a fire occurred, high temperature, thermal radiation and smoke
generated by combustion will cause great difficulties for personnel evacuation. Aiming at the
requirements of Chinese GA1264-2015, this paper carried out full-scale fire-extinguishing
experiments in the bus cabin. The parameters of temperature distribution, smoke
concentration and fire-fighting time in the passenger cabin of the bus are analyzed, so as to
verify the fire-fighting performance of the self-developed water-based fire extinguishing
agent and the fixed fire-extinguishing system in the passenger cabin. The system can put out
fire quickly and efficiently. Moreover, it can effectively reduce temperature and
concentration of toxic and harmful gases in the cabin after the fire. It could obviously
improve the fire protection capacity.
AUTHOR:
Yanying Cheng; Chunjie Mou; Ke Chen; Hui Bai; Yu Liu; Yuchun Zhang
YEAR:2021
4. The Application of Water Mist Fire Extinguishing System in Bus
Abstract:
Based on the characteristics of bus fire, the applicability of water mist
extinguishing bus fire was analyzed. The structures of self-contained water mist
fire extinguishing system and pump supplied system were summarized. Taking a
12-meter bus as an example, the application of pump supplied water mist fire
extinguishing system using in bus cabin was introduced in detail. The fire
extinguishing efficiency of water mist using in buses was verified by full scale fire
test. The flame was extinguished 11 seconds after the system started and the
average temperature of cabin was 39.9° 58 seconds later. Technical guidance for
the application and design of water mist system using in bus is provided in this
paper.
CHAPTER 3
SYSTEM ANALYSIS
EXISTING SYSTEM:
A poorly structured NN model may cause the model to underfit the training
dataset . On the other hand, exaggeration in restructuring the system to suit every
single item in the training dataset may cause the system to be overfitted . One
possible solution to avoid the Overfitting problem is by restructuring the NN
model in terms of tuning some parameters, adding new neurons to the hidden
layer or sometimes adding a new layer to the network. A NN with a small number
of hidden neurons may not have a satisfactory representational power to model
the complexity and diversity inherent in the data. On the other hand, networks
with too many hidden neurons could overfit the data. However, at a certain stage
the model can no longer be improved, therefore, the structuring process should
be terminated. Hence, an acceptable error rate should be specified when creating
any NN model, which itself is considered a problem since it is difficult to
determine the acceptable error rate a priori . For instance, the model designer may
set the acceptable error rate to a value that is unreachable which causes the model
to stick in local minima or sometimes the model designer may set the acceptable
error rate to a value that can further be improved.
Disadvantage:-
PROPOSED SYSTEM:
CHAPTER 4
SYSTEM ARCHITECTURE
MODULES
1. DATA COLLECTION
2. DATA PRE-PROCESSING
3. FEATURE EXTRATION
4. EVALUATION MODEL
DATA COLLECTION
Data collection is a process in which information is gathered from many sources which is
later used to develop the machine learning models. The data should be stored in a way that
makes sense for problem. In this step the data set is converted into the understandable
format which can be fed into machine learning models.
Data used in this paper is a set of data with features . This step is concerned with selecting
the subset of all available data that you will be working with. ML problems start with data
preferably, lots of data (examples or observations) for which you already know the target
answer. Data for which you already know the target answer is called labelled data.
DATA PRE-PROCESSING
Organize your selected data by formatting, cleaning and sampling from it.
Three common data pre-processing steps are:
● Formatting: The data you have selected may not be in a format that is suitable for you to work
with. The data may be in a relational database and you would like it in a flat file, or the data
may be in a proprietary file format and you would like it in a relational database or a text file.
● Cleaning: Cleaning data is the removal or fixing of missing data. There may be data instances
that are incomplete and do not carry the data you believe you need to address the problem.
These instances may need to be removed. Additionally, there may be sensitive information in
some of the attributes and these attributes may need to be anonymized or removed from the
data entirely.
● Sampling: There may be far more selected data available than you need to work with. More
data can result in much longer running times for algorithms and larger computational and
memory requirements. You can take a smaller representative sample of the selected data that
may be much faster for exploring and prototyping solutions before considering the whole
dataset.
FEATURE EXTRATION
Next thing is to do Feature extraction is an attribute reduction process. Unlike feature
selection, which ranks the existing attributes according to their predictive significance, feature
extraction actually transforms the attributes. The transformed attributes, or features, are linear
combinations of the original attributes. Finally, our models are trained using Classifier
algorithm. We use classify module on Natural Language Toolkit library on Python. We use the
labelled dataset gathered. The rest of our labelled data will be used to evaluate the models.
Some machine learning algorithms were used to classify pre-processed data. The chosen
classifiers were Random forest. These algorithms are very popular in text classification tasks.
EVALUATION MODEL
Model Evaluation is an integral part of the model development process. It helps to find the best
model that represents our data and how well the chosen model will work in the future.
Evaluating model performance with the data used for training is not acceptable in data science
because it can easily generate overoptimistic and over fitted models. There are two methods of
evaluating models in data science, Hold-Out and Cross-Validation. To avoid over fitting, both
methods use a test set (not seen by the model) to evaluate model performance.
Performance of each classification model is estimated base on its averaged. The result will be
in the visualized form. Representation of classified data in the form of graphs.
Accuracy is defined as the percentage of correct predictions for the test data. It can be
calculated easily by dividing the number of correct predictions by the number of total
predictions.
Proposed Approach Steps
1. First, we take input dataset.
2. Filter dataset according to requirements and create a new dataset which has attribute
according to analysis to be done
3. Perform Pre-Processing on the dataset
4. Split the data into training and testing
5. Train the model with training data then analyze testing dataset over classification algorithm
6. Finally you will get results as accuracy metrics
DATA FLOW DIAGRAM
LEVEL 0
Dataset
Collection
Pre-
processing
Random
selection
Trained &
Testing
dataset
LEVEL 1
Dataset
collection
Pre-
processing
Feature
Extraction
Apply
Algorithm
LEVEL 2
Classify
the
dataset
Accuracy
of Result
Prediction of
output of car
price
Finalize the
accuracy of
algorithm
UML DIAGRAMS
The Unified Modeling Language (UML) is used to specify, visualize, modify, construct and
document the artifacts of an object-oriented software intensive system under development.
UML offers a standard way to visualize a system's architectural blueprints, including elements
such as:
● actors
● business processes
● (logical) components
● activities
● programming language statements
UML combines best techniques from data modeling (entity relationship diagrams), business
modeling (work flows), object modeling, and component modeling. It can be used with all
processes, throughout the software development life cycle, and across different implementation
technologies. UML has synthesized the notations of the Booch method, the Object-modeling
technique (OMT) and Object-oriented software engineering (OOSE) by fusing them into a
single, common and widely usable modeling language. UML aims to be a standard modeling
language which can model concurrent and distributed systems.
Sequence Diagram:
Sequence Diagrams Represent the objects participating the interaction horizontally and time
vertically. A Use Case is a kind of behavioral classifier that represents a declaration of an
offered behavior. Each use case specifies some behavior, possibly including variants that the
subject can perform in collaboration with one or more actors. Use cases define the offered
behavior of the subject without reference to its internal structure. These behaviors, involving
interactions between the actor and the subject, may result in changes to the state of the subject
and communications with its environment. A use case can include possible variations of its
basic behavior, including exceptional behavior and error handling.
• Activity Diagrams-:
Usecase diagram:
• UML is a standard language for specifying, visualizing, constructing, and documenting
the artifacts of software systems.
• UML was created by Object Management Group (OMG) and UML 1.0 specification
draft was proposed to the OMG in January 1997.
• OMG is continuously putting effort to make a truly industry standard.
Class diagram
The class diagram is the main building block of object-oriented modeling. It is used for general
conceptual modeling of the systematic of the application, and for detailed modeling translating
the models into programming code. Class diagrams can also be used for data modeling.[1] The
classes in a class diagram represent both the main elements, interactions in the application, and
the classes to be programmed.
In the diagram, classes are represented with boxes that contain three compartments:
The top compartment contains the name of the class. It is printed in bold and centered, and the
first letter is capitalized.
The middle compartment contains the attributes of the class. They are left-aligned and the first
letter is lowercase.
The bottom compartment contains the operations the class can execute. They are also left-
aligned and the first letter is lowercase.
Use Case Diagram
CLASS DIAGRAM
SEQUENCE DIAGRAM
ACTIVITY DIAGRAM
PYTHON OVERVIEW
Python is a high-level, interpreted, interactive and object-oriented scripting language. Python
is designed to be highly readable. It uses English keywords frequently where as other languages
use punctuation, and it has fewer syntactical constructions than other languages.
Python is Interpreted: Python is processed at runtime by the interpreter. You do not
need to compile your program before executing it. This is similar to PERL and PHP.
Python is Interactive: You can actually sit at a Python prompt and interact with the
interpreter directly to write your programs.
History of Python
Python was developed by Guido van Rossum in the late eighties and early nineties at the
National Research Institute for Mathematics and Computer Science in the Netherlands.
Python is derived from many other languages, including ABC, Modula-3, C, C++, Algol-68,
SmallTalk, Unix shell, and other scripting languages.
Python is copyrighted. Like Perl, Python source code is now available under the GNU General
Public License (GPL).
Python is now maintained by a core development team at the institute, although Guido van
Rossum still holds a vital role in directing its progress.
Python Features
Python's features include:
Easy-to-learn: Python has few keywords, simple structure, and a clearly defined syntax. This
allows the student to pick up the language quickly.
Easy-to-read: Python code is more clearly defined and visible to the eyes.
Easy-to-maintain: Python's source code is fairly easy-to-maintain.
A broad standard library: Python's bulk of the library is very portable and cross-platform
compatible on UNIX, Windows, and Macintosh.
Interactive Mode: Python has support for an interactive mode which allows interactive testing
and debugging of snippets of code.
Portable: Python can run on a wide variety of hardware platforms and has the same interface
on all platforms.
Extendable: You can add low-level modules to the Python interpreter. These modules enable
programmers to add to or customize their tools to be more efficient.
GUI Programming: Python supports GUI applications that can be created and ported to many
system calls, libraries, and windows systems, such as Windows MFC, Macintosh, and the X
Window system of Unix.
Scalable: Python provides a better structure and support for large programs than shell scripting.
Apart from the above-mentioned features, Python has a big list of good features, few are listed
below:
It can be used as a scripting language or can be compiled to byte-code for building large
applications.
It provides very high-level dynamic data types and supports dynamic type checking.
Python is available on a wide variety of platforms including Linux and Mac OS X. Let's
understand how to set up our Python environment.
ANACONDA NAVIGATOR
Anaconda Navigator is a desktop graphical user interface (GUI) included in Anaconda
distribution that allows you to launch applications and easily manage conda packages,
environments and channels without using command-line commands. Navigator can search for
packages on Anaconda Cloud or in a local Anaconda Repository. It is available for Windows,
mac OS and Linux.
Why use Navigator?
The command line program conda is both a package manager and an environment manager, to
help data scientists ensure that each version of each package has all the dependencies it requires
and works correctly.
Navigator is an easy, point-and-click way to work with packages and environments without
needing to type conda commands in a terminal window. You can use it to find the packages
you want, install them in an environment, run the packages and update them, all inside
Navigator.
WHAT APPLICATIONS CAN I ACCESS USING NAVIGATOR?
● Jupyter Lab
● Jupyter Notebook
● QT Console
● Spyder
● VS Code
● Glue viz
● Orange 3 App
● Rodeo
● RStudio
Advanced conda users can also build your own Navigator applications
How can I run code with Navigator?
The simplest way is with Spyder. From the Navigator Home tab, click Spyder, and write and
execute your code.
You can also use Jupyter Notebooks the same way. Jupyter Notebooks are an increasingly
popular system that combine your code, descriptive text, output, images and interactive
interfaces into a single notebook file that is edited, viewed and used in a web browser.
What’s new in 1.9?
● Add support for Offline Mode for all environment related actions.
● Add support for custom configuration of main windows links.
● Numerous bug fixes and performance enhancements.
TESTING
● Functional Testing
Functional tests provide systematic demonstrations that functions tested are available as
specified by the business and technical requirements, system documentation, and user manuals.
Functional testing is centered on the following items:
● Functions: Identified functions must be exercised.
● Output: Identified classes of software outputs must be exercised.
● Systems/Procedures: system should work properly
Integration Testing
Software integration testing is the incremental integration testing of two or more integrated
software components on a single platform to produce failures caused by interface defects.
Here in machine learning we are dealing with dataset which is in excel sheet format so if any
test case we need means we need to check excel file. Later on classification will work on the
respective columns of dataset .
Test Case 1 :
CONCLUSION
The development and extinguishing process of kitchen fire in different scenarios are shown in
Fig.11. In oil pan fire scenario, the fire can not be controlled by using dry powder fire
extinguisher, so the fire situation is still expanding after 1 minute of fire extinguishing, while
in kitchen flue fire, the fire situation is controlled after 1 minute of fire extinguishing by
using water-based fire extinguisher specially for edible oil. It can be seen from Fig.11 that the
smoke of oil pan fire and kitchen flue fire containing edible oil is stronger than that of cabinet
fire. The cabinet fire can be easily controlled by dry powder fire extinguishing, while the oil
pan fire and kitchen flue fire need the combination of water-based fire extinguisher and dry
powder fire extinguisher for edible oil to effectively extinguish fire.
REFERENCE:
[1] Y. Y. Li, “Investigation of kitchen fire accidents in the united states, Fire Science and Thechnology,” vol. 32,
pp. 651, June 2013.
[2] T. T. Wei, H. L. Jiang, and Y. Wang, “Analysis on the fire risk of commercial kitchens in China and it’s
countermeasure”, China Safety Science Journal, vol. 19, pp. 134-139, April 2009.
[3] T. Q. Chen, Y. T. Zhang, and Y. R. Liu, “Numerical simulation of restaurant fire under different smoke
exhaust modes based on FDS,” Journal of XI’AN University of Science and Technology, vol. 38, pp. 893-900,
November 2018.
[4] L. Q. Lu, “Study on the extinguishment of class F fires,” Fire Science and Thechnology, vol. 28, pp. 349-353,
May 2009.
[5] X. Y. Xu, X. Y. Liu, and Z. T. Li, “Experimental study on characteristics of edible oil pool fires different in
diameter,” China Safety Science Journal,” vol. 28, pp. 48-52, November 2018.
[6] Z. W. Xie, X. Y. Liang, and Q. Yuan, “Summary of Research on Oil Fume Fire in Catering Industry,” Journal of
Catasitrophology, vol. 24, pp. 89-94, December 2009.