Case Study 207
Case Study 207
oduction
n e
70 a large growth in human life expectancy has been observed in
1 8 7 0
This growth has expanded in the whole world principally due to the
hievements in health care field. As a result, the proportion of elderly
splei
la is rapidly increasing. Aging people in general lives in isolated
ans, In addition to that some of them are not capable to live normally
pditions.
aking will improve their quality of life. For better decision making these
ote monitoring systems needs some regular and trustful information about
tients.
ject Origins:
has
tlfil remote monitoring systems' requirements Jorae uis Reves Ortiz
able to detect and
eloped a complete Human Activity Recognition Syste
in their daily living in
gnize different activities performed by humans
ine mode using smartphones. The recognition part of his system has already
ntested which is capable of predicting activities performed by users.
ESsary datasets of users' movements will be collected from smartphone
srs (accelerometer and gyroscope), processed and then fed to the prediction
del to recognize performed activities.
Ac and ayrc
signats
Machine Leaming
Atcelerometer Signal processing
model SVM Activities predictions
capable of
Features creation predicting the
Datasets generation
Gyroscope pata inputs activity pertormed Additional info such as:
|bythe user The user name
-Time
tphone Recognition
Data Processing Location.
Dched to the user
Mobiie application
Project Overview:
The goal of this project is to build a machine learning model in offline mod.
node
capable of processing signals collected using smart phone inertial sensors ar
producing useful datasets will be used as inputs of a machine learning model
capable of recognizing some of human daily activities (sitting., walking .)
included in the dataset (see datasets and Inputs section) with a low error
rate.
The signal processing pipeline and the final model could be used as a good
source of information about patient's daily activities needed by remote
monitoring systems mentioned earlier
The Experiment
The experiments were carried out with a group of 30 volunteers within an age
bracket of 19-48 years. They performed a protocol of activities
composed of six
basic activities: three static postures (standing,
sitting, lying) and three
dynamic activities (walking, walking downstairs and walking upstairs).
The
experiment also included postural transitions that occurred between the static
postures. These are: stand-to-sit, sit-to-stand, sit-to-lie, lie-to-
sit, stand-to-lie, and lie-to-stand. All the
participants were wearing a
smartphone (Samsung Galaxy S II) on the waist during the experiment
execution. They captured 3-axial linear acceleration and3-axial angular
velocity at a constant rate of 5OHz using the embedded accelerometer and
gyroscope of the device. The experiments were video-recorded to label the
data manually. The resulted dataset is stored and is
considered as the original
labelled dataset and from it different subsets would
be generated.
27thim:
Expenments
mo
Sitting Lying
ppadpondueioery
Conclusion:
From the processed signals presented in the picture above, they extract data
points and activity labels related only to the first six activities which
standing, sitting, lying, walking, walking downstairs and walking
are:
following:
this project we willneed the
Raspberry Pi
One LED
. One 330 Ohm resistor
Jumper wires
Breadboard
below:
snnect the components as shownin the wiring diagram
yvVV
limiting resistors should always
resistor. Current
S0 Ohm resistor is a current limiting LED is connected to a
GPIO pin
LEDs to the GPIO pins. If a n
ed when connecting which c a n damage the Raspberry
a resistor, the LED will
draw too much current,
ut
burn out the LED.
HON CODE:
TugtRPi.GPIO as GPIO
import time
GPIO. setmode (GPIo. BOARD)
GPIO. setup (7, GPIO.OUT)
fo1 X in range ( 0 , 5 ) :
Explanation:
GPIO.setup(7.GPIO.0OUT) -
Bpberry pi.
atroduction to Object Detection for Self Driving Cars:
Rollno:207
Introduction:
ISed.
Building aClassifier:
A classic
approach is to first design a classifier
car
images trom non-car that
differentiate
can
an entire frame
images, and then run that classiler
sampling small patches
acu
that classified as
car
along the way. The paee
atehes
arethe desired
Work
properly, we detections. For this appro tu
must train our
non-car images. classifier to distingush cauand
To train
any classifier,
the two classes we
we need labeled data. cASC
samples of frames
s, we see elaborate shapes and edges that our big brains can
e together to make up a face. But to a computer, they see an
y of many numbers.
many
Convolutional Neural Networks CNNs):are used
Vision because of their amazing ability to inn Computer
understand snspacial TH
information. This means if I had a picture of a
rotate it. move it around, squeeze and
human
human,. even if CC
stretch it, the CNN
recognize that it is a human. would sa 11
L
Convolutional Neural Networks by themselves
image classification: given an image, are
mainly used in
assign a given class. For example, usingenetwork
a CNN
will aaccurate in
lesion dataset, it can then learn
to diagnose
trained on c a
from a given
image. different skin Can
cancers
But what IfI wanted to know more than
specifically, where exactly thejustobjectanisimage belonoc.
class, more if
onclusion
2essence, YOLO locates and classifies objects within an
age/video. Object Detection algorithms like YOLO. combined
th the many other sensors on a self-driving car like Li-Dar. allow
sto build
fully autonomous cars that can drive faster. sater. and
tter than any human can.