Unit IV
Unit IV
Unit IV
a. Machine-to-Machine Communication
i. Machine-to-machine (M2M) uses a device (sensor, meter, etc.) to capture an event
(temperature, inventory level, etc.), which is sent through a network (wireless, wired, or
hybrid) to an application (software program) that translates the captured event into a
meaningful information (e.g., items need to be restocked).
ii. With the availability of implementation technologies, a large number of intelligent
machines sharing information and making decisions without direct human intervention is
getting realized.
iii. The M2M applications are various and vast:
◾ Physical and homeland security through connected security and surveillance cameras,
alarm systems, and access control
◾ Object tracking and tracing: Fleet management, supply chain management, order
management, asset tracking and management, road tolling, and traffic optimization or
steering
◾ Automated payment through the integrated point of sales (PoS), vending machines,
gaming consoles, integrated dashboards, and so on
◾ Smarter health care through continuous monitoring vital signs, ambient assisted living
(AAL), telemedicine, remote diagnostics, and so on
◾ Remote diagnostics or maintenance or control of machines, sensors, and actuators,
lighting, pumps, valves, elevators, roadside vending machines, transport vehicles, and so on
◾ Advanced metering infrastructure (AMI) for power, gas, water, heating, grid control,
industrial metering, and so on
◾ Industry automation through production chain monitoring and automation
◾ Home or building or campus networking and automation
◾ Mobile enablement for machine operations
iv. A variety of protocols exist that allow machines to connect and communicate with one
another. Machine data is collected and sent to centralised control and analytical systems in
order to extract and may be utilised for later decision-making and appropriate actuation.
1
v. Cloud-based machine-centric services can be downloaded and installed on a wide range of
ground-level machines to make them more adaptable and clever. Thus, machine connectivity
brings about a wide range of new abilities and capacities for both machines and people.
2
Figure 1: The DPWS (Devices Profile for Web Services) OSI model.
Data acquisition
a. Data generation
Data generates at devices that later on, transfers to the Internet through a gateway. Data
generates as follows:
i. Passive devices data: Data generate at the device or system, following the result of
interactions. A passive device does not have its own power source. An external source helps
such a device to generate and send data. Examples are an RFID or an ATM debit card. The
device may or may not have an associated microcontroller, memory and transceiver. A
contactless card is an example of the former and a label or barcode is the example of the
latter.
ii. Active devices data: Data generates at the device or system or following the result of
interactions. An active device has its own power source. Examples are active RFID,
streetlight sensor or wireless sensor node. An active device also has an associated
microcontroller, memory and transceiver.
iii. Event data: A device can generate data on an event only once. For example, on detection
of the traffic or on dark ambient conditions, which signals the event. The event on darkness
communicates a need for lighting up a group of streetlights. A system consisting of security
cameras can generate data on an event of security breach or on detection of an intrusion. A
waste container with associate circuit can generate data in the event of getting it filled up
90% or above. The components and devices in an automobile generate data of their
performance and functioning.
iv. Device real-time data: An ATM generates data and communicates it to the server
instantaneously through the Internet. This initiates and enables Online Transactions
Processing (OLTP) in real time.
v. Event-driven device data: A device data can generate on an event only once. Examples are:
(a) a device receives command from Controller or Monitor, and then performs action(s) using
3
an actuator. When the action completes, then the device sends an acknowledgement; (b)
when an application seeks the status of a device, then the device communicates the status.
b. Data Acquisition
i. Data acquisition means collecting data from IoT or M2M devices. The data transfers after
the interactions with a data acquisition system of the application. The application interacts
and communicates with a number of devices for acquiring the needed data. The devices send
data on demand or at programmed intervals. Data of devices communicate using the network,
transport and security layers.
ii. When devices have the ability to configure themselves, an application can setup them for
data. The system, for example, can configure devices to send data at predefined intervals. The
frequency of data generation is determined by the device settings. For example, the system
can design an umbrella device to obtain weather data from an Internet weather service once
every week.
iii. Every hour, Vending machine can be programmed to communicate machine sales
statistics and other information. The Vending machine system can be programmed to
communicate instantly in the event of a problem or when a specific chocolate flavour requires
the fill service.
iv. The data-adaptation layer allows application to customize data sending after filtering or
enriching. The application-to-device’s gateway can do transcoding, data management, and
device management. Data management may include data integration, compaction, and fusion.
vi. Device-management software maintains the device ID or address, activation, configuring
registering, deregistering, attaching, and detaching.
c. Data Validation
i. Data acquired from the devices does not mean that data are correct, meaningful or
consistent. Data consistency means within expected range, data or as per pattern or data not
corrupted during transmission. The applications or services depend on valid data. Therefore,
data needs validation checks.
ii. Data validation software do the validation checks on the acquired data. Validation software
applies logic and rules. Then only the analytics, predictions, prescriptions, diagnosis and
decisions can be acceptable.
iii. Large magnitude of data is acquired from a large number of devices, especially, from
machines in industrial plants or embedded components data from large number of
automobiles or health devices in ICUs or wireless sensor networks, and so on.
iv. Therefore, validation software consumes significant resources. An appropriate strategy
needs to be adopted.
Data Integration
i. Data integration is the process of combining data from different sources into a single,
unified view. Data integration ultimately enables analytics tools to produce effective,
actionable business intelligence.
ii. There is no universal approach to data integration. However, data integration solutions
typically involve a few common elements, including a network of data sources, a master
server, and clients accessing data from the master server.
4
iii. In a typical data integration process, the client sends a request to the master server for
data. The master server then intakes the needed data from internal and external sources. The
data is extracted from the sources, then consolidated into a single, organize data set.
Integration helps businesses succeed
iv. Advantages of data integration are Improves collaboration and unification of systems,
saves time and boosts efficiency, reduces errors (and rework), delivers more valuable data.
5
Uniform access integration is a type of data integration that focuses on creating data reliable
when accessed from different sources. The data, however, is left within the original source.
Using this method, object-oriented database management systems can be used to create the
appearance of uniformity between unlike databases.
Common storage integration is the most frequently used approach to storage within data
integration. A copy of data from the original source is kept in the integrated system and
processed for a unified view.
Data store
A data store is a data repository of a set of objects which integrate into the store. Features
of data store are:
i. Objects in a data-store are modeled using Classes which are defined by the database
schemas.
6
ii. A data store is a general concept. It includes data repositories such as database, relational
database, flat file, spreadsheet, mail server, web server, directory services and VMware
iii. A data store may be distributed over multiple nodes. Apache Cassandra is an example of
distributed data store.
iv. A data store may consist of multiple schemas or may consist of data in only one scheme.
Example of only one scheme data store is a relational database. Repository in English means
a group, which can be related upon to look for required things, for special information or
knowledge. For example, a repository of paintings of artists. A database is a repository of
data which can be relied upon for reporting, analytics, process, knowledge discovery and
intelligence. A flat file is another repository. Flat file means a file in which the records have
no structural interrelationship.
Server management
Server management means managing services, setup and maintenance of systems of all types
associated with the server. A server needs to serve around the clock. Server management
includes managing the following:
i. Short reaction times when the system or network is down
ii. High security standards by routinely performing system maintenance and updation
iii. Periodic system updates for state-of-the art setups
iv. Optimised performance
v. Monitoring of all critical services, with SMS and email notifications
vi. Security of systems and protection
vii. Maintaining confidentiality and privacy of data
viii. High degree of security and integrity and effective protection of data, files and databases
at the organisation
ix. Protection of customer data or enterprise internal documents by attackers which includes
spam mails, unauthorised use of the access to the server, viruses, malwares and worms
x. Strict documentation and audit of all activities.
Unstructured data is the data which does not follows to a data model and has no easily
identifiable structure such that it cannot be used by a computer program easily.
7
Unstructured data is not organised in a pre-defined manner or does not have a pre-defined
data model, thus it is not a good fit for a mainstream relational database.
9
Overview: Authentication and Authorization
Authentication and authorization are two critical components in the everyday mission to
secure clients and devices on the Internet. That makes these components essential to any IoT
project because the Internet of Things is simply devices-from simple sensors to complicated
cars and mobile devices-connecting together to share data. These connections must be
secured, and authentication and authorization.
The two concepts have some similarities, but really each one means something very specific
for this discussion:
Authentication is the process of identifying the device. For Message Queuing Telemetry
Transport (MQTT), the process of authentication is to confirm that the device’s client ID is
valid; that is, the ID belongs to the device in question.
Authorization provides a mechanism to bind a specific device to certain permissions. With
Edge Connect, authorization is broken into two tasks:
Binding devices to groups
Binding groups to topics
Authentication
1. Identity (ID) establishment and authentication are essential elements of access control. A
hash function or MD5 gives the irreversible result after many operations on that and the
operations are just one way. The algorithm generates a fixed size, say, 128 or 256-bit hash or
digest value using authentication data and secret key.
2. Only the hash or digest value communicates. The receiver-end receives the value, and
compares that with a stored value. If both are equal then the sender is authenticated.
3. Hash function characteristic features are pre-image resistance, hash function should not
alter, before or after communication and should be as per the previous image (original
message). Second pre-image resistance: hash function should not be altered by an in between
entity (called eavesdropper), should remain the same as one for the previous image (original
message) should be collision-resistance and should not be the same for any form of altered
message.
Authorisation
1. Access control allows only an authorised device or application/service access to a resource,
such as web API input, IoT device, sensor or actuator data or URL.
2. Authorisation model is an essential element of secure access control. The standard
authorisation models are as follows:
i. Access Control List (ACL) for coarse-grain access control
ii. Role-Based Access Control (RBAC) for fine-grain access control
10
iii. Attribute-Based Access Control (ABAC) or other capability-based fine grain access
control An access control server and data communication gateway can be centrally used to
control accesses between application/service and IoT devices. The server central control can
be on a cloud server. Each device can access the server and communicate data to another
server. Alternatively, a distributed architecture enables:
iv. Each device to request access to the server and the server grants application/service access
token
v. Each application/service to request access to the server and the server grants device access
token for the device.
11
Figure 3: Steps during key exchanges, management, authentication and authorisations
followed by secure communication of application/service message to the device/gateway
MD5
https://www.youtube.com/watch?v=S9PMQsbMqUk
https://www.youtube.com/watch?v=Q2H2ndbHUFQ
Additional information
Unstructured data cloud storage system structure design
In order to achieve effective management of unstructured data, many domestic and
foreign companies or individuals be a lot of research. The most important management is
divided into two: one is based on technology, semi-structured data to unstructured data
conversion; the other is unstructured data to structured data conversion, data will eventually
be stored in a relational database in. Unstructured to Structured Data Conversion mostly used
the "unstructured data, structured data half a structured data" gradual conversion. Thus, the
structure of the data obtained through the conversion of its relational database storage and
management. Based on the project requirements, the use of "unstructured data structured data
half a structured data" gradual conversion method and further expand on its basis, the concept
of the standard structure members to implement the data structure of the file name conversion
versatility introduction of templates to save the converted file to extract the file metadata,
create document templates, documents related table to achieve the association unstructured
data with structured data, as shown in Figure 4.
12
Figure 4: Unstructured data cloud storage system structure
System consists of database, file system, template libraries, file format definition module,
metadata extraction module, template creation and management module, intermediate module
data representation and data conversion modules and other components. On the whole system
architecture is divided into three levels: the interface application layer, application logic
layer, data storage layer. Interface application layer provides a graphical interface to the user
data conversion, through the application interface, users can use unstructured to structured
data conversion related operations, without having to be concerned about the specific data
conversion.
Program logic layer consists of five functional modules of the system structure, work
focused on achieving business logic structured to unstructured data conversion system.
Interface application layer client file system after obtaining simulation output file, issue a
request for data conversion, then, the application receives the request sent by the client, will
need to convert the file is passed to the data conversion module. After the module receives
the file, depending on the file type classification to determine which program to use to
convert. Then, five functional modules to work, extract metadata of the file, establish the
appropriate document templates, and then implement the unstructured to semi-structured data
conversion, the processed data is written to the simulation results table in the database.
Application and then convert the result back to the user, and prompts the user whether the
next data conversion, to finalize the whole process of data conversion.
Data storage layer collection system used by the database table, such as document templates,
documents associated table, the simulation results tables. Document templates, documents
13
associated table needs to be created before the system is running. Data simulation result table
is unstructured file data after converted structured data. After the data conversion is
completed, the system will associate the relevant information into the file table.
MD5: Digest is a process which gives the irreversible result involving many operations. A
standard algorithm called MD5 (Message Digest 5) is also used for digest, similar to the hash
value. Receiver-end stores the digest value expected to be obtained after the MD5 operations,
and compares that with received value. If both are equal then the sender message is
authenticated.
14