DataAnalytics Assignment
DataAnalytics Assignment
Data analytics (DA) is the process of examining data sets in order to draw
conclusions about the information they contain, increasingly with the aid of
specialized systems and software. Data analytics technologies and techniques are
widely used in commercial industries to enable organizations to make more-
informed business decisions and by scientists and researchers to verify or
disprove scientific models, theories and hypotheses.
Data analytics initiatives can help businesses increase revenues, improve
operational efficiency, optimize marketing campaigns and customer service
efforts, respond more quickly to emerging market trends and gain a competitive
edge over rivals , all with the ultimate goal of boosting business performance.
Data analytics can also be separated into quantitative data analysis and
qualitative data analysis. The former involves analysis of numerical data with
quantifiable variables that can be compared or measured statistically. The
qualitative approach is more interpretive. it focuses on understanding the content
of non-numerical data like text, images, audio and video, including common
phrases, themes and points of view.
Many businesses are turning to data analytics to provide insight for making
operational decisions. Two areas in particular where data analytics can help
companies is (1) improved service delivery to customers, and (2) more efficient
and effective resource allocation. To arrive at actionable insights, the analysis
often relies on multiple data sets of varying size and content. In this article, we
will discuss one simple example where data engineering, data analysis, and the
merging of two data sets can help a company in both the above areas.
Answering these questions can make a big impact on your business , And
because the tools and techniques used in data analytics are advancing constantly,
new applications for data analytics are always being found.
Three things make operational analytics tough. One is that to make it work,
you have to integrate it with transactional or workflow systems. Two is that you
often have to pull data from a variety of difficult places. And problem three is that
embedding analytics within operational processes means that you have to change
the behavior of the people who perform that process.
Integration
Unfortunately, to succeed with operational analytics, a company has to
combine transaction systems, workflow systems, analytical systems, databases,
and display/user experience tools. Integrating with transactional systems takes a
good deal of effort, although modern systems architectures make it a bit easier.
Most transactional systems these days (including SAP and Oracle ERP systems)
allow API-based connections. But there is usually a fair amount of effort involved
in integrating an operational system sucking out the data you need, doing the
analytics somewhere (the cloud, in-database processing), and embedding the
result into an interface for the front-line user.
Changing Behavior
Measuring responsiveness
A good metric for support staff responsiveness is the amount of time it
takes for a support technician to take a first action in response to a service ticket.
In this scenario, we are imagining that issues are raised through a software
interface that generates a service ticket and that the entire service team has the
ability to respond to tickets in the service queue. The goal of the analysis, then, is
to understand how this metric varies by staffing level and determine if any
adjustments need to be made.
The data table stub below shows a sample of what such a time series table
might look like
Using these data, we can derive another data set that gives the amount of time
elapsed between when a ticket is created and the time of first action. These
derived data (sample shown below) will serve as the basis for our analysis.
This plot shows the average initial response time, by hour of day that a
ticket was created (in red). For comparison, we also show the average number of
tasks opened, by hour of day (in blue). The black dashed line shows the number of
support staff working during each hour of day. Perhaps the most striking feature
of this plot is that large increase in response time for tickets opened between 16h
to 19h (4 p.m. to 7 p.m.); this increase also coincides with a drop in staffing levels
from three people to one person.
The black dashed line shows a linear fit to these data, used to determine if
there is a trend in this relationship. There is also a significant outlier
corresponding to the 18h to 19h period, with an unusually high response time. If
we remove this point when performing the linear fit, we obtain the trend shown
in the solid black line. After removing the outlying point for either fit, does not
indicate a substantial difference (1.4 versus 1.5), and its also apparent by visual
inspection that neither fit provides much predictive value.
For example, based on the fit alone, we might expect that at no ticket
creation rate below 4/hr should we expect the initial response time to exceed five
minutes, but the data clearly show five hours of day when the response time is
greater than five minutes.
Initial response time vs. staffing level
Another relationship we should examine is between the staffing level and
the average initial response time. The plot below indicates a much more apparent
trend where the higher the staffing level, the shorter the initial response time. For
each hour-of-day, the blue dots show the average response time for any given
staffing level. Once again, we perform a simple linear fit to the data with (dashed
black line) and without (solid black line) the outlier corresponding to the 18h to
19h block.
In this case, removing the outlier does appear to have a more significant
impact on the trendline, though the predictive value of both fits are similar. To
use the example above, both fits would suggest that keeping the minimum staff
level at three people or more would result in an average initial response time
below five minutes. There are only two hours of day where these models are
incorrect (11h to 12h and 13h to 14h), and in both of those cases, the average
response time is still below six minutes.
Given this trend, the longer initial response time during the 18h to 19h
block is less surprising, though this hour remains a significant outlier. To better
understand this, we can go back to our initial plot, which showed the data as a
time series. When we do so, it becomes obvious that the increase in initial
response time occurs during the final hour of a three-hour block, where there is
only one person staffed. Analysis of additional data concerning the other duties
the support staff are attending to may offer better insight into this outlier. Initial
hypotheses, though, could be that either the staff member on duty begins to feel
fatigue during her third hour alone, slowing down her overall performance or she
develops a backlog of work from her competing responsibilities, which slows
down her initial response time.
2.https://www.techopedia.com/definition/29495/operational-analytics
3.https://www.slideshare.net/OracleAnalytics/optimizing-manufacturing-operations-using-big-
data-and-analytics?qid=8e2d5f5c-7fd3-448b-b0c7-66fd9edff778&v=&b=&from_search=3
4.http://searchdatamanagement.techtarget.com/definition/data-analytics
5.https://www.oreilly.com/ideas/improving-operations-using-data-analytics