CDL Study Guide November - 2021
CDL Study Guide November - 2021
Cloud Digital
Leader Study
Guide
2
Introduction
The Google Cloud Digital Leader training and exam are intended for tech-adjacent individuals
who want to demonstrate an overall knowledge of cloud technology concepts and Google Cloud.
The exam validates a candidate’s ability to complete the following course objectives:
The Cloud Digital Leader exam is job-role independent. The exam assesses the knowledge and skills
of any individuals who want (or are required to) understand the purpose and application of Google
Cloud products.
Sign up for the Cloud Digital Leader Learning Path via Google Cloud Skills Boost, Coursera
or Pluralsight.
Learn more about how and where to take the exam on the Cloud Digital Leader website.
4
Course 1
Introduction to Digital Transformation with Google Cloud
Course 2
Innovating with Data and Google Cloud
Course 3
Infrastructure and Application Modernization with
Google Cloud
Course 4
Understanding Google Cloud Security and Operations
• How cloud technology enables data to Principles and best practices for data
be applied in new ways governance in the cloud
VM VM VM Container orchestration
Infastructure Infastructure
11
Users
Days
Apigee
Speed of Mission
Pace of change
Scale
innovation critical
• Google Cloud solutions for infrastructure Migrate workloads to the public cloud:
modernization and application modernization an essential guide & checklist
Software
Hardware
14
Company
Project 2 Project 3
Glossary
Course 1
The cloud – A metaphor for the network of data centers that store and compute information available
through the Internet. It includes the complex web of software, computers, networks and security
systems involved.
Cloud technology/computing – The technology and processes needed to store, manage, and
access data that is transferred over the Cloud (as opposed to data that remains on your computer’s
hard drive).
Computing – A machine’s ability to process, store, retrieve, compare and analyze information, and
to automate tasks often done by computer programs (otherwise known as software or applications).
Data – Any information that is useful to an organization. Can be numbers on a spreadsheet, text in
an email, audio or video recordings, images, or even ideas in employees’ heads. Includes internal and
external information.
Think 10x – Also ‘generating big ideas’. Finding solutions that make improvements by 10 times, rather
than 10 percent.
Launch and iterate – Also ‘continuous learning’. A mindset and a practice where, instead of starting
off with a perfect solution, you figure it out through experimentation.
17
Course 2
Data point – A piece of information (e.g. a customer purchase or return).
Structured data – Highly organized, quantitative data (e.g. names or credit card numbers).
Easily stored and managed in databases.
Unstructured data – Data that has no organization and tends to be qualitative (e.g. word processing
documents or images). Can be stored as objects, which consist of the data in its native format along
with metadata such as unique identifiers.
Object storage – A data storage architecture for large stores of unstructured data, designating each
piece of data as an object (e.g. audio or multimedia files).
Database – An organized collection of data generally stored in tables and accessed electronically
from a computer system. Built and optimized to enable the efficient ingestion of large amounts of
data from many different sources.
Data integrity – Also known as transaction integrity, this refers to the accuracy and consistency of
data stored in a database.
Data warehouse – The central hub for all business data, it assembles data from multiple sources,
including databases. When combined with connector tools, it can transform unstructured data into
semi-structured data that can be used for analysis. Data warehouses are built to rapidly analyse and
report massive and multi-dimensional datasets on an ongoing basis, in real-time.
18
Data lake – A data lake is a repository designed to store, process, and secure large amounts of
structured, semistructured, and unstructured data. It can store data in its native format and process
any variety of it, ignoring size limits and serves many purposes, such as exploratory data analysis.
Artificial intelligence (AI) – A broad field or term that describes any kind of machine capable
of a task that normally requires human intelligence, such as visual perception, speech recognition,
decision-making, or translation between languages.
Machine learning (ML) – A branch within the field of AI. Computers that can “learn” from data and
make predictions or decisions without being explicitly programmed to do so, using algorithms or
models to analyze data. These algorithms use historical data as input to predict new output values.
Course 3
Virtual machines (VM) – A VM is a virtualized instance of a server that re-creates the functionality
of a dedicated physical server. It uses a partitioned space inside a physical server which makes it
easy to optimize and reallocate resources and allow multiple systems to run on the same hardware.
Hypervisor – The software layer that sits on top of physical hardware. Multiple VMs are built on top
of the hypervisor and are enabled by it.
Container – Follows the same principle as a VM, providing an isolated environment to run
software services and optimize resources from one piece of hardware. Containers are more
efficient than VMs because they do not recreate a full representation of the hardware, but only
recreate or virtualize the operating system.
Serverless computing – A cloud computing execution model in which the cloud provider
allocates machine resources on demand, taking care of the servers on behalf of their customers.
Businesses provide code for the function they want to run and the cloud provider handles all
infrastructure management. Resources such as compute power are automatically provisioned
behind the scenes as needed.
19
Private cloud – When an organization has virtualized servers in its own data centers to create its own
private on-premises environment.
Hybrid cloud – When an organization uses some combination of on-premises or private cloud
infrastructure and public cloud services.
Multi-cloud – When an organization is using multiple public cloud providers as part of its architecture.
Application (or app) – A computer program or software that is designed to carry out a specific digital
task, typically used or run by an end-user. In this digital age, customers expect applications to be
intuitive, well-functioning, and efficient.
Application programming interface (API) – A piece of software that interfaces with or connects
different applications and enables information to flow between systems. In contrast to a user interface,
which connects a computer to a person, an API connects computers or pieces of software to each
other. One purpose of APIs is to hide the internal details of how a system works, exposing only
those parts a developer wants to allow a user or program to interface with. In this way APIs can help
organizations to adapt to modern business needs by allowing access to older legacy systems.
Course 4
Total cost of ownership (TCO) – A comprehensive assessment of all layers within the infrastructure
and other associated costs across the business over time. Includes acquiring hardware and software,
management and support, communications, and user expenses, and the cost of service downtime,
training and other productivity losses.
Shared responsibility model – A model in which the responsibility to secure data is shared
between a business and the cloud provider. The cloud service provider is the data processor, while
the organization is the data controller.
Defense-in-depth – The cloud service provider manages the security of its infrastructure and its
data centers, and customers gain the benefits of their infrastructure’s multiple built in security layers.
Resource hierarchy – How an IT team can organize a business’s Google Cloud environment and
how that service structure maps to the organization’s actual structure. It determines what resources
users can access.
DevOps – Developer operations. A philosophy that seeks to create a more collaborative and
accountable culture within developer and operations teams. Five objectives of DevOps include
reducing silos, accepting failure as normal, implementing gradual change, leveraging tooling and
automation and measuring everything.
SRE – Site reliability engineering. A discipline that applies aspects of software engineering to
operations. The goals of SRE are to create ultra-scalable and highly reliable software systems. Best
practices central to SRE align with DevOps objectives.
Monitoring – Gathering predefined sets of metrics or logs. Monitoring is the foundation for site
reliability engineering because it provides visibility into the performance, uptime, and overall health
of cloud powered applications.
Log file – A text file where applications (including the operating system) write events. Log files
make it easier for developers, DevOps and system administrators to get insights and identify the
root cause of issues within applications and the infrastructure.
Logging – A process that allows IT teams to analyze selected logs and accelerate
application troubleshooting.
21
Cloud SQL – Google Cloud’s database service (relational database management service)
Cloud Spanner – A fully managed Google Cloud database service designed for global scale
Cloud Storage – Google Cloud’s object storage service for structured, semi-structured and structured
data. One of several products used in data lake solutions
TensorFlow – An end-to-end open source platform for machine learning, with a comprehensive,
flexible ecosystem of tools, libraries and community resources, originally created by Google
VMware Engine – An engine for migrating and running VMware workloads natively on Google Cloud
App Engine – A platform for building scalable web applications and mobile back-ends
Cloud Functions – An event-driven compute platform for cloud services and apps
Cost Management – Tools for monitoring, controlling, and optimizing business costs
Cloud Identity – A unified platform for IT administrators to manage user devices and apps
Cloud Console – A web-based interface for managing and monitoring cloud apps
Cloud Monitoring – A tool monitoring infrastructure and application health with rich metrics
Cloud Debugger – A real-time application state inspection and in-production debugging tool
Cloud Profiler – Continuous CPU and heap profiling to improve performance and reduce costs