CPDS Cortez Activity 4
CPDS Cortez Activity 4
MW 9:00 – 10:30
1. Some vendors’ incremental backups record only files changed since the last incremental. How
does this affect how restores are done? What are the pros and cons of this technique?
Incremental backups, focusing on changes since the last incremental backup, provide benefits like
faster backups and reduced storage. However, challenges include restore complexity, a higher risk of data
loss, slower restores over time, and increased management demands. Success with this approach hinges
on factors such as backup frequency, retention policies, and overall strategy, requiring careful planning
and testing to address potential issues.
2. What are the benefits and risks of using a backup system that can continue a dump onto a second
tape if the first tape gets full?
Using a backup system that seamlessly continues onto a second tape offers benefits like extended
backup capacity and reduced downtime. However, there are risks, including dependency on tape
sequencing, increased complexity, potential for tape errors, and limited granularity for restores. Effective
management, rigorous procedures, regular testing, and monitoring are essential to mitigate these risks
and ensure the integrity of multi-tape backups.
3. Which aspects of your current backup and restore system should be automated better?
Automating essential elements within a backup and restore system, including scheduling, data
selection, retention policies, monitoring, storage management, restore processes, and other functions, is
crucial to improve efficiency and reliability. This automation promotes consistent operations, minimizes
errors, and allows for adaptability in changing data environments, ultimately enhancing the overall
effectiveness of the backup and restore system.
4. “A modern backup system must be automated to minimize human labor, human thought, human
decisions, and human mistakes.” What is the point being made here?
COLLEGE OF ENGINEERING
The point being made is that automation is vital in contemporary backup systems to lessen
dependence on human labor, decision-making, and the possibility of errors. Through automating tasks
like scheduling, data selection, and monitoring, the system seeks to diminish the need for human
involvement, thereby reducing the likelihood of inconsistencies and mistakes. This not only boosts
efficiency but also lessens the risks linked to human error, ensuring a more dependable and efficient
backup process.
A software repository stores and makes available a library of software packages that can be
installed on client machines. A software repository functions as a centralized hub for distributing
software, enabling users to access, download, and install applications and updates. Widely employed
in operating systems and package management systems, these repositories offer an organized method
for storing and disseminating software packages. The main objective is to streamline the procedures
related to software installation, updates, and maintenance.
A software package is a unified file containing all the elements necessary for installing an
application, patch, or library. These packages facilitate straightforward installation, updating, and
removal of software, handling dependencies to ensure system compatibility. Examples of package
management systems include APT and dpkg for Debian-based systems, RPM for Red Hat-based
systems, and language-specific managers like pip for Python and npm for Node.js.
3. What are some of the differences between the various software package and repository systems
available?
Various software package and repository systems differ in package format, management
commands, repository structure, dependency resolution, system integration, ecosystem specifics,
versioning, security measures, community support, and customization options. These differences have
implications for how users engage with and oversee software packages, underscoring the importance
of comprehending the unique attributes of each system for proficient package management.
A repository for software comprises packages, metadata, an index, a signing key, and an organized
directory structure. Software packages encompass executable files, libraries, configurations,
documentation, dependencies, and metadata. On the client side, interacting with the repository involves
components like a package manager, dependency resolver, update manager, configuration files, local
package database, and a user interface. Grasping these elements is vital for effective software
management and upkeep.
6. If HTTPS provides end-to-end encryption and assures that no bits are changed during transport,
why would digitally signed packages be needed as well?
While HTTPS guarantees secure data transfer through end-to-end encryption, digitally signed
packages are pivotal in validating the authenticity, source, and integrity of software. Digital signatures offer
trust, non-repudiation, and safeguard against unauthorized modifications during transit, enhancing the
security provided by encryption. The combination of HTTPS and digital signatures establishes a holistic
strategy for secure and reliable software distribution.
7. Which access controls do software repositories need and why are they needed?
Access controls in software repositories are essential for governing user permissions and
preventing unauthorized access. They play a crucial role in safeguarding the repository's security and
integrity by overseeing actions like uploading, modifying, or retrieving software packages. The
implementation of these controls is vital to thwart unauthorized modifications, verify package
authenticity, and uphold the repository's overall trustworthiness.
8. When providing a software repository service on behalf of others, what can be done to make it
self-service?
initiatives empower users to navigate, contribute, and participate autonomously, resulting in improved
efficiency and user satisfaction.
1. Which issues need to be considered differently when providing external web services versus internal
services to your organization?
Providing external web services involves unique considerations such as enhanced security against
external threats, managing cross-domain user identities, adhering to external regulations, addressing
performance issues, meeting negotiated SLAs, managing costs, prioritizing user experience for a
broader audience, ensuring high availability, addressing data privacy, integrating with external
systems, handling legal agreements, and establishing robust feedback and support mechanisms.
2. Pick a web service in your organization, and develop a plan for scaling it, assuming that it will
receive 5 times as many queries. Now do the same while assuming it will receive 100 times as
many queries.
To scale a web service anticipating a 5x increase in queries, consider horizontal scaling, caching
strategies, database optimization, CDN usage, and asynchronous processing. For a 100x increase,
additionally explore advanced load balancing, auto-scaling, distributed architecture, serverless
computing, optimized code, data partitioning, global content distribution, advanced caching, database
sharding, and machine learning for predictive scaling. These strategies collectively optimize
performance, ensure reliability, and accommodate varying levels of increased demand.
3. Is your organization’s external web site hosted by a third-party hosting provider? Why or why not?
Evaluate the pros and cons of moving it to a third-party hosting provider or bringing it internal to
your organization.
Hosting Internally:
- Pros: Safeguarding of data privacy and security, and foreseeable expenses, Full control.
- Cons: Difficulties in scalability, possibility of latency, Higher initial investment, and resource
intensiveness.
COLLEGE OF ENGINEERING
The decision between third-party hosting and internal hosting relies on considerations such as
budget, scalability, control needs, and long-term strategy. Many organizations choose a hybrid
approach, considering a comprehensive cost-benefit analysis and their specific organizational
requirements. Remaining knowledgeable about hosting technologies and industry best practices is
essential for continuous optimization.