0% found this document useful (0 votes)
4 views2 pages

Parallel_and_distributed

Parallel and Distributed Computing are essential for efficiently processing and analyzing big data. Parallel Computing breaks down complex problems into smaller tasks for simultaneous execution, offering advantages like speed, scalability, and efficiency, while Distributed Computing allows multiple systems to share resources for handling large datasets. Both approaches face challenges such as energy consumption, coordination, and security concerns.

Uploaded by

Awanish Gupta
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views2 pages

Parallel_and_distributed

Parallel and Distributed Computing are essential for efficiently processing and analyzing big data. Parallel Computing breaks down complex problems into smaller tasks for simultaneous execution, offering advantages like speed, scalability, and efficiency, while Distributed Computing allows multiple systems to share resources for handling large datasets. Both approaches face challenges such as energy consumption, coordination, and security concerns.

Uploaded by

Awanish Gupta
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

Parallel Computing and Distributed Computing in

Big Data
Parallel and Distributed Computing are useful in handing Big data.
The data is ef ciently processed and analysed.

Parallel Computing-

Parallel Computing is a type of computing architecture in which several processors are


simultaneously executed, smaller calculations are broken down from an overall larger complex
problem.

Parallel Computing refers to the process of breaking down larger problem into smaller, independent,
often similar parts that can be executed simultaneously by multiple processors Communicating via
shared memory, This is the rst one, And Second ome

Advantages

• Speed: Reduces the time needed for training and inference.


• Scalability: Handles larger datasets and models effectively.
• Ef ciency: Maximises the utilisation of available computational resources.

Applications
• Image processing: Parallel computing can be used to process large amounts of
images for biomedical imaging.

• Banking: Parallel computing can be used to improve the way the banking sector
operates.

• Solar power: Parallel computing can be used to enhance solar power.

Challenges-
• As parallel systems scale, particularly in high-performance computing environments, energy
consumption and the associated heat dissipation become signi cant concerns.
• Designing energy-efficient parallel algorithms and implementing advanced cooling solutions are
necessary to mitigate these issues

Distributed Computing-

Distributed computing is used in big data as large data can't be stored on a single system so multiple
system with individual memories are used.
Distributed computing allows different users or Computers to share information .
Distributed Computing can allow an application on one machine to leverage processing power,
memory, or storage on another machine.
fi
fi
fi
fi
Applications

• Cloud Computing-

Cloud computing platforms like Amazon Web Services (AWS), Microsoft Azure, and Google Cloud
Platform (GCP) are built on distributed systems.

• Social Media Platforms-

Platforms like Facebook, Twitter, and Instagram use distributed systems to handle millions of users,
posts, and interactions every second. Distributed databases, caching systems, and content delivery
networks are essential components in ensuring fast and reliable service.

• Online Marketplaces

E-commerce platforms like Amazon and eBay use distributed systems to manage product catalogs,
process orders, and handle inventory across multiple locations

Challenges-
• Coordination and synchronization between nodes is necessary to ensure the task is completed
correctly and ef ciently, as well as to keep data consistent and up-to-date.

• This can introduce additional overhead and complexity, requiring sophisticated mechanisms such
as consensus algorithms, distributed locks, or message queues.

• Additionally, security and privacy of data must be addressed to protect against unauthorized
access, modi cation, or leakage when it is transmitted or stored across nodes and networks. This
may involve encryption, authentication, authorization, or auditing techniques that can affect
performance and cost.
fi
fi

You might also like