Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

A Memory-Aware Spark Cache Replacement Strategy

View through CrossRef
<p>Spark is currently the most widely used distributed computing framework, and its key data abstraction concept, Resilient Distributed Dataset (RDD), brings significant performance improvements in big data computing. In application scenarios, Spark jobs often need to replace RDDs due to insufficient memory. Spark uses the Least Recently Used (LRU) algorithm by default as the cache replacement strategy. This algorithm only considers the most recent use time of RDDs as the replacement basis. This characteristic may cause the RDDs that need to be reused to be evicted when performing cache replacement, resulting in a decrease in Spark performance. In response to the above problems, this paper proposes a memory-aware Spark cache replacement strategy, which comprehensively considers the cluster memory usage, RDD size, RDD dependencies, usage times and other information when performing cache replacement and selects the RDDs to be evicted. Furthermore, this paper designs extensive corresponding experiments to test and analyze the performance of the memory-aware Spark cache replacement strategy. The experimental data show that the proposed strategy can improve the performance by up to 13% compared with the LRU algorithm in different scenarios.</p> <p>&nbsp;</p>
Title: A Memory-Aware Spark Cache Replacement Strategy
Description:
<p>Spark is currently the most widely used distributed computing framework, and its key data abstraction concept, Resilient Distributed Dataset (RDD), brings significant performance improvements in big data computing.
In application scenarios, Spark jobs often need to replace RDDs due to insufficient memory.
Spark uses the Least Recently Used (LRU) algorithm by default as the cache replacement strategy.
This algorithm only considers the most recent use time of RDDs as the replacement basis.
This characteristic may cause the RDDs that need to be reused to be evicted when performing cache replacement, resulting in a decrease in Spark performance.
In response to the above problems, this paper proposes a memory-aware Spark cache replacement strategy, which comprehensively considers the cluster memory usage, RDD size, RDD dependencies, usage times and other information when performing cache replacement and selects the RDDs to be evicted.
Furthermore, this paper designs extensive corresponding experiments to test and analyze the performance of the memory-aware Spark cache replacement strategy.
The experimental data show that the proposed strategy can improve the performance by up to 13% compared with the LRU algorithm in different scenarios.
</p> <p>&nbsp;</p>.

Related Results

A Hierarchical Cache Architecture-Oriented Cache Management Scheme for Information-Centric Networking
A Hierarchical Cache Architecture-Oriented Cache Management Scheme for Information-Centric Networking
Information-Centric Networking (ICN) typically utilizes DRAM (Dynamic Random Access Memory) to build in-network cache components due to its high data transfer rate and low latency....
Smart hardware designs for probabilistically-analyzable processor architectures
Smart hardware designs for probabilistically-analyzable processor architectures
Future Critical Real-Time Embedded Systems (CRTES), like those is planes, cars or trains, require more and more guaranteed performance in order to satisfy the increasing performanc...
Design and Optimization of 4-way set Associative Mapped Cache Controller
Design and Optimization of 4-way set Associative Mapped Cache Controller
Abstract: In the realm of modern computer systems, the 4-way set associative mapped cache controller emerges as a cornerstone, revolutionizing memory access efficiency. This explor...
Concurrent Evaluation of Web Cache Replacement and Coherence Strategies
Concurrent Evaluation of Web Cache Replacement and Coherence Strategies
When studying Web cache replacement strategies, it is often assumed that documents are static. Such an assumption may not be realistic, especially when large-size caches are consid...
In-Memory Caching for Enhancing Subgraph Accessibility
In-Memory Caching for Enhancing Subgraph Accessibility
Graphs have been utilized in various fields because of the development of social media and mobile devices. Various studies have also been conducted on caching techniques to reduce ...
C-Aware: A Cache Management Algorithm Considering Cache Media Access Characteristic in Cloud Computing
C-Aware: A Cache Management Algorithm Considering Cache Media Access Characteristic in Cloud Computing
Data congestion and network delay are the important factors that affect performance of cloud computing systems. Using local disk of computing nodes as a cache can sometimes get bet...
Analytical characterization of cache replacement policy impact on content delivery time in information‐centric networks
Analytical characterization of cache replacement policy impact on content delivery time in information‐centric networks
SummaryInformation‐centric networking (ICN) has emerged as a promising candidate for designing content‐based future Internet paradigms. ICN increases the utilization of a network t...
Analisis Variasi Busi Terhadap Performa dan Bahan Bakar Motor Bensin 2 Langkah Yamaha F1ZR 110CC
Analisis Variasi Busi Terhadap Performa dan Bahan Bakar Motor Bensin 2 Langkah Yamaha F1ZR 110CC
Spark plugs have various types and specifications that can improve motorcycle performance. The purpose of this study was to determine the ratio of torque, power, and fuel consumpti...

Back to Top