TY - JOUR
T1 - Optimizing Video Quality in Distributed Storage Systems via Deep Reinforcement Learning (DRL)-Based Adaptive Replication
AU - Lee, Dayoung
AU - Song, Minseok
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - Distributed storage systems, such as Hadoop distributed file system (HDFS), are widely used for video storage due to their outstanding scalability. However, they frequently face challenges related to data unavailability arising from issues like network disconnection, server downtimes, and storage failures. This necessitates file replication, leading to significant storage requirements. To address this, we propose a novel deep reinforcement learning (DRL) algorithm that relies on the tradeoff between video quality and storage demands. We first formulate an optimization problem with the objective of maximizing video quality while constraining storage requirements necessary for replication. The video storage system is then modeled with time-varying video streaming workloads as the DRL environment, where the agent determines the placement of replica files without foreknowledge of future storage availability and video popularity. To address this uncertainty, we use a deep double-Q network (D3QN), which includes an action space that finds the number of replicas for each file, an observation space featuring storage utilization and file placement, and a reward model calculating the expected video quality under various data unavailability probabilities. The implementation of our method is examined within the HDFS. Experimental results show that our method improves video quality by up to 39% compared to benchmarks, achieves quality comparable to Oracle when all bitrates are accessible, and even surpasses HDFS's triple redundancy method while using only 20% of the storage space.
AB - Distributed storage systems, such as Hadoop distributed file system (HDFS), are widely used for video storage due to their outstanding scalability. However, they frequently face challenges related to data unavailability arising from issues like network disconnection, server downtimes, and storage failures. This necessitates file replication, leading to significant storage requirements. To address this, we propose a novel deep reinforcement learning (DRL) algorithm that relies on the tradeoff between video quality and storage demands. We first formulate an optimization problem with the objective of maximizing video quality while constraining storage requirements necessary for replication. The video storage system is then modeled with time-varying video streaming workloads as the DRL environment, where the agent determines the placement of replica files without foreknowledge of future storage availability and video popularity. To address this uncertainty, we use a deep double-Q network (D3QN), which includes an action space that finds the number of replicas for each file, an observation space featuring storage utilization and file placement, and a reward model calculating the expected video quality under various data unavailability probabilities. The implementation of our method is examined within the HDFS. Experimental results show that our method improves video quality by up to 39% compared to benchmarks, achieves quality comparable to Oracle when all bitrates are accessible, and even surpasses HDFS's triple redundancy method while using only 20% of the storage space.
KW - deep reinforcement learning (DRL)
KW - distributed file systems
KW - Multimedia systems
KW - streaming media
UR - http://www.scopus.com/inward/record.url?scp=85217689765&partnerID=8YFLogxK
U2 - 10.1109/TCSVT.2025.3538916
DO - 10.1109/TCSVT.2025.3538916
M3 - Article
AN - SCOPUS:85217689765
SN - 1051-8215
JO - IEEE Transactions on Circuits and Systems for Video Technology
JF - IEEE Transactions on Circuits and Systems for Video Technology
ER -