Deep Reinforcement Learning for a Dictionary Based Compression Schema (Student Abstract)

Authors

  • Keren Nivasch Ariel University
  • Dana Shapira Ariel University
  • Amos Azaria Ariel University

Keywords:

Compression, Reinforcement Learning, Lempel–Ziv–Welch Algorithm

Abstract

An increasingly important process of the internet age and the massive data era is file compression. One popular compression scheme, Lempel–Ziv–Welch (LZW), maintains a dictionary of previously seen strings. The dictionary is updated throughout the parsing process by adding new encountered substrings. Klein, Opalinsky and Shapira (2019) recently studied the option of selectively updating the LZW dictionary. They show that even inserting only a random subset of the strings into the dictionary does not adversely affect the compression ratio. Inspired by their approach, we propose a reinforcement learning based agent, RLZW, that decides when to add a string to the dictionary. The agent is first trained on a large set of data, and then tested on files it has not seen previously (i.e., the test set). We show that on some types of input data, RLZW outperforms the compression ratio of a standard LZW.

Downloads

Published

2021-05-18

How to Cite

Nivasch, K., Shapira, D., & Azaria, A. (2021). Deep Reinforcement Learning for a Dictionary Based Compression Schema (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 35(18), 15857-15858. Retrieved from https://ojs.aaai.org/index.php/AAAI/article/view/17925

Issue

Section

AAAI Student Abstract and Poster Program