REFERENCES
[1] [n. d.]. Apache Hadoop. https://hadoop.apache.org/. ([n. d.]).
[2] [n. d.]. Apache Parquet. https://parquet.apache.org/. ([n. d.]).
[3] [n. d.]. Apache Spark. https://spark.apache.org/. ([n. d.]).
[4] [n. d.]. bzip2. http://sourceware.org/bzip2/. ([n. d.]).
[5]
[n. d.]. Census. http://archive.ics.uci.edu/ml/datasets/US+Census+
Data+(1990). ([n. d.]).
[6]
[n. d.]. Corel. http://archive.ics.uci.edu/ml/datasets/Corel+Image+
Features. ([n. d.]).
[7]
[n. d.]. Covtype. http://archive.ics.uci.edu/ml/datasets/covertype. ([n.
d.]).
[8]
[n. d.]. Criteo. http://labs.criteo.com/2013/12/conversion-logs-dataset/.
([n. d.]).
[9] [n. d.]. gzip. https://www.gnu.org/software/gzip/. ([n. d.]).
[10] [n. d.]. Monitor. https://github.com/crottyan/mgbench. ([n. d.]).
[11] [n. d.]. Tesla Autopilot. https://www.tesla.com/autopilot. ([n. d.]).
[12]
Daniel J. Abadi, Samuel Madden, and Miguel Ferreira. 2006. Integrating
compression and execution in column-oriented database systems. In
SIGMOD. 671–682.
[13]
N. Ahmed, T. Natarajan, and K. R. Rao. 1974. Discrete Cosine Trans-
form. IEEE Trans. Computers 23, 1 (1974), 90–93.
[14]
Shivnath Babu, Minos N. Garofalakis, and Rajeev Rastogi. 2001. SPAR-
TAN: A Model-Based Semantic Compression System for Massive Data
Tables. In SIGMOD. 283–294.
[15]
Samy Chambi, Daniel Lemire, Owen Kaser, and Robert Godin. 2014.
Better bitmap performance with Roaring bitmaps. CoRR abs/1402.6407
(2014).
[16]
Yann Collet and Murray S. Kucherawy. 2018. Zstandard Compression
and the application/zstd Media Type. RFC 8478 (2018), 1–54.
[17]
Andrew Crotty, Alex Galakatos, Emanuel Zgraggen, Carsten Binnig,
and Tim Kraska. 2015. Vizdom: Interactive Analytics through Pen and
Touch. PVLDB 8, 12 (2015), 2024–2027.
[18]
Andrew Crotty, Alex Galakatos, Emanuel Zgraggen, Carsten Binnig,
and Tim Kraska. 2016. The case for interactive data exploration accel-
erators (IDEAs). In HILDA@SIGMOD.
[19]
Scott Davies and Andrew W. Moore. 1999. Bayesian Networks for
Lossless Dataset Compression. In SIGKDD. 387–391.
[20]
Peter Deutsch. 1996. DEFLATE Compressed Data Format Specication
version 1.3. RFC 1951 (1996), 1–17.
[21]
Alex Galakatos, Andrew Crotty, Emanuel Zgraggen, Carsten Binnig,
and Tim Kraska. 2017. Revisiting Reuse for Approximate Query Pro-
cessing. PVLDB 10, 10 (2017), 1142–1153.
[22]
Yihan Gao and Aditya G. Parameswaran. 2016. Squish: Near-Optimal
Compression for Archival of Relational Datasets. In SIGKDD. 1575–
1584.
[23]
Georey E. Hinton and Ruslan Salakhutdinov. 2006. Reducing the
Dimensionality of Data with Neural Networks. Science 313, 5786 (2006),
504–507.
[24]
David A. Human. 1952. A Method for the Construction of Minimum-
Redundancy Codes. Proceedings of the IRE 40, 9 (1952), 1098–1101.
[25]
H. V. Jagadish, J. Madar, and Raymond T. Ng. 1999. Semantic Com-
pression and Pattern Extraction with Fascicles. In VLDB. 186–198.
[26]
H. V. Jagadish, Raymond T. Ng, Beng Chin Ooi, and Anthony K. H.
Tung. 2004. ItCompress: An Iterative Semantic Compression Algo-
rithm. In ICDE. 646–657.
[27]
Andrew Lamb, Matt Fuller, Ramakrishna Varadarajan, Nga Tran, Ben
Vandier, Lyric Doshi, and Chuck Bear. 2012. The Vertica Analytic
Database: C-Store 7 Years Later. PVLDB 5, 12 (2012), 1790–1801.
[28]
Daniel Lemire, Owen Kaser, Nathan Kurz, Luca Deri, Chris O’Hara,
François Saint-Jacques, and Gregory Ssi Yan Kai. 2017. Roaring
Bitmaps: Implementation of an Optimized Software Library. CoRR
abs/1709.07821 (2017).
[29]
Mu Li, Wangmeng Zuo, Shuhang Gu, Debin Zhao, and David Zhang.
2018. Learning Convolutional Networks for Content-Weighted Image
Compression. In CVPR. 3214–3223.
[30]
Bernard Marr. 2018. The Amazing Ways Tesla Is Using Articial Intel-
ligence And Big Data. https://www.forbes.com/sites/bernardmarr/
2018/01/08/the-amazing-ways-tesla-is-using-articial-intelligence-
and-big-data/. (2018).
[31]
Vijayshankar Raman and Garret Swart. 2006. How to Wring a Table
Dry: Entropy Compression of Relations and Querying of Compressed
Relations. In VLDB. 858–869.
[32]
Jorma Rissanen. 1978. Modeling by shortest data description. Autom.
14, 5 (1978), 465–471.
[33]
Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis,
Quoc V. Le, Georey E. Hinton, and Je Dean. 2017. Outrageously
Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer.
In ICLR.
[34]
Jasper Snoek, Hugo Larochelle, and Ryan P. Adams. 2012. Practical
Bayesian Optimization of Machine Learning Algorithms. In NIPS. 2960–
2968.
[35]
Michael Stonebraker, Daniel J. Abadi, Adam Batkin, Xuedong Chen,
Mitch Cherniack, Miguel Ferreira, Edmond Lau, Amerson Lin, Samuel
Madden, Elizabeth J. O’Neil, Patrick E. O’Neil, Alex Rasin, Nga Tran,
and Stanley B. Zdonik. 2005. C-Store: A Column-oriented DBMS. In
VLDB. 553–564.
[36]
James A. Storer and Thomas G. Szymanski. 1982. Data compression
via textual substitution. J. ACM 29, 4 (1982), 928–951.
[37]
Nikolaj Tatti and Jilles Vreeken. 2008. Finding Good Itemsets by
Packing Data. In ICDM. 588–597.
[38]
Lucas Theis, Wenzhe Shi, Andrew Cunningham, and Ferenc Huszár.
2017. Lossy Image Compression with Compressive Autoencoders. In
ICLR.
[39]
George Toderici, Sean M. O’Malley, Sung Jin Hwang, Damien Vincent,
David Minnen, Shumeet Baluja, Michele Covell, and Rahul Sukthankar.
2016. Variable Rate Image Compression with Recurrent Neural Net-
works. In ICLR.
[40]
Dmitry Ulyanov, Andrea Vedaldi, and Victor S. Lempitsky. 2018. Deep
Image Prior. In CVPR. 9446–9454.
[41]
Jilles Vreeken. 2009. Making Pattern Mining Useful. Ph.D. Dissertation.
Utrecht University, Netherlands.
[42]
Terry A. Welch. 1984. A Technique for High-Performance Data Com-
pression. IEEE Computer 17, 6 (1984), 8–19.
[43]
Jacob Ziv and Abraham Lempel. 1977. A universal algorithm for
sequential data compression. IEEE Trans. Inf. Theory 23, 3 (1977),
337–343.
[44]
Jacob Ziv and Abraham Lempel. 1978. Compression of individual
sequences via variable-rate coding. IEEE Trans. Inf. Theory 24, 5 (1978),
530–536.
[45]
Marcin Zukowski, Sándor Héman, Niels Nes, and Peter A. Boncz. 2006.
Super-Scalar RAM-CPU Cache Compression. In ICDE. 59.
Research 19: Machine Learning Systems and Applications
SIGMOD ’20, June 14–19, 2020, Portland, OR, USA