The term data compression refers to decreasing the number of bits of information which should be saved or transmitted. This can be achieved with or without losing data, so what will be removed in the course of the compression will be either redundant data or unneeded one. When the data is uncompressed subsequently, in the first case the information and the quality shall be the same, while in the second case the quality will be worse. There're various compression algorithms which are better for various sort of data. Compressing and uncompressing data generally takes a lot of processing time, so the server performing the action must have adequate resources in order to be able to process the data quick enough. A simple example how information can be compressed is to store just how many consecutive positions should have 1 and how many should have 0 in the binary code instead of storing the actual 1s and 0s.

Data Compression in Shared Hosting

The compression algorithm used by the ZFS file system which runs on our cloud internet hosting platform is known as LZ4. It can improve the performance of any website hosted in a shared hosting account on our end since not only does it compress data more effectively than algorithms used by various file systems, but it also uncompresses data at speeds that are higher than the hard disk drive reading speeds. This can be done by using a lot of CPU processing time, that is not a problem for our platform since it uses clusters of powerful servers working together. One more advantage of LZ4 is that it enables us to make backup copies a lot faster and on reduced disk space, so we shall have a couple of daily backups of your databases and files and their generation will not change the performance of the servers. That way, we could always recover any kind of content that you may have deleted by mistake.