No Data Corruption & Data Integrity in Shared Hosting
The integrity of the data which you upload to your new shared hosting account will be guaranteed by the ZFS file system which we employ on our cloud platform. Most of the web hosting suppliers, including our firm, use multiple hard disks to keep content and since the drives work in a RAID, identical data is synchronized between the drives all of the time. When a file on a drive becomes damaged for some reason, yet, it is more than likely that it will be copied on the other drives as other file systems do not offer special checks for that. In contrast to them, ZFS uses a digital fingerprint, or a checksum, for every single file. In the event that a file gets corrupted, its checksum will not match what ZFS has as a record for it, so the damaged copy shall be swapped with a good one from a different hard disk. Due to the fact that this happens right away, there is no risk for any of your files to ever get damaged.
No Data Corruption & Data Integrity in Semi-dedicated Hosting
We have avoided any chance of files getting corrupted silently due to the fact that the servers where your semi-dedicated hosting account will be created use a powerful file system named ZFS. Its basic advantage over other file systems is that it uses a unique checksum for each file - a digital fingerprint that's checked in real time. As we keep all content on numerous SSD drives, ZFS checks whether the fingerprint of a file on one drive matches the one on the other drives and the one it has stored. If there is a mismatch, the bad copy is replaced with a good one from one of the other drives and since this happens instantly, there is no chance that a damaged copy can remain on our website hosting servers or that it can be copied to the other drives in the RAID. None of the other file systems use this kind of checks and what is more, even during a file system check following a sudden blackout, none of them will discover silently corrupted files. In comparison, ZFS will not crash after a power failure and the constant checksum monitoring makes a lenghty file system check obsolete.
