Image-centric compression of protein structures improves space savings

Background Because of the rapid generation of data, the study of compression algorithms to reduce storage and transmission costs is important to bioinformaticians. Much of the focus has been on sequence data, including both genomes and protein amino acid sequences stored in FASTA files. Current standard practice is to use an ordinary lossless compressor such as gzip on a sequential list of atomic coordinates, but this approach expends bits on saving an arbitrary ordering of atoms, and it also prevents reordering the atoms for compressibility. The standard MMTF and BCIF file formats extend this approach with custom encoding of the coordinates. However, the brand new Foldcomp tool introduces a new paradigm of compressing local angles, to great effect. In this article, we explore a different paradigm, showing for the first time that image-based compression using global angles can also significantly improve compression ratios. To this end, we implement a prototype compressor ‘PIC’, specialized for point clouds of atom coordinates contained in PDB and mmCIF files. PIC maps the 3D data to a 2D 8-bit greyscale image and leverages the well developed PNG image compressor to minimize the size of the resulting image, forming the compressed file. Results PIC outperforms gzip in terms of compression ratio on proteins over 20,000 atoms in size, with a savings over gzip of up to 37.4% on the proteins compressed. In addition, PIC’s compression ratio increases with protein size. Conclusion Image-centric compression as demonstrated by our prototype PIC provides a potential means of constructing 3D structure-aware protein compression software, though future work would be necessary to make this practical.

(macromolecular Crystallographic Information File) files, contain structural information about the protein [6].Although the Protein Data Bank is no longer growing exponentially, the number of new structures deposited is still quite formidable [4]; furthermore, the recent publication of AlphaFold predicted structures has increased total available structures by orders of magnitude [7].
FASTA files are used for storing both protein and genomic sequence information, and much work has been done to create customized sequence compression algorithms.It bears mentioning that the genomic sequence compression literature has recently seen significantly more activity with the advent of next-generation sequencing [8][9][10][11], and many protein sequence compressors take advantage of that work.For protein sequences, [12] introduce a single and double pass version of a amino acid sequence compressor for FASTA files that makes use of substitution matrices.MFCompress was introduced by [13], and converts the amino acid sequences to their corresponding DNA bases, divides the data into three streams, and compresses the resulting streams.CoMSA is another compression algorithm for FASTA files introduced by [14] based on a generalized Burrows-Wheeler transform.Similarly to MFCompress, The Nucleotide Archival Format (NAF) introduced by [15] is another compressor that works on amino acid sequences converted to their corresponding DNA bases by dictionary encoding this transformed string.
In addition to directly transforming and compressing the sequences in FASTA files, a significant amount of research has gone into read-reordering algorithms for genomic sequences in the BEETL [16], SCALCE, [17], MINCE [18], and more.These methods are applicable when FASTA (and the related FASTQ) files are used to store multiple small fragments ('reads') of sequences; next-generation sequencing produces these reads in no particular order, so the reads can be safely reordered without losing important information.When properly performed, this reordering can significantly improve the compression ratio of standard compressors.
On the other hand, the primary data component of PDB and mmCIF protein structure files is a point cloud of coordinates belonging to the atoms that make up the protein.In the standard formats, each atom has its own separate ASCII-formatted line entry in the file that contains the type of atom, type of amino acid to which it belongs, atom and amino acid identifiers, followed by three floating point Cartesian coordinates, along with other information.The coordinates are measured in units of Angstroms Å, where 1 µm = 10,000 Å [19].Unlike their FASTA counterparts, comparatively less work has been done to create compressors customized for the structural data contained in PDB and mmCIF files, though there have been a number of recent tools/formats like MMTF [20], BCIF [21], and the brand new Foldcomp [22].We note with especial interest Foldcomp, which introduces a new paradigm for compressing atomic coordinates using local angles, which is a radical shift from what both MMTF and BCIF do.In this manuscript, we explore yet another different direction in the form of image-centric compression and global angle computations.
Valasatava et al. [23] did a deep investigation on compressing 3D coordinates of atoms in proteins by investigating a full gamut of compression techniques.Their final recommendation was to apply "intramolecular compression", which aims to reduce the size of each protein via three steps: encoding, packing, and entropy compression.The encoding step transforms floating point coordinates into alternate representations, such as Integer, Delta, Predictive, Wavelet, and Unit Vector encodings.Integer encoding as described by [23] multiplies the floating point location coordinates by a power of 10 and rounds the result to the nearest integer.This encoding strategy is lossy when not all decimal places of precision are kept in the integer encoded value, but it can be lossless when used in MMTF and BCIF with a sufficiently large power of 10.However, some amount of loss of precision can be acceptable because of both measurement error, and due to the natural uncertainty of exact atom locations in a protein-PIC will use a lossy variant of integer encoding.Going back to [23], the authors suggest that after packing the encoded coordinate vectors using either recursive indexing or variable packing, the resulting packed coordinates are entropy encoded using standard methods like gzip [24] or brotli [25], which are both combinations of LZ77 dictionary based encoding and Huffman encoding.
However, [23]'s investigation focused primarily on compression of atomic coordinates as sequential objects stored within a text file, treating the data as sequential, much like in FASTA files without reordering.However, unlike protein/genomic sequences, 3D atomic point clouds are not naturally sequential, and the sequence of atoms listed is purely an artifact introduced by using a sequential file format to store the atoms.Thus, preserving the order of the atoms as listed out in a PDB or mmCIF file is largely irrelevant for the purposes of compression, so long as the original information can be reconstructed.Given the background above, one logical next step would be to perform a principled reordering of the atoms to improve compressibility, similar to the technique used by read-reordering algorithms (where again, the order of reads output by the sequencer is inconsequential).The remaining question is of course how to perform that reordering, as point clouds are very different from sequenced genomic reads in underlying structure.
To resolve this question, we turn to an alternate paradigm for compressing point cloud data sets proposed in the field of LIDAR (light detection and ranging) imaging.Houshiar and Nüchter [26] proposed a new compression algorithm for the 5D point cloud data generated by LIDAR scans of real-world scenes.The LIDAR scans produced tuples of data points containing coordinates of a point in space in the scene, along with reflectance and colour data of the surface at that location.Their compression algorithm converts the Cartesian coordinates to spherical coordinates, maps the angular coordinates to the axes of an image, and the radial component, colour, and reflectance data to pixel's fields at the mapped location.The radial component, colour, and reflectance data are written to the R, G, and B components respectively of a single coloured image as well as the greyscale intensity field of three separate consecutive images.The resulting images were compressed using PNG, JPEG 100 (lossless, perfect quality JPEG), JPEG2000, no compression TIFF, LZW TIFF, and Pack Bits TIFF lossless image compressors.The authors of [26] found that compressing three greyscale images using the PNG compressor performed the best in terms of compression ratio.
In this manuscript, we take inspiration from the next-generation sequencing readreordering literature and combine the intramolecular compression techniques of [23] with the image-centric methods of [26].In "Implementation" section, we outline our new compression algorithm, PIC, for the structural protein data contained in PDB and mmCIF files.Design choices and methodology are examined in detail followed by a pseudo-code outline of the compression algorithm.In "Results" section, we give compression results for the atomic coordinates of 20 proteins of a variety of different sizes compressed using both PIC and gzip and show PIC outperforms gzip in terms of compression ratio for proteins over 20,000 atoms in size.We also give the images that constitute the compressed files for a few of the compressed proteins.Furthermore, although PIC is not a full compressor as it does not compress metadata, for the sake of completeness, we also compare PIC file sizes against full compression software MMTF, BCIF, and Foldcomp.In "Discussion" section, we highlight some trends in the compression results and make note of the advantages of the PIC compressor over gzip for structural protein data compression.

Implementation
The PIC compression algorithm has has three main components, namely mapping each atom to a position in an image, encoding information at that position, and compressing the resulting image.A high-level overview is given in Algorithm 1 and Fig. 1, and details are furnished in the following text.

Mapping
Cartesian coordinates of atoms stored in the protein's PDB or mmCIF file are extracted and the global centroid µ of all the coordinates is computed.The coordinates are trans- lated by −µ so that the global centroid becomes the new reference point or origin for the coordinates.This transformation minimizes the instances of collisions when mapping the coordinates to the image.To decompress the images, µ is stored along with the images.
The translated coordinates are then transformed to spherical coordinates.Each spherical coordinate component is rounded to a precision of one decimal place.Valasatava et al. [23] noted that experimental measurements that produce the Cartesian coordinates determine an atom's position with a degree of uncertainty, greater than 0.2 Å .This allows for the exploitation of lossy compression to store the coordinates only up to a tenth of an Å , which is generally sufficient to preserve the essential structural informa- tion provided by lossless representation.The radial r and azimuth φ spherical coordinate components of each atom are posi- tionally encoded to the horizontal and vertical axis of an eight bit pixel greyscale image as follows where ε is a user-defined parameter that sets the number of pixels per azimuth angle degree.Letting r * be the maximal radial component across all spherical coordinates, the dimensions of the resulting image are 10r * × 360ε .Further note that while x ∈ Z ≥0 , y is not necessarily an integer.However, ε is chosen such that 360ε, 8y ∈ Z ≥0 for all y and ε ≥ 1.25 .This ensures there is at least one bit available per tenth of an azimuth angle degree and each y coordinate has an integer bit-level position on the vertical axis.In this way, we view each column in the image as a bit string that is being written to.
Care must be taken when choosing ε .Setting ε too large will produce a large image, degrading the compression ratio.On the contrary, choosing a small ε will induce more collisions when data is mapped to the image.This results in increased compression time, as alternate data storage locations need to be considered.A decrease in the compression ratio may also be experienced in this case as more data points will need pointers to their intended locations and additional images may need to be populated to store all the required data.
The remaining elevation angle θ is stored in the image's pixel intensity values begin- ning at the data point's (x, y) encoding position in the image.Further details on how the elevation angle is formatted or packed and stored in the image is described in "Packing" section.This encoding scheme was selected as it positionally encodes the spherical coordinates r and φ with the largest range of values and encodes the small- est ranging coordinate θ in the image's pixel's intensity values.Thus each coordinate takes up the fewest amount of pixels when encoded into the image, allowing for more data to be stored in the image before another image needs to be generated.
In the event that a data point is mapped to a position that does not have availability to hold all the required information, an alternate encoding position (x * , y * ) is determined systematically.A position (x, y) has availability if all bits at positions between and including (x, y) and (x, y + l − 1/8) , where l is the length of the encoded elevation angle in bytes, have not had data previously written to them.Beginning at the data point's target encoding position (x, y), the positions (x, y + i/8 mod 360ε), 0 < i < 8 • 360ε = 2880ε are scanned subsequently to find the first position with availability.This position is the alternate encoding position.All encoding positions (x, y) also satisfy y < y + l − 1/8 < 360ε , ensuring no data points begin at the bottom of the image and finish at the top to enable proper decompression of the image.
If (x * , y * ) is the alternate encoding position for a data point with target position (x, y), and y ≤ y * < y + 0.1ε , the encoded elevation angle is stored begining at (x * , y * ) as is.Otherwise, a pointer p is encoded and stored along with the encoded elevation angle at (x * , y * ) .p points to the largest y ′ ∈ {i/8|0 ≤ i < 2880ε} that satisfies y ≤ y ′ < y + 0.1ε , namely y ′ = y + 0.1ε − 1/8 .The stored pointer is the integer p = 8(y * − y ′ ) .Note that p > 0 as y * > y ′ .The decompressor then knows that the intended azimuth angle for the data point is that belonging to the position (y * − p/8) = y ′ .
(x, y) = (10r, εφ) In the case that an alternate encoding position cannot be found in the current image, another image is generated, if not already done by a previous data point.The above mapping procedure is repeated in that image to locate an encoding position for the data point.This process repeats until an encoding position is determined for each atom's coordinate.

Packing
Elevation angles are stored in pixels' greyscale values beginning at their corresponding data point's (x, y) encoding position.Each pixel has an 8-bit intensity field.Due to the variable lengths of the binary elevation angles and use of pointers, the following packing scheme is used to store the elevation angles so they can be properly decompressed.
If no pointer is needed, the elevation angle is integer encoded as 10θ and con- verted into its binary representation.If the binary representation has length less than ⌈log 2 (1801)⌉ = 11 bits, 0 bits are added to the front until the representation is 11 bits long.Two additional bits 1 and 0 are added to the front of the resulting binary string in that order to signify the start of a new data point and to notify the decompressor the data point has no pointer, respectively.
If a pointer is required, a similar but expanded packing scheme is used.The second bit is set to 1 instead of 0 to signify to the decompressor that the data point has a pointer.The pointer p is converted to its binary representation and prefixed with 0 bits until it has length ⌈log 2 (2880ε)⌉ .The adjusted binary representations of the pointer and eleva- tion angles follows the two bit prefix in that order.
For 0 ≤ i < 8l , bit i of the packed string is mapped to the bit at position (x, y + i/8) in the image.This packing scheme ensures that each data point has one of two possible lengths, the exact length of which can be determined by the second bit located at (x, y + 1/8) .This is a key feature that allows for the proper decompression of the image.

Cropping and compression
The resulting image(s) are cropped and compressed using the PNG lossless image compressor on the highest compression ratio setting.These image(s) make up the compressed version of the protein's point cloud of atom coordinates in the PDB or mmCIF file.
Images are cropped to remove any all-black rows and columns on the edge of the image.To decompress the images, two cropping parameters are stored along with each image generated to reverse the cropping.
Other lossless image compressors investigated in [26] were also examined.Similarly to the results found by Houshiar at al., PNG was selected for use in the algorithm as it offers the highest compression ratios of the aforementioned compressors at comparable compression times.

Decompressed file
The original and decompressed files are identical up to the coordinates of the atoms.As noted in "Mapping" section, since there is a tolerance of up to 0.2 Å in each coordinate component, each decompressed coordinate is within a euclidean ball of radius 0.2 √ 3 Å about the original coordinate.

Atomic cloud coordinate compression
We benchmarked PIC against the gzip compression after the integer compression/ precision reduction of [23], the primary relevant prior work.Valasatava et al. [23] explored a variety of methods for entropy compression, but we found the differences between methods to be swamped out by the integer encoding step, and thus chose gzip as a representative method for sequential compression.We did not compare against plain gzip for that reason, as the compression ratios without the integer compression were not at all comparable.All further references to gzip are to gzip after the [23] integer compression.
Table 1 gives statistics and compression results on 20 proteins compressed using gzip and PIC where ε = 2.5 and the decompressed files are identical to the original with the lossy coordinate transform.Figures 2 and 3 compare the 3D structures of proteins to the images created by the PIC compressor.Figure 4 compares PIC vs Fig. 2 3D structure [27] and PIC compressor PNG image output for 2ign.Some attributes and symmetries in the 3D structure are observed in the corresponding PIC-compressed image.The upper and lower parts of the 3D structure of protein 2ign can be seen in PIC generated image as two separate masses of black pixels, one over the other Fig. 3 3D structure [28] and PIC compressor PNG image output for 4v60.Some attributes and symmetries in the 3D structure are observed in the corresponding PIC-compressed image.The spiked edge of the 4v60 protein can be seen on the right side of the first outputted image from the PIC compressor Fig. 4 PIC compression ratios plotted against gzip compression ratios (using integer precision reduction for both methods reduced to a tenth of an angstrom for comparability) for each protein compressed in Table 1.Points in the region above the diagonal indicates a protein with better compression ratios using PIC than gzip.Vice versa below the diagonal.PIC demonstrates substantially higher compression ratios for nearly all proteins tested Fig. 5 PIC and gzip compression ratios (using integer precision reduction for both methods as suggested by [23] for comparability) for the proteins compressed in Table 1 plotted against the number of atoms that make up the compressed protein.All but the smallest proteins showed a higher compression ratio when using PIC; for the small proteins, the extra overhead of PIC dominates, but for any large protein, PIC performs better.Note that this comparison is fair to gzip, as instead of gzipping the original files, we only apply gzip after using the same lossy precision encoding that PIC uses; thus, the comparison here is really between sequential storage of a text file using gzip, and spherical storage using PIC gzip compression ratios, whereas Fig. 5 visualizes some of the same results found in Table 1, but plotted against atom size.
As can be seen from Table 1, the proposed PIC algorithm has superior compression ratio performance than the standard gzip text compressor for all proteins over 20,000 atoms in size.This is seen visually in Fig. 5, as all except two points belonging to the two proteins with the fewest number of atoms lie above the diagonal, the region where PIC has better compression ratio performance.In Fig. 5, the gzip compression ratio decays while PIC's compression ratio increases with atom count.
Furthermore, unlike most compression algorithms, we can visually inspect the transformed image because it is itself a projection mapping of the original 3D structure.In Figs. 2 and 3, we show the PIC outputted images.For easier viewing, these images are inverted, five-fold contrast enhanced versions of the actual images outputted by the PIC compressor.
These results were obtained by running the PIC.py script in the command terminal with the "-e" option.These experiments were ran on a Ubuntu 20.04.4 LTS machine with an AMD Ryzen Threadripper 3970X 32-Core Processor and 256 GB of memory in single-thread mode without parallelization.However, the code has also been tested on an Apple MacBook Pro with a 3.5 GHz dual-core processor and 16 GB of memory, with comparable results.Thus, the code can run nearly as well on personal laptops.

Full PDB/mmCIF compression benchmark
Although PIC is designed as a prototype to showcase image-centric compression and thus only compressed the atomic point clouds, it is still instructive to compare against full compression software, such as MMTF, BCIF, and Foldcomp.In order to create a fair comparison, the total metadata space also needs to be included when comparing PICas such, we decided to use use MMTF to compress only the metadata, and then add that size to the size of the PIC image output.This is of course impractical for use as a compressor, but simply serves to level the playing field.
In Table 2, we compare the same benchmark proteins as in Table 1 with original PDB size, BCIF, MMTF, MMTF-lossy, PIC+MMTF-meta, and Foldcomp.MMTF-lossy notably both decreases precision to tenth of an Angstrom (same as PIC), but also only stores the C-alpha atoms, which allows them to take the least space at the cost of not storing all atoms.We were only able to get Foldcomp v0.0.5 to work on one of our proteins, 4v60, because most of our benchmark proteins had discontinuous chains, which is not supported by Foldcomp, and several of the other proteins caused segfaults.However, on both 2jan and 4v60, Foldcomp does substantially better than PIC or any of the other compressors other than MMTF-lossy.

Discussion
As expected, as atom count increases, more images are populated by PIC and more of the image space of the constructed images is used.In addition to the increased data load in only a slightly larger image width wise, this is due to an increased number of collisions as atom count increases.This causes the use of pointers, increasing the average number of bits used per data point, and, when no alternate location can be found in the current image, the population of a new image, increasing the image count.

Table 1 Compression and runtime comparisons of gzip and PIC
PIC compression algorithm, ε = 2.5 , results.Rounded Coordinates Text Size and Binary Size are the sizes of the text and binary files (in kilobytes, i.e. 1000 × bytes, rather than kibibytes), respectively, that contain only the Cartesian coordinates found in the original file, rounded to one decimal place.The binary file (which uses a variable-length encoding) is then gzipped.The gzip and PIC compression ratios (CR) are the ratios of the Rounded Coordinates Text Size to the size the gzip file and PNG image output(s) from the PIC compressor, respectively.Bolded values are the best of gzip and PIC.Compression and decompression times are for the PIC algorithm; note that our code is unoptimized, as the focus is on compression ratios, but we include these times here for completeness.As an aside, (de)compression for gzip takes negligible time for files of this size.We also include RMSD values to measure the lossiness of PIC compression.Image Space Used gives the proportion of the image space that was used to encode the protein coordinate data, or part thereof, in each image constructed by the PIC compressor (for large proteins, more than one image is needed to represent all the atoms)  Furthermore, as noted in the respective figures, our prototype PIC implementation is not optimized for speed.It is not intended as a drop-in replacement for gzip or MMTF, but is instead meant to show that image-centric compression of protein atomic point clouds can provide significant space savings.The Python implementation takes on the order of a few minutes for a single compression/decompression, which is significantly slower than the order of seconds for gzip compression.

Protein
Other values of ε investigated include {1.25, 5, 10} .Only the results for ε = 2.5 are shown as this value produced the best compression ratios.As stated in "Mapping" section, higher ε increased image sizes and consequently decreased the compression ratios.Setting ε = 1.25 increased compression times as collisions increased due to the decreased image size and alternate mapping locations needed to be considered.Compression ratios also decreased slightly as gains in the compression ratios from decreased image sizes were overcame by the additional use of pointers and higher number of images generated by the PIC algorithm.Importantly, in this prototype study, we have given results from only a single set of parameters for all sizes of proteins for principled benchmarking.In future work, it may be preferable to set parameters dynamically for each protein and to store them, as in standard practice in many file compression Table 2 Compression of entire PDB/mmCIF files benchmark Actual compressed file sizes.We compare PIC, MMTF, Foldcomp, and BCIF formats.However, because PIC does not compress atomic metadata, we compressed the metadata-only with MMTF and then added that to the PNG sizes from PIC.All uses of MMTF were followed-up with standard Gzip compression on the MMTF file, as is standard, whereas BCIF is already a fully compressed file format.All sizes are in kilobytes (1000 × bytes).We also include the RMSD for PIC coordinates here.Lastly, unfortunately, most of the proteins we chose for our benchmark were discontinuous, or had other quirks, so Foldcomp 0.0.5 was unable to compress them after running for 24 h.However, Foldcomp does substantially better on both 2jan and 4v60 than all but MMTF-reduced (Cα-lossy), which not only decreases precision but also only keeps the alpha carbons formats.This will play a greater role as the structures of more complex proteins are deconstructed, stored in databases, and transmitted amongst researchers.Further, as the PIC algorithm leverages the standard and widely used PNG image compressor, the algorithm can be easily implemented on a variety of platforms and systems.Although we ran benchmarks, the comparisons against MMTF, BCIF, and Foldcomp were not especially informative for a couple of reasons.First, PIC does not compress metadata, so we had to use MMTF for that portion to create reasonable comparisons.Second, different tools made different design choices about what to focus on: MMTFlossy also only stores the C-alpha atoms, rather than all atoms as PIC and Foldcomp do, and Foldcomp is not designed for discontinuous chains, which were common in our benchmark data set.Still, it does seem that if only C-alpha atomic coordinates are needed, MMTF-lossy is better, and where Foldcomp works, it is also better.

Conclusion
In this paper, we have introduced PIC, an new compression algorithm that leverages positional encoding techniques and the well-developed, widely available PNG image compressor to encode and compress structural protein data in PDB and mmCIF files.The algorithm encodes two of the three dimensions of an atomic coordinate from the point cloud stored in the file to a position in the image space and stores the remaining dimension in pixels' intensity values around that location.The resulting image is then compressed with the lossless image compressor PNG.We showed PIC has a compression ratio superior of that of gzip for proteins with more than 20,000 atoms, and improves with the size of the protein being compressed, reaching up to 37.4% on the proteins we examined.Although as of September 2023, only 6.7% of the structures in the RCSB PDB are over 20,000 atoms, they represent at least 21% of the database in terms of total atom count, so PIC has fair applicability.Additionally, the advent of Cryo-EM in recent years has substantially increased the number of large structures deposited in the PDB, and this growth in the number of large structures is expected to continue.The improvement in compression is also orthogonal to the lossy storage of atomic coordinates to a precision of only a tenth of an angstrom, as we compared against gzip on files with that precision level.
More important than just providing a prototype, we demonstrate in this paper that the paradigm of image-centric compression is superior in efficacy than simply applying a standard sequential compressor to the atomic point clouds.This result is consistent with examples from point-cloud compression in LIDAR imaging, read-reordering for NGS sequence compression, and also the recent Foldcomp compressor which stores internal angles instead.Importantly, this improvement in compression ratios persists even though we store the necessary metadata to undo any atom-reorderings; thus, the only lossy portion of PIC is in coordinate positions.Still, we would recommend that the ordering information be entirely discarded, as it is for LIDAR and read re-ordering-we only kept all of that information to ensure that we performed a fair comparison in our benchmarks.Were we to discard that information, PIC's benchmark results would be even stronger.
We do note that as a prototype implementation, our runtimes and file formats are not suitable for everyday use, but our hope is that future compression algorithms will be designed with our findings in mind.Ultimately, we hope that this study points the way for future image-centric (or more generally structure-aware) compression of protein structures.Indeed, the contemporaneous Foldcomp [22] makes use of internal bond angles and torsions in protein compression, which is a different means of exploiting the 3D structure than our image-centric approach, and that shows greater promise even than PIC-as the Foldcomp software matures, we expect that it will potentially be the benchmark to beat in the future.
The PIC algorithm itself, if reimplemented in a faster language, is certainly competitive on compression ratios already, and furthermore is easy to implement because the PNG image format is already implemented on many platforms.However, we mostly envision that image-centric compression will simply form a part of other more complex compression methods.As structural protein files with increasing complexity are deconstructed, added to databases, and transmitted amongst researchers, targeted compression techniques will become ever more necessary.

Fig. 1
Fig.1Flow chart diagram of the PIC compression algorithm.µ is the global centroid of all atoms used to center the image, and r * is the maximal radial component that needs to be stored after centering.The basic intuition is to store atoms and their coordinate data in a pixel corresponding to the radial coordinates, and then compress with PNG