Optimizing the Future of Data Storage
Batch Optimization of DNA-Based Archival Storage Systems
This is a technical article. I’d suggest you’d know 2+2, and bio, applied math, and comp sci! But don’t let that intimidate you; anyone can learn. Ultimately, feel free to approach with any level of knowledge, and search for what you don’t understand! Or, you can ask me in the chat for clarifications. To see diagrams in more detail, click on them.
As humanity continues to scale up the amount of collective knowledge being stored on the cloud, there will be a vastly increasing demand for more powerful data storage methods. Unfortunately, we’re encountering a major Rubicon that needs to be crossed, as our traditional storage methods are becoming obsolete.
Most recent data suggest that there currently exists an estimated 8.4 million exabytes of information. As the digital cloud continues to expand as we amalgamate more advanced technologies into our network, and implement more complex architectures for data storage (i.e. blockchain and edge computing paradigms), the information will become intractable.
The quagmire with drives is that there is minimal scalability with large physical storage systems, and their capacity is delimited by their electronic limitations, as typical hard disk drives and computational…