Here at Quobyte we are obsessed with storage! For that reason, we decided to create a short introduction to software storage so that you can get more familiar with it. We have broken down this series of articles into topics so that you can gradually learn about the concepts, protocols, and technologies in enterprise storage.
If you are entirely new to enterprise storage, we recommend starting with the first section, which focuses on general storage. Once you have a general idea of software storage, we recommend exploring the File Systems or Protocols used to manage data. Then, you can learn about taking a storage solution to scale with our scalability articles. Next, you can explore data protection, which is one of the most important aspects of a good software storage solution. Lastly, you can explore storage for Kubernetes.
If you find this content useful and would like to see more, or if you have any questions or general feedback, please fill out our Storage 101 Feedback form, and we will try our best to get back to you as soon as possible. Also, if you want to learn more about Quobyte, please visit our resources page, where you will find case studies, videos, whitepapers, and more! You can also visit our Quobyte’s Blog page to catch up on our latest blogs.
These days anything seems to be software defined. So what does software-defined storage or software storage really mean?
Unified storage refers to storage systems that make data accessible via multiple protocols. Some of these protocols include NFS, SMB, HDFS and others.
Enterprise storage systems are divided into three categories today: SAN, NAS and Object.
The main differences between file and object storage are the access protocol, performance, scalability, and the consistency guarantees they offer.
The NFS protocol was designed to allow several client machines to transparently access a file system on a single server.
The POSIX standard describes how system calls must behave, and to what degree one section defines the semantics (behavior) of a POSIX compatible file system.
The term parallel file system is often used in two ways: The ability to do IO in parallel to multiple servers, and specific IO patterns of parallel HPC applications.
A DFS, as the name suggests, is a file system that is distributed on multiple file servers or locations. Its primary purpose is to reliably store files.
NAS consolidation refers to the process of replacing several NAS systems with a unified, more efficient system that behaves like a virtual system.
Scale-out refers to the ability of a system to scale certain dimensions when you add more components. In a storage or file system - sometimes also called scale-out NAS - these components are hard drives (hard disk, NVMe) and servers.
High-performance computing (HPC) is the field of IT that deals with solving large - often scientific or research problems - using large supercomputers or compute clusters; In other words, by aggregating computing power of many machines.
Hadoop Distributed File System (HDFS for short) is the primary data storage system used by Apache Hadoop applications as a means for managing pools of big data and supporting related big data analytics applications.
Data protection is the process of safeguarding important information from corruption, compromise or loss using technologies like RAID, Erasure Coding or Replication.
Containers are composed of container images as a method to package software and its dependencies, namespaces for resource isolation, and Linux kernel cgroups for resource limits, such as CPUs, and RAM.
Kubernetes, also known as k8s, is an open-source container orchestration system that allows you to automatically manage containerized workloads and services.
Application storage is essential to Kubernetes. It helps you keep your data safe whenever a container crashes, and allows you to share files between containers running in the same Pod.