Flexible multiprotocol connectivity (“speaks” S3/Swift, RADOS, RBD, NFS/CIFS), integrates easily with core enterprise applications
Performance scales in parallel with capacity. Which means no I/O pre-planning, since throughput increases with added storage
No Single Point of Failure
Data stays available even with a failure of one or more components, whether it is a drive, controller, NIC, switch, or node
Zero administrator intervention to activate a recovery; system redistributes data across all remaining nodes and drives – automatically – eliminating downtime
Real-time operating system optimized for storage clusters.
Open Source Ceph® Engine
Self-healing architecture delivers unmatched resiliency and maximum availability. Because downtime is not an option.
Simplified installation and management to radically streaming storage administration and cluster expansion.
Aquari provides flexible cluster configurations to support multiple workload types simultaneously. Try our Configurator for yourself and see how Aquari can meet your storage needs:
|Terabytes||Aquari Performance Data Nodes||Small Aquari clusters can scale in 20TB to 60TB increments.|
|Petabytes||Aquari Capacity Data Nodes||Scale larger clusters in 60 to 180 TB increments.|
|Exabytes||Aquari Capacity Data Nodes||Aquari can scale to exabytes - either in one large object pool or in a mix of multi-workload pools to support all of your application workloads.|
|Archive Speed||Aquari Capacity Data Nodes||Deploy capacity optimized nodes and Aquari can support nearline or offline workloads efficiently.|
|Collaborative Rate||Aquari Performance Data Nodes||General collaboration usage rates are easily supported on the Aquari Performance Nodes. Depending on the usage characteristics of the application, you might even be able to use Capacity Data Nodes also.|
|Flashy Fast||Coming Soon||Some of your applications are high I/O. We get that. Aquari is flexible today and will be more flexible in the future. Let's talk directly to see how we can address your needs.|
|Object||Provision Object Pools
Protect with Erasure Coding or Replication.
|Aquari is object storage at the core. The RADOS object libraries are included in every major Linux distribution.|
|S3||File & Obect Services Nodes||Aquari supports a rich set of S3 APIs. If your applications are written to S3 already, no need to re-write to leverage Aquari.|
|File||File & Obect Services Nodes||Aquari support legacy file protocols - CIFS and NFS - to support all of your applications today and when you re-architect them in the future.|
|Cinder||Provision RBD Pool(s)||OpenStack Cinder and Aquari are a natural fit. Aquari uses the Ceph storage engine, which is the de facto block storage for the OpenStack community.|
|Swift||File & Obect Services Nodes||Aquari supports Swift and Cinder to address all OpenStack storage needs. Add Aquari FOS Nodes to deliver Swift object storage for your applications.|
|Backup||Aquari Capacity Data Nodes
File & Obect Services Nodes (optional)
Provision Object Pools
Protect with Erasure Coding
|Aquari's object storage foundation is a perfect solution to your backup needs. With support for CIFS, NFS, S3 and Swift, Aquari probably works with your storage solution already.|
|Collaboration||Aquari Performance Data Nodes||Sync & Share and file workflow applications perform well on Aquari. Depending on the usage characteristics, and the application integration, configure for file or object interfaces to the cluster.|
|Video Streaming||Aquari Performance Data Nodes||Aquari is already battle tested in demanding video streaming enviroments. Deploy with performance nodes and write directly to RADOS to get maximum stream counts.|
|OpenStack||Aquari Performance Data Nodes||OpenStack Cinder and Aquari are a natural fit. Aquari uses the Ceph storage engine, which is the de facto block storage for the OpenStack community.|