The StorageOS volume plugin turns your Docker node into a hyper-converged storage platform. Each node that runs the StorageOS plugin can contribute available local or attached storage into a distributed pool, which is then available to all cluster members via a global namespace.
Volumes are available across the cluster so if a container gets moved to another node it still has access to its data. Data can be protected with synchronous replication. Compression, caching, and QoS are enabled by default, and all volumes are thin-provisioned.
No other hardware or software is required, except an optional KV store for multi-node deployments.
During beta, StorageOS is freely available for testing and experimentation. DO NOT USE FOR PRODUCTION DATA. A Developer edition will be free forever.
Full documentation is available at https://docs.storageos.com. To stay informed about new features and production-ready releases, sign up on our customer portal.
To install a single StorageOS node for testing:
$ docker plugin install --alias storageos storageos/plugin KV_BACKEND=boltdb
That's it - you can now start containers with StorageOS-backed volumes:
$ docker run --name postgres01 \
-e PGDATA=/var/lib/postgresql/data/db \
-v postgres01:/var/lib/postgresql/data \
--volume-driver=storageos -d postgres
Or pre-create and manage StorageOS volumes using the docker volume
command:
$ docker volume create \
--driver storageos --opt size=20 --opt storageos.feature.replicas=2 vol01
To install multiple StorageOS nodes in a clustered configuration, see the installation section below.
To get the most out of StorageOS, try:
- Running the CLI to manage volumes, rules, and cluster configuration
- Joining more nodes to the cluster (must use Consul as the KV store)
The docker plugin install
method requires Docker 1.13.0 or above. Older versions (from Docker 1.10.0 onwards) can use the node container install method.
Multi-node StorageOS installations require an external key-value store for configuration data and cluster management. Consul is currently required, though support for etcd is being tested and should be available in the future.
We believe that the KV store is best managed separately from the StorageOS plugin so that the plugin can remain stateless. Most organizations will already be familiar with managing Consul or etcd as they are common components in cloud-native architectures.
For single-node testing, BoltDB is embedded and can be used in place of an external KV store by specifying KV_BACKEND=boltdb
when installing (consul
is the default).
For help setting up Consul, consult the documentation.
The plugin (or node container) should be installed on each Docker node where you want to consume StorageOS volumes or to present capacity to other nodes.
(Optional) NBD is a default Linux kernel module that allows block devices to be run in userspace. Enabling NBD is recommended as it will increase performance for some workloads. To enable the module and increase the number of allowable devices, you must either run:
$ sudo modprobe nbd nbds_max=1024
or add the following line to /etc/modules
:
nbd nbds_max=1024
If Consul is running locally, the defaults will work:
$ docker plugin install --alias storageos storageos/plugin
Plugin "storageos/plugin" is requesting the following privileges:
- network: [host]
- mount: [/var/lib/storageos]
- mount: [/dev]
- device: [/dev/fuse]
- allow-all-devices: [true]
- capabilities: [CAP_SYS_ADMIN]
Do you grant the above permissions? [y/N]
If using Consul for the KV store and it is not local, supply the IP address of the Consul service using the KV_ADDR
parameter:
$ docker plugin install --alias storageos storageos/plugin KV_ADDR=127.0.0.1:8500
Alternatively, to setup a single test StorageOS instance for testing, you can use the built-in BoltDB. Note that each additional StorageOS node will be isolated, so features such as replication and volume failover will not be available.
$ docker plugin install --alias storageos storageos/plugin KV_BACKEND=boltdb
Other configuration parameters (see Configuration Reference below) may be set in a similar way. For most environments, only the KV_ADDR will need to be set if Consul is not running locally on the node.
NOTE: In order to make plugin upgrades easier, install the plugin using --alias storageos
. This ensures that volumes provisioned with a previous version of the plugin continue to function after the upgrade. By default, Docker ties volumes to the plugin version. Volumes should then be provisioned using the alias e.g. --volume-driver storageos
.
Although the default settings should work for most environments, a number of settings are configurable:
HOSTNAME
: Hostname of the Docker node, only if you wish to override it.KV_ADDR
: IP address/port of the Key/Vaue store. Defaults to127.0.0.1:8500
ADVERTISE_IP
: IP address of the Docker node, for incoming connections. Defaults to first non-loopback address.USERNAME
: Username to authenticate to the API with. Defaults tostorageos
.PASSWORD
: Password to authenticate to the API with. Defaults tostorageos
.KV_ADDR
: IP address/port of the Key/Vaue store. Defaults to127.0.0.1:8500
KV_BACKEND
: Type of KV store to use. Defaults toconsul
.boltdb
can be used for single node testing.API_PORT
: Port for the API to listen on. Defaults to5705
(IANA Registered).NATS_PORT
: Port for NATS messaging to listen on. Defaults to4222
.NATS_CLUSTER_PORT
: Port for the NATS cluster service to listen on. Defaults to8222
.SERF_PORT
: Port for the Serf protocol to listen on. Defaults to13700
.DFS_PORT
: Port for DirectFS to listen on. Defaults to17100
.LOG_LEVEL
: One ofdebug
,info
,warning
orerror
. Defaults toinfo
.LOG_FORMAT
: Logging output format, one oftext
orjson
. Defaults tojson
.