The OpenStack Cinder driver enables communication between OpenStack
SolidFire storage system. The user can use information from
the SF-Series cluster to configure the driver by modifying the
/etc/cinder/cinder.conf
service file on the controller host.
For more information on the configuration and best practices please visit
the following link: http://www.netapp.com/us/media/tr-4620.pdf
Table 4.18 lists the required storage system attributes used in the
/etc/cinder/cinder.conf
configuration file.
SolidFire Attribute |
Default |
Description |
---|---|---|
|
None |
SolidFire cluster MVIP |
|
None |
SolidFire cluster administrative user |
|
None |
SolidFire cluster administrative password |
Table 4.18. Required SolidFire Attributes
Add the following lines to the file, replacing login and password with the cluster admin login credentials
[solidfire]
volume_backend_name=solidfire
volume_driver=cinder.volume.drivers.solidfire.SolidFireDriver
san_ip=172.17.1.182
san_login=login
san_password=password
[DEFAULT]
enabled_backends=solidfire
You can optionally use the following attributes specific to SolidFire
in the [solidfire]
section of the /etc/cinder/cinder.conf
configuration file to control the interaction between the storage
system and the OpenStack Cinder service. (See Table 4.19.)
SolidFire Attribute |
Default |
Description |
---|---|---|
|
None |
Create storage system accounts with this prefix. The default is no prefix. |
|
True |
Create an internal cache of image copies when a bootable volume is created to eliminate the fetch from OpenStack Glance and QEMU conversion on subsequent calls. |
|
False |
Allow individual tenants to specify QoS on volume creation. Setting this value to True disables dynamic QOS as well as administrative control over QOS. Changing from the default is not recommended. |
|
443 |
The SolidFire storage system API port. Change this value if the device API is behind a proxy on a different port. |
|
True |
Set 512-byte emulation on volume creation. Note: The default is to enable 512 emulation. Required for KVM hypervisors. |
|
False |
Use volume access groups on a per tenant basis instead of using CHAP secrets. |
|
True |
Create an internal mapping of volume IDs and accounts. This option optimizes look-ups and performance at the expense of memory. Use with caution. For very large deployments, this can cause performance issues. |
|
None |
Overrides the default cluster SVIP with the one specified. This option should only be used when multiple OpenStack instances are accessing the storage cluster from non-default VLANs. |
|
|
The account name on the SF-Series cluster to use as the owner of the template and cache volumes. |
|
|
Create volumes on the cluster with this prefix. Volumes use the prefix format |
|
None |
SolidFire Target Cluster for Replication. This option uses the format |
Table 4.19. Optional SolidFire Attributes
Note
Set sf_allow_template_caching to False and use the generalized Cinder global caching feature instead of the SolidFire internal image caching feature.
Tip
NetApp recommends the use of the generalized Cinder global caching feature over the SolidFire internal image cache. As of the Queens release, the default value of sf_allow_template_caching will be set to False. Also in the Queens release, the SolidFire internal image caching feature will be deprecated.
There are several facilities in the SolidFire driver that can accommodate scenarios involving multiple OpenStack cloud instances using a single SolidFire cluster. These features include account prefixes, volume prefixes, virtual networks, and multiple virtual interfaces (VIFs).
You can configure an account prefix using the variable
sf_account_prefix
in the /etc/cinder/cinder.conf
file. You
can set this variable differently for each OpenStack cloud
instance, allowing the storage cluster administrator to
distinguish between accounts for each OpenStack instance.
By default, the sf_account_prefix
variable is not set,
and account names have no prefix.
The OpenStack Mitaka release introduced the sf_volume_prefix
variable to allow creation of unique volume prefixes for each
OpenStack cloud instance.
The default value of the sf_volume_prefix
variable is
UUID-, which automatically configures the driver to work
as it has prior to the OpenStack Mitaka release. If you have multiple
OpenStack cloud instances attached to a single SolidFire system,
NetApp recommends that you change this variable for each instance
using the SolidFire system.
Note
The variable sf_volume_prefix
should not be changed within a
cloud after a volume has been provisioned on the SolidFire storage
system with OpenStack.
You can use VLANs and multiple VIFs to separate multiple OpenStack cloud instances. Element OS supports multiple storage virtual interfaces (SVIPs) on separate virtual networks. Each OpenStack cloud instance accesses the storage system through the shared management port. NetApp recommends that you configure each OpenStack cloud instance with a unique cluster admin account.
Note
When you use a single SVIP with OpenStack, the SolidFire
driver acquires the SVIP by querying the cluster. If multiple SVIP
addresses are configured, the query returns the default SVIP on
the native virtual network. You can configure an alternate virtual
network for the OpenStack cloud instance by modifying the
/etc/cinder/cinder.conf
file. Set the sf_svip
variable in the
[solidfire]
section of the /etc/cinder/cinder.conf
file for that
OpenStack cloud instance to the IP address you want the iSCSI
initiator to use to access volumes on the storage system.
In order to use SolidFire with Replication enabled you must have a secondary
target backend configured and being referenced by primary host under
replication_device
attribute. Example:
[solidfire]
volume_backend_name=solidfire
volume_driver=cinder.volume.drivers.solidfire.SolidFireDriver
san_ip=172.17.1.182
san_login=login
san_password=password
replication_device=backend_id:solidfire2,mvip:172.17.1.142,login:login2,password:password2
[solidfire-2]
volume_backend_name=solidfire2
volume_driver=cinder.volume.drivers.solidfire.SolidFireDriver
san_ip=172.17.1.142
san_login=login2
san_password=password2
[DEFAULT]
enabled_backends=solidfire
Note
The secondary cluster is not required to be in the enabled_backends
like in the example above.
You also need a volume type with replication_enabled=<is> True
set as an
extra-spec:
$ cinder type-show solidfire
+---------------------------------+--------------------------------------+
| Property | Value |
+---------------------------------+--------------------------------------+
| description | None |
| extra_specs | replication_enabled : <is> True |
| | volume_backend_name : solidfire |
| id | 6910843e-0d49-4f8b-84f5-288d3672699d |
| is_public | True |
| name | solidfire |
| os-volume-type-access:is_public | True |
| qos_specs_id | None |
+---------------------------------+--------------------------------------+
When using SolidFire with Replication enabled you can use three different replication modes:
Real-time (Asynchronous): Writes are acknowledged to the client after they are committed on the source cluster.
Real-time (Synchronous): Writes are acknowledged to the client after they are committed on both the source and target clusters.
Snapshot-Only: Only snapshots created on the source cluster are replicated. Active writes from the source volume are not replicated.
The default mode is Real-time (Asynchronous)
, and a new volume type extra-spec
must be set in order to change it. This extras-spec is
solidfire:replication_mode
and its possible values are Sync
, Async
and SnapshotsOnly
. For example:
$ cinder type-show solidfire
+---------------------------------+--------------------------------------+
| Property | Value |
+---------------------------------+--------------------------------------+
| description | None |
| extra_specs | replication_enabled : <is> True |
| | solidfire:replication_mode : Sync |
| | volume_backend_name : solidfire |
| id | 6910843e-0d49-4f8b-84f5-288d3672699d |
| is_public | True |
| name | solidfire |
| os-volume-type-access:is_public | True |
| qos_specs_id | None |
+---------------------------------+--------------------------------------+
This document is licensed under Apache 2.0 license.