Ceph is a distributed storage and network file system designed to provide excellent performance, reliability, and scalability. . This charm provides the RADOS HTTP gateway supporting S3 and Swift protocols for object storage.
Ceph is a unified, distributed storage system designed for excellent performance, reliability, and scalability.
The ceph-radosgw charm deploys the RADOS Gateway, a S3 and Swift compatible HTTP gateway. The deployment is done within the context of an existing Ceph cluster.
This section covers common and/or important configuration options. See file
config.yaml for the full list of options, along with their descriptions and
default values. See the Juju documentation for details
on configuring applications.
pool-type option dictates the storage pool type. See section 'Ceph pool
type' for more information.
source option states the software sources. A common value is an OpenStack
UCA release (e.g. 'cloud:xenial-queens' or 'cloud:bionic-ussuri'). See Ceph
and the UCA. The underlying host's existing apt sources
will be used if this option is not specified (this behaviour can be explicitly
chosen by using the value of 'distro').
Ceph pool type
Ceph storage pools can be configured to ensure data resiliency either through
replication or by erasure coding. This charm supports both types via the
pool-type configuration option, which can take on the values of 'replicated'
and 'erasure-coded'. The default value is 'replicated'.
For this charm, the pool type will be associated with Object storage.
Note: Erasure-coded pools are supported starting with Ceph Luminous.
Replicated pools use a simple replication strategy in which each written object is copied, in full, to multiple OSDs within the cluster.
ceph-osd-replication-count option sets the replica count for any object
stored within the rgw pools. Increasing this value increases data resilience at
the cost of consuming more real storage in the Ceph cluster. The default value
ceph-osd-replication-countoption must be set prior to adding the relation to the ceph-mon application. Otherwise, the pool's configuration will need to be set by interfacing with the cluster directly.
Erasure coded pools
Erasure coded pools use a technique that allows for the same resiliency as replicated pools, yet reduces the amount of space required. Written data is split into data chunks and error correction chunks, which are both distributed throughout the cluster.
Note: Erasure coded pools require more memory and CPU cycles than replicated pools do.
When using erasure coded pools for Object storage multiple pools will be
created: one erasure coded pool ('rgw.buckets.data' for storing actual RGW
data) and several replicated pools (for storing RGW omap metadata). The
ceph-osd-replication-count configuration option only applies to the metadata
Erasure coded pools can be configured via options whose names begin with the
Important: It is strongly recommended to tailor the
ec-profile-moptions to the needs of the given environment. These latter options have default values of '1' and '2' respectively, which result in the same space requirements as those of a replicated pool.
Ceph BlueStore compression
This charm supports BlueStore inline compression
for its associated Ceph storage pool(s). The feature is enabled by assigning a
compression mode via the
bluestore-compression-mode configuration option. The
default behaviour is to disable compression.
The efficiency of compression depends heavily on what type of data is stored in the pool and the charm provides a set of configuration options to fine tune the compression behaviour.
Note: BlueStore compression is supported starting with Ceph Mimic.
To deploy a single RADOS gateway node within an existing Ceph cluster:
juju deploy ceph-radosgw juju add-relation ceph-radosgw:mon ceph-mon:radosgw
Expose the service:
juju expose ceph-radosgw
exposecommand is only required if the backing cloud blocks traffic by default. In general, MAAS is the only cloud type that does not employ firewalling.
The gateway can be accessed over port 80 (as per
juju status ceph-radosgw
By default, Ceph RADOS Gateway puts all tenant buckets into the same global namespace, disallowing multiple tenants to have buckets with the same name. Tenant namespacing can be enabled in this charm by deploying with configuration like:
ceph-radosgw: charm: cs:ceph-radosgw num_units: 1 options: namespace-tenants: True
Enabling tenant namespacing will place all tenant buckets into their own namespace under their tenant id, as well as adding the tenant's ID parameter to the Keystone endpoint registration to allow seamless integration with OpenStack. Tenant namespacing cannot be toggled on in an existing installation as it will remove tenant access to existing buckets. Toggling this option on an already deployed RADOS Gateway will have no effect.
For security reasons the charm is not designed to administer the Ceph cluster. A user (e.g. 'ubuntu') for the Ceph Object Gateway service will need to be created manually:
juju ssh ceph-mon/0 'sudo radosgw-admin user create \ --uid="ubuntu" --display-name="Charmed Ceph"'
Keystone integration (Swift)
Ceph RGW supports Keystone authentication of Swift requests. This is enabled by adding a relation to an existing keystone application:
juju add-relation ceph-radosgw:identity-service keystone:identity-service
When more than one unit is deployed with the hacluster application the charm will bring up an HA active/active cluster.
There are two mutually exclusive high availability options: using virtual IP(s) or DNS. In both cases the hacluster subordinate charm is used to provide the Corosync and Pacemaker backend HA functionality.
This charm supports the use of Juju network spaces (Juju
v.2.0). This feature optionally allows specific types of the application's
network traffic to be bound to subnets that the underlying hardware is
Note: Spaces must be configured in the backing cloud prior to deployment.
API endpoints can be bound to distinct network spaces supporting the network separation of public, internal and admin endpoints.
For example, providing that spaces 'public-space', 'internal-space', and 'admin-space' exist, the deploy command above could look like this:
juju deploy ceph-radosgw \ --bind "public=public-space internal=internal-space admin=admin-space"
Alternatively, configuration can be provided as part of a bundle:
ceph-radosgw: charm: cs:ceph-radosgw num_units: 1 bindings: public: public-space internal: internal-space admin: admin-space
Note: Existing ceph-radosgw units configured with the
os-admin-hostnameoptions will continue to honour them. Furthermore, these options override any space bindings, if set.
This section lists Juju actions supported by the charm.
Actions allow specific operations to be performed on a per-unit basis. To
display action descriptions run
juju actions ceph-radosgw. If the charm is
not deployed then see file
Please report bugs on Launchpad.
For general charm questions refer to the OpenStack Charm Guide.
- (string) Comma-separated list of Swift admin roles; used when integrating with OpenStack Keystone. Admin roles can set the user quota amount.
- (string) Compressor to use (if any) for pools requested by this charm. . NOTE: The ceph-osd charm sets a global default for this value (defaults to 'lz4' unless configured by the end user) which will be used unless specified for individual pools.
- (int) Chunks larger than this are broken into smaller blobs sizing bluestore compression max blob size before being compressed on pools requested by this charm.
- (int) Value of bluestore compression max blob size for rotational media on pools requested by this charm.
- (int) Value of bluestore compression max blob size for solid state media on pools requested by this charm.
- (int) Chunks smaller than this are never compressed on pools requested by this charm.
- (int) Value of bluestore compression min blob size for rotational media on pools requested by this charm.
- (int) Value of bluestore compression min blob size for solid state media on pools requested by this charm.
- (string) Policy for using compression on pools requested by this charm. . 'none' means never use compression. 'passive' means use compression when clients hint that data is compressible. 'aggressive' means use compression unless clients hint that data is not compressible. 'force' means use compression under all circumstances even if the clients hint that the data is not compressible.
- (float) The ratio of the size of the data chunk after compression relative to the original size must be at least this small in order to store the compressed version on pools requested by this charm.
- (int) Number of keystone tokens to hold in local cache.
- (int) This value dictates the number of replicas ceph must make of any object it stores within RGW pools. Note that once the RGW pools have been created, changing this value will not have any effect (although it can be changed in ceph by manually configuring your ceph cluster).
- (string) User provided Ceph configuration. Supports a string representation of a python dictionary where each top-level key represents a section in the ceph.conf template. You may only use sections supported in the template. . WARNING: this is not the recommended way to configure the underlying services that this charm installs and is used at the user's own risk. This option is mainly provided as a stop-gap for users that either want to test the effect of modifying some config or who have found a critical bug in the way the charm has configured their services and need it fixed immediately. We ask that whenever this is used, that the user consider opening a bug on this charm at http://bugs.launchpad.net/charms providing an explanation of why the config was needed so that we may consider it for inclusion as a natively supported config in the charm.
- (boolean) Use DNS HA with MAAS 2.0. Note if this is set do not set vip settings below.
- (string) (lrc plugin) The type of the crush bucket in which each set of chunks defined by l will be stored. For instance, if it is set to rack, each group of l chunks will be placed in a different rack. It is used to create a CRUSH rule step such as step choose rack. If it is not set, no such grouping is done.
- (string) Device class from CRUSH map to use for placement groups for erasure profile - valid values: ssd, hdd or nvme (or leave unset to not use a device class).
- (int) (shec plugin - c) The number of parity chunks each of which includes each data chunk in its calculation range. The number is used as a durability estimator. For instance, if c=2, 2 OSDs can be down without losing data.
- (int) (clay plugin - d) Number of OSDs requested to send data during recovery of a single chunk. d needs to be chosen such that k+1 <= d <= k+m-1. Larger the d, the better the savings.
- (int) Number of data chunks that will be used for EC data pool. K+M factors should never be greater than the number of available zones (or hosts) for balancing.
- (int) (lrc plugin - l) Group the coding and data chunks into sets of size l. For instance, for k=4 and m=2, when l=3 two groups of three are created. Each set can be recovered without reading chunks from another set. Note that using the lrc plugin does incur more raw storage usage than isa or jerasure in order to reduce the cost of recovery operations.
- (int) Number of coding chunks that will be used for EC data pool. K+M factors should never be greater than the number of available zones (or hosts) for balancing.
- (string) Name for the EC profile to be created for the EC pools. If not defined a profile name will be generated based on the name of the pool used by the application.
- (string) EC plugin to use for this applications pool. The following list of plugins acceptable - jerasure, lrc, isa, shec, clay.
- (string) (clay plugin) specifies the plugin that is used as a building block in the layered construction. It can be one of jerasure, isa, shec (defaults to jerasure).
- (string) EC profile technique used for this applications pool - will be validated based on the plugin configured via ec-profile-plugin. Supported techniques are ‘reed_sol_van’, ‘reed_sol_r6_op’, ‘cauchy_orig’, ‘cauchy_good’, ‘liber8tion’ for jerasure, ‘reed_sol_van’, ‘cauchy’ for isa and ‘single’, ‘multiple’ for shec.
- (string) Name of the metadata pool to be created (for RBD use-cases). If not defined a metadata pool name will be generated based on the name of the data pool used by the application. The metadata pool is always replicated, not erasure coded.
- (string) Default network interface on which HA cluster will bind to communication with the other members of the HA Cluster.
- (int) Default multicast port number that will be used to communicate between HA Cluster nodes.
- (int) Client timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 90000ms is used.
- (int) Connect timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 9000ms is used.
- (int) Queue timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 9000ms is used.
- (int) Server timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 90000ms is used.
- (string) Apply system hardening. Supports a space-delimited list of modules to run. Supported modules currently include os, ssh, apache and mysql.
- (string) Key ID to import to the apt keyring to support use with arbitary source configuration from outside of Launchpad archives or PPA's.
- (int) RadosGW debug level. Max is 20.
- (string) Used by the nrpe-external-master subordinate charm. A string that will be prepended to instance name to set the host name in nagios. So for instance the hostname would be something like: . juju-myservice-0 . If you're running multiple environments with the same services in them this allows you to differentiate between them.
- (string) A comma-separated list of nagios servicegroups. If left empty, the nagios_context will be used as the servicegroup
- (boolean) Enable tenant namespacing. If tenant namespacing is enabled, keystone tenants will be implicitly added to a matching tenant in radosgw, in addition to updating the catalog URL to allow radosgw to support publicly-readable containers and temporary URLS. This namespacing also allows multiple tenants to create buckets with the same names, as the bucket names are namespaced into the tenant namespaces in the RADOS gateway. This configuration option will not be enabled on a charm upgrade, and cannot be toggled on in an existing installation as it will remove tenant access to existing buckets.
- (string) Comma-separated list of Swift operator roles; used when integrating with OpenStack Keystone.
- (string) The hostname or address of the admin endpoints created for ceph-radosgw in the keystone identity provider. . This value will be used for admin endpoints. For example, an os-admin-hostname set to 'files.admin.example.com' with will create the following admin endpoint for the ceph-radosgw: . https://files.admin.example.com:80/swift/v1
- (string) The IP address and netmask of the OpenStack Admin network (e.g. 192.168.0.0/24) . This network will be used for admin endpoints.
- (string) The hostname or address of the internal endpoints created for ceph-radosgw in the keystone identity provider. . This value will be used for internal endpoints. For example, an os-internal-hostname set to 'files.internal.example.com' with will create the following internal endpoint for the ceph-radosgw: . https://files.internal.example.com:80/swift/v1
- (string) The IP address and netmask of the OpenStack Internal network (e.g. 192.168.0.0/24) . This network will be used for internal endpoints.
- (string) The hostname or address of the public endpoints created for ceph-radosgw in the keystone identity provider. . This value will be used for public endpoints. For example, an os-public-hostname set to 'files.example.com' with will create the following public endpoint for the ceph-radosgw: . https://files.example.com:80/swift/v1
- (string) The IP address and netmask of the OpenStack Public network (e.g. 192.168.0.0/24) . This network will be used for public endpoints.
- (string) DEPRECATED, use zone instead - pool name can be inherited from the zone config option. The rados gateway stores objects in many different pools. If you would like to have multiple rados gateways each pointing to a separate set of pools set this prefix. The charm will then set up a new set of pools. If your prefix has a dash in it that will be used to split the prefix into region and zone. Please read the documentation on federated rados gateways for more information on region and zone.
- (string) Ceph pool type to use for storage - valid values include ‘replicated’ and ‘erasure-coded’.
- (int) The port that the RADOS Gateway will listen on. . The default is 80 when no TLS is configured and 443 when TLS is configured.
- (boolean) If True enables IPv6 support. The charm will expect network interfaces to be configured with an IPv6 address. If set to False (default) IPv4 is expected. . NOTE: these charms do not currently support IPv6 privacy extension. In order for this charm to function correctly, the privacy extension must be disabled and a non-temporary address must be configured/available on your network interface.
- (string) Name of RADOS Gateway Realm to create for multi-site replication. Setting this option will enable support for multi-site replication, at which point the zonegroup and zone options must also be provided.
- (string) OpenStack region that the RADOS gateway supports; used when integrating with OpenStack Keystone.
- (boolean) Optionally restrict Ceph key permissions to access pools as required.
- (int) Defines a relative weighting of the pool as a percentage of the total amount of data in the Ceph cluster. This effectively weights the number of placement groups for the pool created to be appropriately portioned to the amount of data expected. For example, if the amount of data loaded into the RADOS Gateway/S3 interface is expected to be reserved for or consume 20% of the data in the Ceph cluster, then this value would be specified as 20.
- (int) When the Rados Gatway is installed it, by default, creates pools with pg_num 8 which, in the majority of cases is suboptimal. A few rgw pools tend to carry more data than others e.g. .rgw.buckets tends to be larger than most. So, for pools with greater requirements than others the charm will apply the optimal value i.e. corresponding to the number of OSDs up+in the cluster at the time the pool is created. For others it will use this value which can be altered depending on how big you cluster is. Note that once a pool has been created, changes to this setting will be ignored. Setting this value to -1, enables the number of placement groups to be calculated based on the Ceph placement group calculator.
- (string) Optional repository from which to install. May be one of the following: distro (default), ppa:somecustom/ppa, a deb url sources entry, or a supported Ubuntu Cloud Archive e.g. . cloud:<series>-<openstack-release> cloud:<series>-<openstack-release>/updates cloud:<series>-<openstack-release>/staging cloud:<series>-<openstack-release>/proposed . See https://wiki.ubuntu.com/OpenStack/CloudArchive for info on which cloud archives are available and supported. . Note that a minimum ceph version of 0.48.2 is required for use with this charm which is NOT provided by the packages in the main Ubuntu archive for precise but is provided in the Ubuntu cloud archive.
- (string) SSL CA to use with the certificate and key provided - this is only required if you are providing a privately signed ssl_cert and ssl_key.
- (string) SSL certificate to install and use for API ports. Setting this value and ssl_key will enable reverse proxying, point Glance's entry in the Keystone catalog to use https, and override any certificate and key issued by Keystone (if it is configured to do so).
- (string) SSL key to use with certificate specified as ssl_cert.
- (boolean) If set to True, supporting services will log to syslog.
- (string) Virtual IP(s) to use to front API services in HA configuration. . If multiple networks are being used, a VIP should be provided for each network, separated by spaces.
- (string) Name of RADOS Gateway Zone to create for multi-site replication. This option must be specific to the local site e.g. us-west or us-east.
- (string) Name of RADOS Gateway Zone Group to create for multi-site replication.