Ceph is a distributed storage and network file system designed to provide excellent performance, reliability, and scalability. . This charm provides the RADOS HTTP gateway supporting S3 and Swift protocols for object storage.
Ceph is a distributed storage and network file system designed to provide excellent performance, reliability and scalability.
This charm deploys the RADOS Gateway, a S3 and Swift compatible HTTP gateway for online object storage on-top of a ceph cluster.
In order to use this charm, it is assumed that you have already deployed a ceph storage cluster using the 'ceph' charm with something like this::
juju deploy -n 3 --config ceph.yaml ceph
To deploy the RADOS gateway simple do::
juju deploy ceph-radosgw juju add-relation ceph-radosgw ceph
You can then directly access the RADOS gateway by exposing the service::
juju expose ceph-radosgw
The gateway can be accessed over port 80 (as show in juju status exposed ports).
Note that you will need to login to one of the service units supporting the ceph charm to generate some access credentials::
juju ssh ceph/0 \ 'sudo radosgw-admin user create --uid="ubuntu" --display-name="Ubuntu Ceph"'
For security reasons the ceph-radosgw charm is not set up with appropriate permissions to administer the ceph cluster.
Ceph >= 0.55 integrates with Openstack Keystone for authentication of Swift requests.
This is enabled by relating the ceph-radosgw service with keystone::
juju deploy keystone juju add-relation keystone ceph-radosgw
If you try to relate the radosgw to keystone with an earlier version of ceph the hook will error out to let you know.
Its possible to scale-out the RADOS Gateway itself::
juju add-unit -n 2 ceph-radosgw
and then stick a HA loadbalancer on the front::
juju deploy haproxy juju add-relation haproxy ceph-radosgw
Should give you a bit more bang on the front end if you really need it.
Author: James Page firstname.lastname@example.org Report bugs at: http://bugs.launchpad.net/charms/+source/ceph-radosgw/+filebug Location: http://jujucharms.com/charms/ceph-radosgw
The Ceph RADOS Gateway makes use of a multiverse package libapache2-mod-fastcgi. As such it will try to automatically enable the multiverse pocket in /etc/apt/sources.list. Note that there is noting 'wrong' with multiverse components - they typically have less liberal licensing policies or suchlike.
- (int) Number of keystone tokens to hold in local cache.
- (int) This value dictates the number of replicas ceph must make of any object it stores within RGW pools. Note that once the RGW pools have been created, changing this value will not have any effect (although it can be changed in ceph by manually configuring your ceph cluster).
- (string) Default network interface on which HA cluster will bind to communication with the other members of the HA Cluster.
- (int) Default multicast port number that will be used to communicate between HA Cluster nodes.
- (int) Client timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 30000ms is used.
- (int) Connect timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 5000ms is used.
- (int) Queue timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 5000ms is used.
- (int) Server timeout configuration in ms for haproxy, used in HA configurations. If not provided, default value of 30000ms is used.
- (string) Key ID to import to the apt keyring to support use with arbitary source configuration from outside of Launchpad archives or PPA's.
- (int) RadosGW debug level. Max is 20.
- (string) Comma-separated list of Swift operator roles; used when integrating with OpenStack Keystone.
- (string) The hostname or address of the public endpoints created for ceph-radosgw in the keystone identity provider. This value will be used for public endpoints. For example, an os-public-hostname set to 'files.example.com' with will create the following public endpoint for the ceph-radosgw: https://files.example.com:80/swift/v1
- (int) The port that the RADOS Gateway will listen on.
- (string) OpenStack region that the RADOS gateway supports; used when integrating with OpenStack Keystone.
- (int) Interval between revocation checks to keystone.
- (int) When the Rados Gatway is installed it, by default, creates pools with pg_num 8 which, in the majority of cases is suboptimal. A few rgw pools tend to carry more data than others e.g. .rgw.buckets tends to be larger than most. So, for pools with greater requirements than others the charm will apply the optimal value i.e. corresponding to the number of OSDs up+in the cluster at the time the pool is created. For others it will use this value which can be altered depending on how big you cluster is. Note that once a pool has been created, changes to this setting will be ignored.
- (string) Optional configuration to support use of additional sources such as: - ppa:myteam/ppa - cloud:trusty-proposed/kilo - http://my.archive.com/ubuntu main The last option should be used in conjunction with the key configuration option. Note that a minimum ceph version of 0.48.2 is required for use with this charm which is NOT provided by the packages in the main Ubuntu archive for precise but is provided in the Ubuntu cloud archive.
- (boolean) By default apache2 and libapache2-mod-fastcgi will be installed from the Ubuntu archives. This option allows for an alternate ceph.com install source which contains patched versions with added support for HTTP 100-continue. See the following page for more info: http://ceph.com/docs/dumpling/radosgw/manual-install/#continue-support
- (boolean) Newer versions of the Ceph RADOS Gateway support use of an embedded web container instead of Apache + mod-fastcgi, avoiding some of the nuances of using the stock mod-fastcgi packages from Ubuntu. Enable this option to disable use of Apache and enable the embedded web container feature.
- (boolean) If set to True, supporting services will log to syslog.
- (string) Virtual IP(s) to use to front API services in HA configuration. If multiple networks are being used, a VIP should be provided for each network, separated by spaces.