This document explains the processes and practices recommended for contributing enhancements to this operator.
- Generally, before developing enhancements to this charm, you should consider opening an issue explaining your use case.
- If you would like to chat with us about your use-cases or proposed implementation, you can reach us at Canonical Mattermost public channel or Discourse.
- Familiarising yourself with the Charmed Operator Framework library will help you a lot when working on new features or bug fixes.
- All enhancements require review before being merged. Code review typically examines
- code quality
- test coverage
- user experience for Juju administrators of this charm.
- Please help us out in ensuring easy to review branches by rebasing your pull request branch onto
the
main
branch. This also avoids merge commits and creates a linear Git commit history.
Build the charm in this git repository using tox.
There are two alternatives to build the charm: using the charm cache or not. Cache will speed the build by downloading all dependencies from charmcraftcache-hub.
First, ensure you have the right dependencies:
- charmcraft v2.5.4+
- charmcraftcache
By running the following commands:
pipx install charmcraftcache
tox -e build-dev
To run the traditional build only using charmcraft
, run the following command:
charmcraftcache pack
You can create an environment for development with tox
:
tox devenv -e integration
source venv/bin/activate
To run tests, first build the charm as described above, then run the following
tox -e format # update your code according to linting rules
tox -e lint # code style
tox -e unit # unit tests
tox -e integration # integration tests, running on juju 3.
tox # runs 'format', 'lint', and 'unit' environments
Integration tests can be run for separate files:
tox -e integration -- tests/integration/tls/test_tls.py
tox -e integration -- tests/integration/relations/test_charm.py
tox -e integration -- tests/integration/plugins/test_plugins.py
tox -e integration -- tests/integration/ha/test_storage.py
tox -e integration -- tests/integration/ha/test_large_deployments.py
tox -e integration -- tests/integration/ha/test_horizontal_scaling.py
tox -e integration -- tests/integration/ha/test_ha_networking.py
tox -e integration -- tests/integration/ha/test_ha_multi_clusters.py
tox -e integration -- tests/integration/relations/test_opensearch_provider.py
tox -e integration -- tests/integration/ha/test_ha.py
tox -e integration -- tests/integration/ha/test_backups.py
For integration tests, libjuju must be in-sync with the target juju version. Make sure that the version of libjuju installed is compatible with the bootstrapped controller version. If not, update it with:
poetry add --lock --group integration juju@<YOUR CHOSEN VERSION>
Backup testing installs microceph and can run on S3 (aws) object stores. To setup your environment, you should set the: access / secret / service account information as environment variables.
To run the test only against microceph:
tox -e integration -- tests/integration/ha/test_backups.py --group='microceph' # test backup service for microceph
And against public clouds + microceph:
SECRETS_FROM_GITHUB=$(cat <path-to>/credentials.json) tox -e integration -- tests/integration/ha/test_backups.py
Where, for AWS only, credentials.json
should look like:
$ cat credentials.json
{ "AWS_ACCESS_KEY": ..., "AWS_SECRET_KEY": ...}
OpenSearch has a set of system requirements to correctly function, you can find the list here. Some of those settings must be set using cloudinit-userdata on the model, while others must be set on the host machine:
cat <<EOF > cloudinit-userdata.yaml
cloudinit-userdata: |
postruncmd:
- [ 'echo', 'vm.max_map_count=262144', '>>', '/etc/sysctl.conf' ]
- [ 'echo', 'vm.swappiness=0', '>>', '/etc/sysctl.conf' ]
- [ 'echo', 'net.ipv4.tcp_retries2=5', '>>', '/etc/sysctl.conf' ]
- [ 'echo', 'fs.file-max=1048576', '>>', '/etc/sysctl.conf' ]
- [ 'sysctl', '-p' ]
EOF
echo "vm.max_map_count=262144" | sudo tee -a /etc/sysctl.conf
echo "vm.swappiness=0" | sudo tee -a /etc/sysctl.conf
sudo sysctl -p
Then create a new model and set the previously generated file in it.
# Create a model
juju add-model dev
# Enable DEBUG logging
juju model-config logging-config="<root>=INFO;unit=DEBUG"
# Add cloudinit-userdata
juju model-config --file=./cloudinit-userdata.yaml
# Increase the frequency of the update-status event
juju model-config update-status-hook-interval=1m
You can then deploy the charm with a TLS relation.
# Deploy the self-signed-certificates operator
juju deploy self-signed-certificates --channel=latest/stable --show-log --verbose
# generate a CA certificate
juju config \
self-signed-certificates \
ca-common-name="CN_CA" \
certificate-validity=365 \
root-ca-validity=365
# Deploy the opensearch charm
juju deploy -n 1 ./opensearch_ubuntu-22.04-amd64.charm --series jammy --show-log --verbose
# Relate the opensearch charm with the self-signed-certificates operator
juju integrate self-signed-certificates opensearch
Note: The TLS settings shown here are for self-signed-certificates, which are not recommended for production clusters. The TLS Certificates Operator offers a variety of configurations. Read more on the self-signed-certificates Operator here.
Canonical welcomes contributions to the Charmed Template Operator. Please check out our contributor agreement if you're interested in contributing to the solution.