Currently we define the letsencrypt certs for each host in its individual host variables. With recent work we have a trusted CA and SAN names setup in our testing environment; introducing the possibility that we could accidentally reference the production host during testing (both have valid certs, as far as the testing hosts are concerned). To avoid this, we can use our naming scheme to move our testing hosts to "99" and avoid collision with the production hosts. As a bonus, this really makes you think more about your group/host split to get things right and keep the environment as abstract as possible. One example of this is that with letsencrypt certificates defined in host vars, testing and production need to use the same hostname to get the right certificates created. Really, this should be group-level information so it applies equally to host01 and host99. To cover "hostXX.opendev.org" as a SAN we can include the inventory_hostname in the group variables. This updates one of the more tricky hosts, static, as a proof of concept. We rename the handlers to be generic, and update the testing targets. Change-Id: Id98768e29a06cebaf645eb75b39e4dc5adb8830d
|6 months ago|
|assets||1 year ago|
|doc||6 months ago|
|docker||7 months ago|
|hiera||8 months ago|
|inventory||6 months ago|
|kubernetes||4 years ago|
|launch||1 year ago|
|manifests||8 months ago|
|modules/openstack_project||8 months ago|
|playbooks||6 months ago|
|roles||9 months ago|
|roles-test||2 years ago|
|testinfra||6 months ago|
|tools||11 months ago|
|zuul.d||6 months ago|
|.gitignore||7 years ago|
|.gitreview||4 years ago|
|COPYING.GPL||4 years ago|
|Gemfile||4 years ago|
|README.rst||11 months ago|
|Rakefile||9 years ago|
|bindep.txt||6 years ago|
|install_modules.sh||3 years ago|
|install_puppet.sh||5 years ago|
|modules.env||8 months ago|
|run_k8s_ansible.sh||4 years ago|
|run_puppet.sh||8 years ago|
|setup.cfg||10 months ago|
|setup.py||5 years ago|
|tox.ini||1 year ago|
OpenDev System Configuration
This is the machinery that drives the configuration, testing, continuous integration and deployment of services provided by the OpenDev project.
Services are driven by Ansible playbooks and associated roles stored here. If you are interested in the configuration of a particular service, starting at
playbooks/service-<name>.yaml will show you how it is configured.
Most services are deployed via containers; many of them are built or customised in this repository; see
A small number of legacy services are still configured with Puppet. Although the act of running puppet on these hosts is managed by Ansible, the actual core of their orchestration lives in
The files in this repository are provided as an opinionated example service deployment, and to allow the OpenDev Collaboratory to use public software development workflows in order to coordinate changes and improvements to the systems it runs. This repository is not intended as a reconsumable project on its own, and anyone wishing to adjust it to suit their own needs should do so with a fork. The system-config reviewers are unable to evaluate and support use cases for the contents here other than their own.
OpenDev infrastructure runs a complete testing and continuous-integration environment, powered by Zuul.
Any changes to playbooks, roles or containers will trigger jobs to thoroughly test those changes.
Tests run the orchestration for the modified services on test nodes assigned to the job. After the testing deployment is configured (validating the basic environment at least starts running), specific tests are configured in the
testinfra directory to validate functionality.
Once changes are reviewed and committed, they will be applied automatically to the production hosts. This is done by Zuul jobs running in the
deploy pipeline. At any one time, you may see these jobs running live on the status page or you could check historical runs on the pipeline results (note there is also an
opendev-prod-hourly pipeline, which ensures things like upstream package updates or certificate renewals are incorporated in a timely fashion).
Contributions are welcome!
You do not need any special permissions to make contributions, even those that will affect production services. Your changes will be automatically tested, reviewed by humans and, once accepted, deployed automatically.
Bug fixes or modifications to existing code are great places to start, and you will see the results of your changes in CI testing. Please remember that this repository consists of configuration and orchestration for OpenDev Collaboratory production systems, so contributions to it will be evaluated on the basis of whether they're useful or applicable to OpenDev's services. Changes intended to make the contents more easily reusable outside OpenDev itself are not in scope, and so will be rejected by reviewers.
You can develop all the playbooks, roles, containers and testing required for a new service just by uploading a change. Using a similar service as a template is generally a good place to start. If deploying to production will require new compute resources (servers, volumes, etc.) these will have to be deployed by an OpenDev administrator before your code is committed. Thus if you know you will need new resources, it is best to coordinate this before review.
The #opendev IRC on OFTC channel is the main place for interactive discussion. Feel free to ask any questions and someone will try to help ASAP. The OpenDev meeting is a co-ordinated time to synchronize on infrastructure issues. Issues should be added to the agenda for discussion; even if you can not attend, you can raise your issue and check back on the logs later. There is also the service-discuss mailing list where you are welcome to send queries or questions.
The latest documentation is available at https://docs.opendev.org/opendev/system-config/latest/
That documentation is generated from this repository. You can geneate it yourself with
tox -e docs.