Compare commits
51 Commits
transcaffe
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
|
2cce72a5fd
|
|||
|
d4a86c667e
|
|||
|
b994778d66
|
|||
|
5dac2b5175
|
|||
|
fb35a0acc6
|
|||
|
afcea9439c
|
|||
|
d337407685
|
|||
|
a9c66729dd
|
|||
|
f0f7c05564
|
|||
|
91c793c3c4
|
|||
|
d0ade80f19
|
|||
|
19e41c15ad
|
|||
|
b82fb86d83
|
|||
|
3c0f9efbb3
|
|||
|
acf1e32eca
|
|||
|
0b11837c72
|
|||
|
07b9a3f3d2
|
|||
|
3e43cca13b
|
|||
|
f765345da3
|
|||
|
1f63757567
|
|||
|
b0796f9812
|
|||
|
044ee49795
|
|||
|
4fc0a671b3
|
|||
|
05ace71da7
|
|||
|
2268d803a7
|
|||
|
4634749378
|
|||
|
6688cf4039
|
|||
|
fb639db97c
|
|||
|
39b7190342
|
|||
|
a3ed7081f8
|
|||
|
b564b1579f
|
|||
|
52f6f5b45b
|
|||
|
d0203a5dd7
|
|||
|
72bb3fd273
|
|||
|
766957254a
|
|||
|
5e777934ed
|
|||
|
b44110704b
|
|||
|
416044692b
|
|||
|
31c68f5e89
|
|||
|
7b9ac4b0d5
|
|||
|
dcab6c5404
|
|||
|
498834eb41
|
|||
|
c827446a9f
|
|||
|
8117dab057
|
|||
|
0aa69e03a9
|
|||
|
06cc920f18
|
|||
|
0ce218e538
|
|||
|
814b5fbf78
|
|||
|
a01ef18ea9
|
|||
|
b692f58fd3
|
|||
|
0b03640e40
|
0
.config/ansible-lint-ignore.txt
Normal file
0
.config/ansible-lint-ignore.txt
Normal file
21
.config/ansible-lint.yml
Normal file
21
.config/ansible-lint.yml
Normal file
@@ -0,0 +1,21 @@
|
||||
---
|
||||
# see https://docs.ansible.com/projects/lint/configuring/#ansible-lint-configuration
|
||||
profile: moderate
|
||||
exclude_paths:
|
||||
- ".ansible/"
|
||||
|
||||
# Enable checking of loop variable prefixes in roles
|
||||
loop_var_prefix: "^(__|{role}_)"
|
||||
|
||||
# Enforce variable names to follow pattern below, in addition to Ansible own
|
||||
# requirements, like avoiding python identifiers. To disable add `var-naming`
|
||||
# to skip_list.
|
||||
var_naming_pattern: "^[a-z_][a-z0-9_]*$"
|
||||
|
||||
skip_list:
|
||||
- "name[template]"
|
||||
enable_list:
|
||||
- galaxy
|
||||
- fqcn
|
||||
|
||||
offline: true
|
||||
2
.gitignore
vendored
Normal file
2
.gitignore
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
.ansible/
|
||||
results/
|
||||
@@ -5,6 +5,8 @@
|
||||
This ansible collection provides various roles for installing
|
||||
and configuring basic system utilities like gnupg, ssh etc
|
||||
|
||||
- [`caddy`](roles/caddy/README.md): configures and runs caddy
|
||||
|
||||
- [`git`](roles/git/README.md): configures git on the target system
|
||||
|
||||
- [`gnupg`](roles/gnupg/README.md): configures gnupg on the target system
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
namespace: finallycoffee
|
||||
name: base
|
||||
version: 0.3.0
|
||||
version: 0.4.1
|
||||
readme: README.md
|
||||
authors:
|
||||
- transcaffeine <transcaffeine@finally.coffee>
|
||||
@@ -8,8 +8,9 @@ description: >-2
|
||||
Roles for base services which are core functionality like managing packages
|
||||
and ssh or common dependencies other services like databases
|
||||
dependencies:
|
||||
"community.docker": "^4.2.0"
|
||||
"community.general": "^10.0.0"
|
||||
"community.docker": ">=4.7.0"
|
||||
"community.general": ">=11.1.2"
|
||||
"containers.podman": ">=1.17.0"
|
||||
license_file: LICENSE.md
|
||||
build_ignore:
|
||||
- '*.tar.gz'
|
||||
@@ -23,6 +24,7 @@ tags:
|
||||
- lego
|
||||
- minio
|
||||
- nginx
|
||||
- caddy
|
||||
- restic
|
||||
- user_management
|
||||
- openldap
|
||||
|
||||
7
playbooks/caddy.yml
Normal file
7
playbooks/caddy.yml
Normal file
@@ -0,0 +1,7 @@
|
||||
---
|
||||
- name: Install and configure caddy
|
||||
hosts: "{{ caddy_hosts | default('caddy') }}"
|
||||
become: "{{ caddy_become | default(false) }}"
|
||||
gather_facts: "{{ caddy_gather_facts | default(false) }}"
|
||||
roles:
|
||||
- role: finallycoffee.base.caddy
|
||||
31
playbooks/caddy_reverse_proxy.yml
Normal file
31
playbooks/caddy_reverse_proxy.yml
Normal file
@@ -0,0 +1,31 @@
|
||||
---
|
||||
- name: Ensure reverse proxy configuration is created
|
||||
hosts: "{{ target_hosts }}"
|
||||
become: "{{ target_become | default(false) }}"
|
||||
gather_facts: "{{ target_gather_facts | default(false) }}"
|
||||
roles:
|
||||
- role: finallycoffee.base.caddy_site
|
||||
vars:
|
||||
caddy_site_cert_basepath: >-2
|
||||
{{ caddy_site_tls_store | default('/tls') }}/{{ caddy_site_name }}/certificates/{{ caddy_site_name }}
|
||||
caddy_site_config: >-2
|
||||
{{ caddy_site_config_override | default(caddy_site_default_config, true) }}
|
||||
caddy_site_default_config: |+2
|
||||
https://{{ caddy_site_name }} {
|
||||
tls {{ caddy_site_cert_basepath}}.crt {{ caddy_site_cert_basepath }}.key
|
||||
header {
|
||||
Strict-Transport-Security "max-age=31536000"
|
||||
}
|
||||
encode zstd gzip
|
||||
{% if caddy_reverse_proxy_template_block | default(true) -%}
|
||||
reverse_proxy {{ caddy_reverse_proxy_backend_addr }} {
|
||||
{{ caddy_reverse_proxy_extra_config | default('') | indent(6) }}
|
||||
{%- if caddy_reverse_proxy_import_proxyheaders | default(true, true) %}
|
||||
import proxyheaders
|
||||
{%- endif +%}
|
||||
}
|
||||
{%- else -%}
|
||||
{{- caddy_reverse_proxy_block | default('') | indent(4) }}
|
||||
{%- endif +%}
|
||||
}
|
||||
|
||||
@@ -13,4 +13,4 @@
|
||||
loop: "{{ docker_registries | default([], true) }}"
|
||||
loop_control:
|
||||
loop_var: "docker_registry"
|
||||
label: "{{ docker_registry.username}}@{{ docker_registry.registry }}"
|
||||
label: "{{ docker_registry.username }}@{{ docker_registry.registry }}"
|
||||
|
||||
54
playbooks/docker_shorewall.yml
Normal file
54
playbooks/docker_shorewall.yml
Normal file
@@ -0,0 +1,54 @@
|
||||
---
|
||||
- name: Configure shorewall for docker egress
|
||||
hosts: "{{ docker_shorewall_hosts | default('docker:&shorewall') }}"
|
||||
become: "{{ docker_shorewall_become | default(true, true) }}"
|
||||
tasks:
|
||||
- name: Add docker interface
|
||||
ansible.builtin.lineinfile:
|
||||
path: /etc/shorewall/interfaces
|
||||
regex: "^dock"
|
||||
line: |
|
||||
dock docker0 bridge
|
||||
- name: Add docker routing policy
|
||||
ansible.builtin.blockinfile:
|
||||
path: /etc/shorewall/policy
|
||||
insertbefore: "^# THE FOLLOWING POLICY MUST BE LAST"
|
||||
content: |
|
||||
# Docker specific configuration
|
||||
dock all ACCEPT
|
||||
- name: Add docker zone
|
||||
ansible.builtin.lineinfile:
|
||||
path: /etc/shorewall/zones
|
||||
regex: "^dock"
|
||||
line: |
|
||||
dock ipv4
|
||||
- name: Add docker egress rules
|
||||
ansible.builtin.blockinfile:
|
||||
path: /etc/shorewall/rules
|
||||
marker: "#{mark} ANSIBLE MANAGED BLOCK - DOCKER EGRESS"
|
||||
content: |
|
||||
#
|
||||
# Docker egress configuration
|
||||
#
|
||||
ACCEPT dock all
|
||||
- name: Add docker dns rules
|
||||
ansible.builtin.blockinfile:
|
||||
path: /etc/shorewall/rules
|
||||
marker: "#{mark} ANSIBLE MANAGED BLOCK - DOCKER DNS"
|
||||
content: |
|
||||
#
|
||||
# Docker dns configuration
|
||||
#
|
||||
DNS(ACCEPT) dock all
|
||||
- name: Enable shorewall docker support
|
||||
ansible.builtin.lineinfile:
|
||||
path: /etc/shorewall/shorewall.conf
|
||||
line: "DOCKER=Yes"
|
||||
regex: "^DOCKER="
|
||||
- name: Ensure shorewall reloaded
|
||||
community.general.systemd_service:
|
||||
service: "{{ item }}"
|
||||
state: reloaded
|
||||
loop:
|
||||
- shorewall.service
|
||||
- shorewall6.service
|
||||
109
playbooks/ldap.yml
Normal file
109
playbooks/ldap.yml
Normal file
@@ -0,0 +1,109 @@
|
||||
---
|
||||
- name: Configure LDAP directory information tree
|
||||
hosts: "{{ ldap_hosts | default('ldap') }}"
|
||||
become: "{{ ldap_become | default(false) }}"
|
||||
gather_facts: "{{ ldap_gather_facts | default(false) }}"
|
||||
vars:
|
||||
_state: "{{ ldap_state | default('present') }}"
|
||||
_ldap_bind_info: &ldap_bind_info
|
||||
server_uri: "{{ ldap_server_uri }}"
|
||||
bind_dn: "{{ ldap_bind_dn }}"
|
||||
bind_pw: "{{ ldap_bind_pw }}"
|
||||
roles:
|
||||
# Ensure all defaults from openldap role are in scope
|
||||
- role: finallycoffee.base.openldap
|
||||
when: false
|
||||
tasks:
|
||||
- name: Ensure org units in '{{ ldap_base_dn }}' are {{ _state }}
|
||||
community.general.ldap_entry:
|
||||
<<: *ldap_bind_info
|
||||
dn: "ou={{ org_unit }},{{ ldap_base_dn }}"
|
||||
objectClass: "organizationalUnit"
|
||||
state: "{{ _state }}"
|
||||
loop: "{{ ldap_org_units | default([], true) }}"
|
||||
loop_control:
|
||||
loop_var: org_unit
|
||||
|
||||
- name: Ensure admin user is {{ _state }}
|
||||
community.general.ldap_entry:
|
||||
<<: *ldap_bind_info
|
||||
dn: "uid={{ ldap_admin_user_rdn }},{{ ldap_admin_user_base }}"
|
||||
objectClass: "{{ ldap_admin_user_object_classes }}"
|
||||
attributes: "{{ ldap_admin_user_attributes }}"
|
||||
state: "{{ _state }}"
|
||||
vars:
|
||||
ldap_admin_user_base: >-2
|
||||
{{ ldap_admin_user_base_dn | default(ldap_base_dn, true) }}
|
||||
when: ldap_admin_user_rdn is defined
|
||||
|
||||
- name: Ensure admin user attributes are correct
|
||||
community.general.ldap_attrs:
|
||||
<<: *ldap_bind_info
|
||||
dn: "uid={{ ldap_admin_user_rdn }},{{ ldap_admin_user_base }}"
|
||||
attributes: "{{ ldap_admin_user_attributes }}"
|
||||
state: "{{ _state }}"
|
||||
vars:
|
||||
ldap_admin_user_base: >-2
|
||||
{{ ldap_admin_user_base_dn | default(ldap_base_dn, true) }}
|
||||
when:
|
||||
- ldap_admin_user_rdn is defined
|
||||
- _state == 'present'
|
||||
|
||||
- name: Ensure ldap groups are {{ _state }}
|
||||
community.general.ldap_entry:
|
||||
<<: *ldap_bind_info
|
||||
dn: "{{ _ldap_group_dn }}"
|
||||
objectClass: "{{ _ldap_group_object_classes }}"
|
||||
attributes: "{{ _ldap_group_attributes }}"
|
||||
state: "{{ _state }}"
|
||||
vars:
|
||||
_ldap_group_dn: >-2
|
||||
cn={{ _ldap_group.name }},{{ ldap_group_base_dn }}
|
||||
_ldap_group_object_classes:
|
||||
- "groupOfNames"
|
||||
_ldap_group_attributes:
|
||||
cn: "{{ _ldap_group.name }}"
|
||||
member: >-2
|
||||
{{ _ldap_group.members | default([]) }}
|
||||
loop: "{{ ldap_groups | default([], true) }}"
|
||||
loop_control:
|
||||
loop_var: _ldap_group
|
||||
label: "{{ _ldap_group.name }}"
|
||||
when:
|
||||
- ldap_groups is defined
|
||||
- ldap_group_base_dn is defined
|
||||
|
||||
- name: Ensure service accounts are {{ _state }}
|
||||
community.general.ldap_entry:
|
||||
<<: *ldap_bind_info
|
||||
dn: "{{ _ldap_service_account_dn }}"
|
||||
objectClass: "{{ _ldap_service_account_object_classes }}"
|
||||
attributes: "{{ _ldap_service_account_attributes }}"
|
||||
state: "{{ _state }}"
|
||||
loop: &ldap_service_account_loop "{{ ldap_service_accounts | default([]) }}"
|
||||
loop_control: &ldap_service_account_loop_control
|
||||
loop_var: "_ldap_service_account"
|
||||
label: "{{ _ldap_service_account.name }}"
|
||||
vars: &ldap_service_account_vars
|
||||
_ldap_service_account_dn: >-2
|
||||
uid={{ _ldap_service_account.name }},{{ ldap_service_account_base_dn }}
|
||||
_ldap_service_account_object_classes:
|
||||
- "account"
|
||||
- "simpleSecurityObject"
|
||||
_ldap_service_account_attributes:
|
||||
uid: "{{ _ldap_service_account.name }}"
|
||||
userPassword: "{{ _ldap_service_account.password }}"
|
||||
when: &ldap_service_account_when
|
||||
- ldap_service_accounts is defined
|
||||
- ldap_service_account_base_dn is defined
|
||||
|
||||
- name: Ensure service accounts attributes are correct
|
||||
community.general.ldap_attrs:
|
||||
<<: *ldap_bind_info
|
||||
dn: "{{ _ldap_service_account_dn }}"
|
||||
attributes: "{{ _ldap_service_account_attributes }}"
|
||||
state: exact
|
||||
loop: *ldap_service_account_loop
|
||||
loop_control: *ldap_service_account_loop_control
|
||||
vars: *ldap_service_account_vars
|
||||
when: *ldap_service_account_when
|
||||
85
playbooks/lego_certificate.yml
Normal file
85
playbooks/lego_certificate.yml
Normal file
@@ -0,0 +1,85 @@
|
||||
---
|
||||
- name: Populate DNS, acquire TSIG key and obtain certificate
|
||||
hosts: "{{ target_hosts | default('all') }}"
|
||||
become: "{{ target_become | default(true) }}"
|
||||
gather_facts: "{{ target_gather_facts | default(false) }}"
|
||||
pre_tasks:
|
||||
- name: Build target dns records
|
||||
ansible.builtin.set_fact:
|
||||
target_dns_records: "{{ target_dns_records + [_dns_record] }}"
|
||||
vars:
|
||||
_dns_record:
|
||||
type: "CNAME"
|
||||
name: "_acme-challenge.{{ _domain }}."
|
||||
content: "{{ target_tsig_key_name }}.{{ target_acme_zone }}."
|
||||
loop: "{{ target_domains }}"
|
||||
loop_control:
|
||||
loop_var: "_domain"
|
||||
- name: Populate dns_server if not given
|
||||
ansible.builtin.set_fact:
|
||||
dns_server: "{{ target_dns_server }}"
|
||||
when: dns_server is not defined
|
||||
roles:
|
||||
- role: finallycoffee.base.dns
|
||||
vars:
|
||||
dns_records: "{{ target_dns_records + target_dns_additional_records }}"
|
||||
dns_tsig_name: "{{ target_dns_tsig_key.name }}"
|
||||
dns_tsig_algo: "{{ target_dns_tsig_key.algorithm }}"
|
||||
dns_tsig_key: "{{target_dns_tsig_key.key }}"
|
||||
delegate_to: localhost
|
||||
- role: finallycoffee.base.powerdns_tsig_key
|
||||
vars:
|
||||
powerdns_tsig_key_algo: "{{ target_powerdns_tsig_key_algo }}"
|
||||
powerdns_tsig_key_name: "{{ target_tsig_key_name }}"
|
||||
powerdns_tsig_key_path: "{{ target_tsig_key_path }}"
|
||||
powerdns_tsig_key_path_owner: "{{ target_acme_user }}"
|
||||
powerdns_tsig_key_path_group: "{{ target_acme_group }}"
|
||||
- role: finallycoffee.base.lego
|
||||
vars:
|
||||
lego_instance: "{{ target_lego_instance }}"
|
||||
lego_instance_base_path: "{{ target_lego_instance_base_path }}"
|
||||
lego_environment: "{{ target_lego_environment }}"
|
||||
lego_cert_domains: "{{ target_lego_domains }}"
|
||||
lego_acme_account_email: "{{ target_acme_account_email }}"
|
||||
lego_acme_challenge_type: "{{ target_lego_acme_challenge_type }}"
|
||||
lego_acme_challenge_provider: "{{ target_lego_acme_challenge_provider }}"
|
||||
lego_acme_server_url: "{{ target_lego_acme_server_url }}"
|
||||
vars:
|
||||
target_domains: []
|
||||
target_acme_zone: ~
|
||||
target_acme_account_email: ~
|
||||
target_dns_server: ~
|
||||
target_dns_additional_records: []
|
||||
target_dns_tsig_key: {}
|
||||
target_lego_instance: "{{ target_domains | first }}"
|
||||
target_lego_instance_base_path: "/opt/acme"
|
||||
target_lego_domains: "{{ target_domains }}"
|
||||
target_lego_acme_challenge_type: "dns"
|
||||
target_lego_acme_challenge_provider: "rfc2136"
|
||||
target_lego_acme_server_url: >-2
|
||||
{{ lego_letsencrypt_server_urls.prod }}
|
||||
target_lego_environment:
|
||||
RFC2136_TSIG_KEY: "{{ target_tsig_key_name }}"
|
||||
RFC2136_TSIG_SECRET_FILE: "{{ target_tsig_key_path }}"
|
||||
RFC2136_TSIG_ALGORITHM: "{{ target_powerdns_tsig_key_algo }}"
|
||||
RFC2136_NAMESERVER: "{{ target_dns_server }}"
|
||||
RFC2136_DNS_TIMEOUT: 15
|
||||
RFC2136_TTL: 60
|
||||
RFC2136_SEQUENCE_INTERVAL: 5
|
||||
RFC2136_POLLING_INTERVAL: 10
|
||||
RFC2136_PROPAGATION_TIMEOUT: >-2
|
||||
{{ (target_lego_domains | length * 120) | int }}
|
||||
LEGO_EXPERIMENTAL_CNAME_SUPPORT: "true"
|
||||
target_tsig_key_name: "{{ target_lego_instance | hash('sha1') }}"
|
||||
target_tsig_key_path: >-2
|
||||
{{ target_lego_instance_base_path }}/{{ target_lego_instance }}/secrets/rfc2136_tsig.key
|
||||
target_tsig_key_path_owner:
|
||||
target_tsig_key_path_group:
|
||||
target_acme_user: "acme-{{ target_lego_instance }}"
|
||||
target_acme_user_id: >-2
|
||||
{{ powerdns_tsig_key_path_owner_info.uid }}
|
||||
target_acme_group: "acme-{{ target_lego_instance }}"
|
||||
target_acme_group_id: >-2
|
||||
{{ powerdns_tsig_key_path_owner_info.gid }}
|
||||
target_powerdns_tsig_key_algo: "hmac-sha256"
|
||||
target_dns_records: []
|
||||
7
playbooks/ntp.yml
Normal file
7
playbooks/ntp.yml
Normal file
@@ -0,0 +1,7 @@
|
||||
---
|
||||
- name: Install and configure network time protocol daemon
|
||||
hosts: "{{ ntp_hosts | default('ntp') }}"
|
||||
become: "{{ ntp_become | default(false) }}"
|
||||
gather_facts: "{{ ntp_gather_facts | default(false) }}"
|
||||
roles:
|
||||
- role: finallycoffee.base.ntp
|
||||
7
playbooks/wg_quick.yml
Normal file
7
playbooks/wg_quick.yml
Normal file
@@ -0,0 +1,7 @@
|
||||
---
|
||||
- name: Configure wireguard interfaces with wg_quick
|
||||
hosts: "{{ wg_quick_hosts | default('wg_quick') }}"
|
||||
become: "{{ wg_quick_become | default(false) }}"
|
||||
gather_facts: "{{ wg_quick_gather_facts | default(false) }}"
|
||||
roles:
|
||||
- role: finallycoffee.base.wg_quick
|
||||
10
roles/caddy/README.md
Normal file
10
roles/caddy/README.md
Normal file
@@ -0,0 +1,10 @@
|
||||
# `finallycoffee.base.caddy` ansible role
|
||||
|
||||
Deploy a (pre-)configure [caddy v2](https://caddyserver.com) web
|
||||
server / proxy using ansible.
|
||||
|
||||
## Configuration
|
||||
|
||||
To change the default configuration of reading all files from
|
||||
`/etc/caddy/sites.d/` (see `caddy_dynamic_config_dir`), specify
|
||||
your desired configuration in `caddy_config`.
|
||||
22
roles/caddy/defaults/main/config.yml
Normal file
22
roles/caddy/defaults/main/config.yml
Normal file
@@ -0,0 +1,22 @@
|
||||
---
|
||||
caddy_config: |+2
|
||||
{
|
||||
auto_https disable_redirects
|
||||
}
|
||||
|
||||
(proxyheaders) {
|
||||
header_up X-Forwarded-Ssl on
|
||||
header_up Host {host}
|
||||
header_up X-Real-IP {remote}
|
||||
header_up X-Forwarded-For {remote}
|
||||
# header_up X-Forwarded-Port {port}
|
||||
header_up X-Forwarded-Proto {scheme}
|
||||
header_up X-Url-Scheme {scheme}
|
||||
header_up X-Forwarded-Host {host}
|
||||
}
|
||||
|
||||
# Import all configurations
|
||||
import {{ caddy_dynamic_configs_dir }}/*/Caddyfile
|
||||
:80 {
|
||||
redir / https://{host}{uri} 301
|
||||
}
|
||||
43
roles/caddy/defaults/main/container.yml
Normal file
43
roles/caddy/defaults/main/container.yml
Normal file
@@ -0,0 +1,43 @@
|
||||
---
|
||||
caddy_container_image_registry: "docker.io"
|
||||
caddy_container_image_namespace: "library"
|
||||
caddy_container_image_repository: "caddy"
|
||||
caddy_container_image_name: >-2
|
||||
{{ [
|
||||
caddy_container_image_registry | default([], true),
|
||||
caddy_container_image_namespace | default([], true),
|
||||
caddy_container_image_repository
|
||||
] | flatten | join('/') }}
|
||||
|
||||
caddy_container_image_tag: ~
|
||||
caddy_container_image: >-2
|
||||
{{ [
|
||||
caddy_container_image_name,
|
||||
caddy_container_image_tag | default(caddy_version, true)
|
||||
] | join(':') }}
|
||||
caddy_container_image_source: "pull"
|
||||
caddy_container_image_force_source: >-2
|
||||
{{ caddy_container_image_tag | ansible.builtin.type_debug != 'NoneType' }}
|
||||
caddy_container_image_state: "{{ caddy_state }}"
|
||||
|
||||
caddy_container_name: "caddy"
|
||||
caddy_container_env: ~
|
||||
caddy_container_ports: ~
|
||||
caddy_container_user: ~
|
||||
caddy_container_labels: ~
|
||||
caddy_container_volumes: ~
|
||||
caddy_container_config_dir: "/etc/caddy"
|
||||
caddy_container_default_volumes:
|
||||
- "{{ caddy_config_dir }}:{{ caddy_container_config_dir }}:ro"
|
||||
- "{{ caddy_dynamic_configs_dir }}:{{ caddy_dynamic_configs_dir }}:ro"
|
||||
- "{{ caddy_config_internal_dir }}:/config:rw"
|
||||
- "{{ caddy_state_dir }}:/data:rw"
|
||||
caddy_container_all_volumes: >-2
|
||||
{{ caddy_container_default_volumes | default([], true)
|
||||
+ caddy_container_volumes | default([], true) }}
|
||||
caddy_container_state: >-2
|
||||
{{ (caddy_state == 'present') | ternary('started', 'absent') }}
|
||||
caddy_container_restart_policy: "on-failure"
|
||||
caddy_container_networks: ~
|
||||
caddy_container_network_mode: ~
|
||||
caddy_container_etc_hosts: ~
|
||||
11
roles/caddy/defaults/main/main.yml
Normal file
11
roles/caddy/defaults/main/main.yml
Normal file
@@ -0,0 +1,11 @@
|
||||
---
|
||||
caddy_user: "caddy"
|
||||
caddy_version: "2.10.2"
|
||||
caddy_config_file: "/etc/caddy/Caddyfile"
|
||||
caddy_config_dir: "{{ caddy_config_file | ansible.builtin.dirname }}"
|
||||
caddy_config_internal_dir: "{{ caddy_config_dir }}/config"
|
||||
caddy_dynamic_configs_dir: "{{ caddy_config_dir }}/sites.d"
|
||||
caddy_state_dir: "/var/lib/caddy"
|
||||
|
||||
caddy_state: "present"
|
||||
caddy_deployment_method: "docker"
|
||||
7
roles/caddy/defaults/main/user.yml
Normal file
7
roles/caddy/defaults/main/user.yml
Normal file
@@ -0,0 +1,7 @@
|
||||
---
|
||||
caddy_user_state: "{{ caddy_state }}"
|
||||
caddy_user_system: true
|
||||
caddy_user_create_home: false
|
||||
|
||||
caddy_run_uid: "{{ caddy_user_info.uid | default(caddy_user) }}"
|
||||
caddy_run_gid: "{{ caddy_user_info.group | default(caddy_user) }}"
|
||||
13
roles/caddy/meta/main.yml
Normal file
13
roles/caddy/meta/main.yml
Normal file
@@ -0,0 +1,13 @@
|
||||
---
|
||||
allow_duplicates: true
|
||||
dependencies: []
|
||||
galaxy_info:
|
||||
role_name: caddy
|
||||
description: Deploy caddy, a webserver
|
||||
galaxy_tags:
|
||||
- caddy
|
||||
- zerossl
|
||||
- http
|
||||
- webserver
|
||||
- docker
|
||||
- podman
|
||||
26
roles/caddy/tasks/deploy-docker.yml
Normal file
26
roles/caddy/tasks/deploy-docker.yml
Normal file
@@ -0,0 +1,26 @@
|
||||
---
|
||||
- name: Ensure container image '{{ caddy_container_image }}' is {{ caddy_container_image_state }}
|
||||
community.docker.docker_image:
|
||||
name: "{{ caddy_container_image }}"
|
||||
state: "{{ caddy_container_image_state }}"
|
||||
source: "{{ caddy_container_image_source }}"
|
||||
force_source: "{{ caddy_container_image_force_source }}"
|
||||
register: caddy_container_image_info
|
||||
until: caddy_container_image_info is success
|
||||
retries: 10
|
||||
delay: 3
|
||||
|
||||
- name: Ensure container '{{ caddy_container_name }}' is {{ caddy_container_state }}
|
||||
community.docker.docker_container:
|
||||
name: "{{ caddy_container_name }}"
|
||||
image: "{{ caddy_container_image }}"
|
||||
state: "{{ caddy_container_state }}"
|
||||
env: "{{ caddy_container_env | default(omit, true) }}"
|
||||
user: "{{ caddy_container_user | default(omit, true) }}"
|
||||
ports: "{{ caddy_container_ports | default(omit, true) }}"
|
||||
labels: "{{ caddy_container_labels | default(omit, true) }}"
|
||||
volumes: "{{ caddy_container_all_volumes }}"
|
||||
networks: "{{ caddy_container_networks | default(omit, true) }}"
|
||||
etc_hosts: "{{ caddy_container_etc_hosts | default(omit, true) }}"
|
||||
network_mode: "{{ caddy_container_network_mode | default(omit, true) }}"
|
||||
restart_policy: "{{ caddy_container_restart_policy }}"
|
||||
52
roles/caddy/tasks/main.yml
Normal file
52
roles/caddy/tasks/main.yml
Normal file
@@ -0,0 +1,52 @@
|
||||
---
|
||||
- name: Ensure state '{{ caddy_state }}' is valid
|
||||
ansible.builtin.fail:
|
||||
msg: >-2
|
||||
Unsupported caddy_state '{{ caddy_state }}'.
|
||||
Supported states are {{ caddy_states | join(', ') }}.
|
||||
when: caddy_state not in caddy_states
|
||||
|
||||
- name: Ensure deployment method '{{ caddy_deployment_method }}' is valid
|
||||
ansible.builtin.fail:
|
||||
msg: >-2
|
||||
Unsupported caddy_deployment_method '{{ caddy_deployment_method }}'.
|
||||
Supported deployment methods are {{ caddy_deployment_methods | join(', ') }}.
|
||||
when: caddy_deployment_method not in caddy_deployment_methods
|
||||
|
||||
- name: Ensure caddy user '{{ caddy_user }}' is {{ caddy_user_state }}
|
||||
ansible.builtin.user:
|
||||
name: "{{ caddy_user }}"
|
||||
state: "{{ caddy_user_state }}"
|
||||
system: "{{ caddy_user_system }}"
|
||||
create_home: "{{ caddy_user_create_home }}"
|
||||
register: "caddy_user_info"
|
||||
|
||||
- name: Ensure base directories are present
|
||||
ansible.builtin.file:
|
||||
path: "{{ dir.name }}"
|
||||
state: "directory"
|
||||
owner: "{{ dir.owner | default(caddy_run_uid) }}"
|
||||
group: "{{ dir.group | default(caddy_run_uid) }}"
|
||||
mode: "{{ dir.mode | default('0750') }}"
|
||||
when: caddy_state == 'present'
|
||||
loop:
|
||||
- name: "{{ caddy_config_dir }}"
|
||||
- name: "{{ caddy_dynamic_configs_dir }}"
|
||||
- name: "{{ caddy_config_internal_dir }}"
|
||||
- name: "{{ caddy_state_dir }}"
|
||||
loop_control:
|
||||
loop_var: "dir"
|
||||
label: "{{ dir.name }}"
|
||||
|
||||
- name: Ensure caddy configuration is up to date
|
||||
ansible.builtin.copy:
|
||||
dest: "{{ caddy_config_file }}"
|
||||
content: "{{ caddy_config }}"
|
||||
owner: "{{ caddy_run_uid }}"
|
||||
group: "{{ caddy_run_gid }}"
|
||||
mode: "0640"
|
||||
when: caddy_state == 'present'
|
||||
|
||||
- name: Ensure caddy is deployed using {{ caddy_deployment_method }}
|
||||
ansible.builtin.include_tasks:
|
||||
file: "deploy-{{ caddy_deployment_method }}.yml"
|
||||
6
roles/caddy/vars/main.yml
Normal file
6
roles/caddy/vars/main.yml
Normal file
@@ -0,0 +1,6 @@
|
||||
---
|
||||
caddy_states:
|
||||
- "present"
|
||||
- "absent"
|
||||
caddy_deployment_methods:
|
||||
- "docker"
|
||||
7
roles/caddy_site/README.md
Normal file
7
roles/caddy_site/README.md
Normal file
@@ -0,0 +1,7 @@
|
||||
# `finallycoffee.base.caddy_site` ansible role
|
||||
|
||||
Provision a single site configuration in caddy.
|
||||
|
||||
Set `caddy_site_name` as a unique
|
||||
site identifier (needs to be a valid filename) and `caddy_site_config`
|
||||
to contain the actual `Caddyfile` contents.
|
||||
13
roles/caddy_site/defaults/main.yml
Normal file
13
roles/caddy_site/defaults/main.yml
Normal file
@@ -0,0 +1,13 @@
|
||||
---
|
||||
caddy_site_name: ~
|
||||
caddy_site_config: ~
|
||||
caddy_site_state: "present"
|
||||
|
||||
caddy_site_configs: "/etc/caddy/sites.d"
|
||||
caddy_site_config_dir: >-2
|
||||
{{ caddy_site_configs }}/{{ caddy_site_name }}
|
||||
caddy_site_config_file: >-2
|
||||
{{ caddy_site_config_dir }}/Caddyfile
|
||||
|
||||
caddy_site_owner: "caddy"
|
||||
caddy_site_group: "caddy"
|
||||
11
roles/caddy_site/meta/main.yml
Normal file
11
roles/caddy_site/meta/main.yml
Normal file
@@ -0,0 +1,11 @@
|
||||
---
|
||||
allow_duplicates: true
|
||||
dependencies: []
|
||||
galaxy_info:
|
||||
role_name: caddy_site
|
||||
description: Deploy a sites' configuration in caddy
|
||||
galaxy_tags:
|
||||
- caddy
|
||||
- zerossl
|
||||
- http
|
||||
- webserver
|
||||
26
roles/caddy_site/tasks/main.yml
Normal file
26
roles/caddy_site/tasks/main.yml
Normal file
@@ -0,0 +1,26 @@
|
||||
---
|
||||
- name: Fail if required variables are not populated
|
||||
ansible.builtin.fail:
|
||||
msg: "Either `caddy_site_name` or `caddy_site_config` is not provided"
|
||||
when: >-2
|
||||
(caddy_site_name | ansible.builtin.type_debug == 'NoneType')
|
||||
or
|
||||
(caddy_site_config | ansible.builtin.type_debug == 'NoneType')
|
||||
|
||||
- name: Ensure directory for caddy site config '{{ caddy_site_name }}' is {{ caddy_site_state }}
|
||||
ansible.builtin.file:
|
||||
path: "{{ caddy_site_config_dir }}"
|
||||
state: >-2
|
||||
{{ (caddy_site_state == 'present') | ternary('directory', 'absent') }}
|
||||
owner: "{{ caddy_site_owner }}"
|
||||
group: "{{ caddy_site_group }}"
|
||||
mode: "0750"
|
||||
|
||||
- name: Ensure caddy site configuration is templated
|
||||
ansible.builtin.copy:
|
||||
dest: "{{ caddy_site_config_file }}"
|
||||
content: "{{ caddy_site_config }}"
|
||||
owner: "{{ caddy_site_owner }}"
|
||||
group: "{{ caddy_site_group }}"
|
||||
mode: "0640"
|
||||
when: caddy_site_state == 'present'
|
||||
@@ -8,7 +8,7 @@
|
||||
state: present
|
||||
marker: "#{mark} ANSIBLE MANAGED BLOCK by finallycoffee.base.git"
|
||||
block: |+2
|
||||
{% if git_config_user_name|default(false, true) and git_config_user_email|default(false, true) %}
|
||||
{% if git_config_user_name | default(false, true) and git_config_user_email | default(false, true) %}
|
||||
[user]
|
||||
name = {{ git_config_user_name }}
|
||||
email = {{ git_config_user_email }}
|
||||
|
||||
@@ -14,7 +14,7 @@ gpg_config_emit_version: false
|
||||
gpg_config_comments: false
|
||||
gpg_config_ignore_time_conflict: false
|
||||
gpg_config_allow_freeform_uid: true
|
||||
gpg_config_keyid_format: 0xlong
|
||||
gpg_config_keyid_format: "0xlong"
|
||||
gpg_config_with_fingerprint: true
|
||||
|
||||
gpg_config_keyserver: hkps://keys.openpgp.org
|
||||
|
||||
@@ -1,55 +1,54 @@
|
||||
---
|
||||
|
||||
- name: Ensure gnupg is installed (RedHat*)
|
||||
package:
|
||||
ansible.builtin.package:
|
||||
name: gnupg2
|
||||
state: latest
|
||||
become: true
|
||||
when: ansible_os_family == "RedHat"
|
||||
|
||||
- name: Ensure gnupg is installed (Arch)
|
||||
package:
|
||||
ansible.builtin.package:
|
||||
name: gnupg
|
||||
state: latest
|
||||
become: true
|
||||
when: ansible_os_family == "Archlinux"
|
||||
|
||||
- name: Ensure ~/.gnupg folder exists with correct permissions
|
||||
file:
|
||||
ansible.builtin.file:
|
||||
path: "{{ gpg_config_folder }}"
|
||||
state: directory
|
||||
mode: 0700
|
||||
mode: "0700"
|
||||
|
||||
- name: Ensure gpg.conf is templated
|
||||
template:
|
||||
ansible.builtin.template:
|
||||
src: gpg.conf.j2
|
||||
dest: "{{ gpg_config_file }}"
|
||||
|
||||
- name: Configure gpg-agent.conf (agent configuration)
|
||||
template:
|
||||
ansible.builtin.template:
|
||||
src: gpg-agent.conf.j2
|
||||
dest: "{{ gpg_agent_config_file }}"
|
||||
|
||||
- name: Configure scdaemon.conf (smartcard daemon)
|
||||
template:
|
||||
ansible.builtin.template:
|
||||
src: scdaemon.conf.j2
|
||||
dest: "{{ gpg_scdaemon_config_file }}"
|
||||
|
||||
- name: Configure sshcontrol (in order for gpg-agent to act as ssh-agent)
|
||||
template:
|
||||
ansible.builtin.template:
|
||||
src: sshcontrol.j2
|
||||
dest: "{{ gpg_agent_sshcontrol_file }}"
|
||||
when: gpg_agent_config_enable_ssh_support
|
||||
|
||||
- name: Copy gnupg_agent script, which makes gpg-agent responsible for ssh-auth
|
||||
copy:
|
||||
ansible.builtin.copy:
|
||||
src: gpg-configure-ssh-auth-socket.sh
|
||||
dest: "{{ gpg_configure_agent_script }}"
|
||||
mode: 0700
|
||||
mode: "0700"
|
||||
when: gpg_agent_config_enable_ssh_support
|
||||
|
||||
- name: Ensure gnupg_agent script is included in bashrc
|
||||
lineinfile:
|
||||
ansible.builtin.lineinfile:
|
||||
path: "~/.bashrc"
|
||||
line: "source {{ gpg_configure_agent_script }}"
|
||||
state: present
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
---
|
||||
lego_user: "lego"
|
||||
lego_version: "4.23.0"
|
||||
lego_version: "4.31.0"
|
||||
lego_instance: default
|
||||
lego_base_path: "/opt/lego"
|
||||
lego_cert_user: "acme-{{ lego_instance }}"
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
#!/usr/bin/env bash
|
||||
set -euo pipefail
|
||||
set -xeuo pipefail
|
||||
|
||||
LEGO_BINARY=$(/usr/bin/env which lego)
|
||||
|
||||
@@ -8,8 +8,11 @@ if [[ -n "${LEGO_HTTP_FALLBACK_PORT:-}" ]]; then
|
||||
echo "nc not found (in PATH), exiting"
|
||||
exit 1
|
||||
fi
|
||||
set +e
|
||||
nc -z 127.0.0.1 $LEGO_HTTP_PORT;
|
||||
if [[ $? -eq 0 ]]; then
|
||||
nc_exit_code=$?;
|
||||
set -e
|
||||
if [[ $nc_exit_code -eq 0 ]]; then
|
||||
LEGO_HTTP_PORT=$LEGO_HTTP_FALLBACK_PORT
|
||||
fi
|
||||
fi
|
||||
|
||||
@@ -1,11 +1,11 @@
|
||||
---
|
||||
|
||||
lego_domain_command_args: >-
|
||||
lego_domain_command_args: >-2
|
||||
{% for domain in lego_cert_domains %}
|
||||
--domains={{ domain }}
|
||||
{%- endfor %}
|
||||
|
||||
lego_config_command_args: >-
|
||||
lego_config_command_args: >-2
|
||||
{% for key in lego_full_command_config %}
|
||||
--{{ key | replace("_", "-") }}
|
||||
{%- if lego_full_command_config[key] != None and lego_full_command_config[key] != '' -%}
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
---
|
||||
minio_container_name: minio
|
||||
minio_container_image_name: "docker.io/minio/minio"
|
||||
minio_container_image_tag: "RELEASE.2025-04-22T22-12-26Z"
|
||||
minio_container_image_tag: "RELEASE.2025-10-15T17-29-55Z"
|
||||
minio_container_image: "{{ minio_container_image_name }}:{{ minio_container_image_tag }}"
|
||||
minio_container_networks: []
|
||||
minio_container_ports: []
|
||||
@@ -23,7 +23,7 @@ minio_container_command:
|
||||
- "--console-address"
|
||||
- ":{{ minio_container_listen_port_console }}"
|
||||
minio_container_restart_policy: "unless-stopped"
|
||||
minio_container_image_force_source: "{{ (minio_container_image_tag == 'latest')|bool }}"
|
||||
minio_container_image_force_source: "{{ (minio_container_image_tag == 'latest') | bool }}"
|
||||
minio_container_state: >-2
|
||||
{{ (minio_state == 'present') | ternary('started', 'absent') }}
|
||||
|
||||
|
||||
@@ -3,15 +3,15 @@
|
||||
ansible.builtin.file:
|
||||
path: "{{ minio_data_path }}"
|
||||
state: directory
|
||||
user: "{{ minio_user|default(omit, True) }}"
|
||||
group: "{{ minio_user|default(omit, True) }}"
|
||||
user: "{{ minio_user | default(omit, True) }}"
|
||||
group: "{{ minio_user | default(omit, True) }}"
|
||||
when: minio_manage_host_filesystem
|
||||
|
||||
- name: Ensure container image '{{ minio_container_image }}' is {{ minio_state }}
|
||||
community.docker.docker_image:
|
||||
name: "{{ minio_container_image }}"
|
||||
state: "{{ minio_state }}"
|
||||
source: pull
|
||||
source: "pull"
|
||||
force_source: "{{ minio_container_image_force_source }}"
|
||||
|
||||
- name: Ensure container '{{ minio_container_name }}' is {{ minio_container_state }}
|
||||
@@ -23,7 +23,7 @@
|
||||
labels: "{{ minio_container_labels }}"
|
||||
networks: "{{ minio_container_networks }}"
|
||||
ports: "{{ minio_container_ports }}"
|
||||
user: "{{ minio_user|default(omit, True) }}"
|
||||
user: "{{ minio_user | default(omit, True) }}"
|
||||
command: "{{ minio_container_command }}"
|
||||
restart_policy: "{{ minio_container_restart_policy }}"
|
||||
state: "{{ minio_container_state }}"
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
---
|
||||
nginx_version: "1.28.0"
|
||||
nginx_version: "1.29.5"
|
||||
nginx_flavour: alpine
|
||||
nginx_base_path: /opt/nginx
|
||||
nginx_config_file: "{{ nginx_base_path }}/nginx.conf"
|
||||
@@ -34,4 +34,3 @@ nginx_container_state: >-2
|
||||
nginx_container_restart_policy: "unless-stopped"
|
||||
nginx_container_volumes:
|
||||
- "{{ nginx_config_file }}:/etc/nginx/conf.d/nginx.conf:ro"
|
||||
|
||||
|
||||
@@ -30,7 +30,7 @@
|
||||
ansible.builtin.copy:
|
||||
dest: "{{ nginx_config_file }}"
|
||||
content: "{{ nginx_config }}"
|
||||
mode: 0640
|
||||
mode: "0640"
|
||||
notify:
|
||||
- restart-nginx
|
||||
when: nginx_state == 'present'
|
||||
|
||||
4
roles/ntp/README.md
Normal file
4
roles/ntp/README.md
Normal file
@@ -0,0 +1,4 @@
|
||||
# `finallycoffee.base.ntp`
|
||||
|
||||
Install `ntp`, configure a timezone by sym-linking `/etc/localtime`
|
||||
and enable the systemd service.
|
||||
14
roles/ntp/defaults/main.yml
Normal file
14
roles/ntp/defaults/main.yml
Normal file
@@ -0,0 +1,14 @@
|
||||
---
|
||||
ntp_state: present
|
||||
ntp_package_name: "ntp"
|
||||
|
||||
ntp_timezone: "Europe/Paris"
|
||||
|
||||
ntp_systemd_service_name: "ntpd.service"
|
||||
ntp_systemd_service_state: >-2
|
||||
{{ (ntp_state == 'present') | ternary('started', 'stopped') }}
|
||||
ntp_systemd_service_enabled: >-2
|
||||
{{ (ntp_state == 'present') }}
|
||||
|
||||
ntp_etc_localtime_file: "/etc/localtime"
|
||||
ntp_usr_share_zoneinfo_path: "/usr/share/zoneinfo"
|
||||
8
roles/ntp/meta/main.yml
Normal file
8
roles/ntp/meta/main.yml
Normal file
@@ -0,0 +1,8 @@
|
||||
---
|
||||
allow_duplicates: true
|
||||
dependencies: []
|
||||
galaxy_info:
|
||||
role_name: ntp
|
||||
description: Install network time protocol daemon
|
||||
galaxy_tags:
|
||||
- ntp
|
||||
28
roles/ntp/tasks/main.yml
Normal file
28
roles/ntp/tasks/main.yml
Normal file
@@ -0,0 +1,28 @@
|
||||
---
|
||||
- name: Check if 'ntp_state' is valid
|
||||
ansible.builtin.fail:
|
||||
msg: >-2
|
||||
Invalid state '{{ ntp_state }}'! Valid
|
||||
states are {{ ntp_states | join(', ') }}.
|
||||
when: ntp_state not in ntp_states
|
||||
|
||||
- name: Ensure system package is {{ ntp_state }}
|
||||
ansible.builtin.package:
|
||||
name: "{{ ntp_package_name }}"
|
||||
state: "{{ ntp_state }}"
|
||||
|
||||
- name: Ensure /etc/localtime is symlinked
|
||||
ansible.builtin.file:
|
||||
src: "{{ ntp_usr_share_zoneinfo_path }}/{{ ntp_timezone }}"
|
||||
dest: "{{ ntp_etc_localtime_file }}"
|
||||
state: "{{ (ntp_state == 'present') | ternary('link', 'absent') }}"
|
||||
|
||||
- name: Ensure ntp systemd service is configured
|
||||
ansible.builtin.systemd:
|
||||
name: "{{ ntp_systemd_service_name }}"
|
||||
enabled: "{{ ntp_systemd_service_enabled }}"
|
||||
|
||||
- name: Ensure ntp systemd service is {{ ntp_systemd_service_state }}
|
||||
ansible.builtin.systemd:
|
||||
name: "{{ ntp_systemd_service_name }}"
|
||||
state: "{{ ntp_systemd_service_state }}"
|
||||
4
roles/ntp/vars/main.yml
Normal file
4
roles/ntp/vars/main.yml
Normal file
@@ -0,0 +1,4 @@
|
||||
---
|
||||
ntp_states:
|
||||
- "present"
|
||||
- "absent"
|
||||
@@ -26,11 +26,13 @@ openldap_additional_schemas: []
|
||||
openldap_schemas: >-2
|
||||
{{ openldap_enabled_schemas + openldap_additional_schemas }}
|
||||
|
||||
openldap_config_db: "cn=config"
|
||||
openldap_config_db_olc_access: >-2
|
||||
to *
|
||||
by dn.base="gidNumber=0+uidNumber=0,cn=peercred,cn=external,cn=auth" manage
|
||||
by * none
|
||||
openldap_config_dn: "cn=config"
|
||||
openldap_config_db_dn: "olcDatabase={0}config,cn=config"
|
||||
openldap_config_db_olc_access:
|
||||
- '{0} to *
|
||||
by dn.base="gidNumber=0+uidNumber=0,cn=peercred,cn=external,cn=auth" manage
|
||||
by * none'
|
||||
openldap_config_attributes: {}
|
||||
openldap_config_db_attributes:
|
||||
olcAccess: "{{ openldap_config_db_olc_access }}"
|
||||
|
||||
@@ -49,7 +51,7 @@ openldap_default_database_directory: >-2
|
||||
openldap_default_database_indices: >-2
|
||||
{{ openldap_default_indices + openldap_indices }}
|
||||
openldap_default_database_config: >-2
|
||||
olcDatabase={1}{{ openldap_default_database_name }},{{ openldap_config_db }}
|
||||
olcDatabase={1}{{ openldap_default_database_name }},{{ openldap_config_dn }}
|
||||
openldap_default_database:
|
||||
name: "{{ openldap_default_database_name }}"
|
||||
object_class: "{{ openldap_default_database_object_class }}"
|
||||
@@ -58,5 +60,6 @@ openldap_default_database:
|
||||
root_pw: "{{ openldap_default_database_root_pw }}"
|
||||
directory: "{{ openldap_default_database_directory }}"
|
||||
indices: "{{ openldap_default_database_indices }}"
|
||||
openldap_default_database_olc_access: "{{ openldap_config_db_olc_access }}"
|
||||
openldap_databases:
|
||||
- "{{ openldap_default_database }}"
|
||||
|
||||
@@ -1,9 +1,31 @@
|
||||
---
|
||||
- name: Ensure ACLs are configured
|
||||
- name: Ensure config attributes are configured
|
||||
community.general.ldap_attrs:
|
||||
dn: "{{ openldap_config_dn }}"
|
||||
attributes: "{{ {entry.key: entry.value} }}"
|
||||
state: exact
|
||||
server_uri: "{{ openldap_socket_url }}"
|
||||
loop: "{{ openldap_config_attributes | dict2items }}"
|
||||
loop_control:
|
||||
loop_var: "entry"
|
||||
label: "{{ entry.key }}"
|
||||
|
||||
- name: Ensure config db attributes are configured
|
||||
community.general.ldap_attrs:
|
||||
dn: "{{ openldap_config_db_dn }}"
|
||||
attributes: "{{ {entry.key: entry.value} }}"
|
||||
state: exact
|
||||
server_uri: "{{ openldap_socket_url }}"
|
||||
loop: "{{ openldap_config_db_attributes | dict2items }}"
|
||||
loop_control:
|
||||
loop_var: "entry"
|
||||
label: "{{ entry.key }}"
|
||||
|
||||
- name: Ensure ACLs for default database are configured
|
||||
community.general.ldap_attrs:
|
||||
dn: "{{ openldap_default_database_config }}"
|
||||
attributes:
|
||||
olcAccess: "{{ openldap_config_db_olc_access }}"
|
||||
olcAccess: "{{ openldap_default_database_olc_access }}"
|
||||
state: "exact"
|
||||
server_uri: "{{ openldap_socket_url }}"
|
||||
retries: 3
|
||||
|
||||
@@ -22,4 +22,3 @@
|
||||
{{ openldap_container_restart_policy | default(omit, true) }}
|
||||
healthcheck: "{{ openldap_container_healthcheck | default(omit, true) }}"
|
||||
state: "{{ openldap_container_state }}"
|
||||
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
- name: Ensure additional schemas are mapped to container
|
||||
ansible.builtin.set_fact:
|
||||
openldap_init_container_volumes: >-2
|
||||
{{ openldap_init_container_volumes + [ schema_mount ] }}
|
||||
{{ openldap_init_container_volumes + [schema_mount] }}
|
||||
vars:
|
||||
schema_file: "{{ openldap_schema_path }}/{{ schema.name }}.ldif"
|
||||
schema_mount: >-2
|
||||
|
||||
@@ -41,7 +41,7 @@
|
||||
community.docker.docker_container_exec:
|
||||
container: "{{ powerdns_tsig_key_container_name }}"
|
||||
command: "pdnsutil list-tsig-keys"
|
||||
delegate_to: "{{ powerdns_tsig_key_hostname }}"
|
||||
delegate_to: "{{ powerdns_tsig_key_hostname | default(inventory_hostname) }}"
|
||||
register: powerdns_tsig_key_powerdns_info
|
||||
changed_when: false
|
||||
check_mode: false
|
||||
@@ -54,7 +54,7 @@
|
||||
when: >-
|
||||
(powerdns_tsig_key_name ~ '. ' ~ powerdns_tsig_key_algo ~ '. ')
|
||||
not in powerdns_tsig_key_powerdns_info.stdout
|
||||
delegate_to: "{{ powerdns_tsig_key_hostname }}"
|
||||
delegate_to: "{{ powerdns_tsig_key_hostname | default(inventory_hostname) }}"
|
||||
register: powerdns_tsig_key_powerdns_generated_tsig_key
|
||||
throttle: 1
|
||||
become: true
|
||||
@@ -67,7 +67,7 @@
|
||||
when: >-
|
||||
(powerdns_tsig_key_name ~ '. ' ~ powerdns_tsig_key_algo ~ '. ')
|
||||
not in powerdns_tsig_key_powerdns_info.stdout
|
||||
delegate_to: "{{ powerdns_tsig_key_hostname }}"
|
||||
delegate_to: "{{ powerdns_tsig_key_hostname | default(inventory_hostname) }}"
|
||||
throttle: 1
|
||||
become: true
|
||||
|
||||
|
||||
@@ -18,10 +18,10 @@
|
||||
msg: "`restic_backup_stdin_command` was set but no filename for the resulting output was supplied in `restic_backup_stdin_command_filename`"
|
||||
|
||||
- name: Ensure backup frequency adheres to systemd's OnCalender syntax
|
||||
command:
|
||||
ansible.builtin.command:
|
||||
cmd: "systemd-analyze calendar {{ restic_policy.frequency }}"
|
||||
register: systemd_calender_parse_res
|
||||
failed_when: systemd_calender_parse_res.rc != 0
|
||||
register: restic_systemd_calender_parse_result
|
||||
failed_when: restic_systemd_calender_parse_result.rc != 0
|
||||
changed_when: false
|
||||
|
||||
- name: Ensure restic is installed
|
||||
@@ -42,13 +42,21 @@
|
||||
state: present
|
||||
when: ansible_os_family not in ['RedHat', 'Debian']
|
||||
|
||||
- name: Ensure backup script is templated
|
||||
ansible.builtin.copy:
|
||||
src: restic-backup-directories.sh
|
||||
dest: /opt/restic-backup-directories.sh
|
||||
owner: root
|
||||
group: root
|
||||
mode: "0755"
|
||||
|
||||
- name: Ensure systemd service file for '{{ restic_job_name }}' is templated
|
||||
template:
|
||||
dest: "/etc/systemd/system/{{ restic_systemd_unit_naming_scheme }}.service"
|
||||
src: restic.service.j2
|
||||
owner: root
|
||||
group: root
|
||||
mode: 0640
|
||||
mode: "0640"
|
||||
notify:
|
||||
- reload-systemd
|
||||
- trigger-restic
|
||||
@@ -59,7 +67,7 @@
|
||||
src: restic.timer.j2
|
||||
owner: root
|
||||
group: root
|
||||
mode: 0640
|
||||
mode: "0640"
|
||||
notify:
|
||||
- reload-systemd
|
||||
|
||||
|
||||
@@ -4,6 +4,9 @@ Description=Run {{ restic_timer_description | default(restic_job_name) }}
|
||||
[Timer]
|
||||
OnCalendar={{ restic_policy.frequency }}
|
||||
Unit={{ restic_systemd_unit_naming_scheme }}.service
|
||||
{% if restic_systemd_timer_randomized_delay_sec %}
|
||||
RandomizedDelaySec={{ restic_systemd_timer_randomized_delay_sec }}
|
||||
{% endif %}
|
||||
|
||||
[Install]
|
||||
WantedBy=timers.target
|
||||
|
||||
5
roles/wg_quick/README.md
Normal file
5
roles/wg_quick/README.md
Normal file
@@ -0,0 +1,5 @@
|
||||
# `finallycoffee.base.wg_quick` ansible role
|
||||
|
||||
Configure a wireguard interface using `wg_quick`. This role writes
|
||||
the configuration files and activates the interface using the systemd
|
||||
template service abstractions.
|
||||
20
roles/wg_quick/defaults/main/connection.yml
Normal file
20
roles/wg_quick/defaults/main/connection.yml
Normal file
@@ -0,0 +1,20 @@
|
||||
---
|
||||
wg_quick_interface_name: ~
|
||||
wg_quick_interface_address: ~
|
||||
wg_quick_interface_listen_port: ~
|
||||
wg_quick_interface_private_key: ~
|
||||
wg_quick_interface_private_key_file: ~
|
||||
wg_quick_interface_peer_endpoint: ~
|
||||
wg_quick_interface_peer_public_key: ~
|
||||
wg_quick_interface_peer_allowed_ips: ~
|
||||
|
||||
wg_quick_interfaces:
|
||||
- name: "{{ wg_quck_interface_name }}"
|
||||
address: "{{ wg_quick_interface_address }}"
|
||||
listen_port: "{{ wg_quick_interface_listen_port }}"
|
||||
private_key: "{{ wg_quick_interface_private_key }}"
|
||||
private_key_file: "{{ wg_quick_interface_private_key_file }}"
|
||||
peers:
|
||||
- endpoint: "{{ wg_quick_interface_peer_endpoint }}"
|
||||
public_key: "{{ wg_quick_interface_peer_public_key }}"
|
||||
allowed_ips: "{{ wg_quick_interface_peer_allowed_ips }}"
|
||||
7
roles/wg_quick/defaults/main/main.yml
Normal file
7
roles/wg_quick/defaults/main/main.yml
Normal file
@@ -0,0 +1,7 @@
|
||||
---
|
||||
wg_quick_state: "present"
|
||||
wg_quick_package_name: "wireguard-tools"
|
||||
wg_quick_system_packages:
|
||||
- "{{ wg_quick_package_name }}"
|
||||
|
||||
wg_quick_configuration_dir: "/etc/wireguard"
|
||||
25
roles/wg_quick/tasks/configure-interface.yml
Normal file
25
roles/wg_quick/tasks/configure-interface.yml
Normal file
@@ -0,0 +1,25 @@
|
||||
---
|
||||
- name: Ensure wg-quick configuration for interface '{{ wg_quick_iface.name }}' is up to date
|
||||
ansible.builtin.template:
|
||||
src: "wg-quick.conf.j2"
|
||||
dest: "{{ wg_quick_configuration_dir }}/{{ wg_quick_iface.name }}.conf"
|
||||
when: wg_quick_iface.state | default(wg_quick_state) == 'present'
|
||||
|
||||
- name: Ensure systemd service is enabled
|
||||
ansible.builtin.systemd_service:
|
||||
name: "wg-quick@{{ wg_quick_iface.name }}.service"
|
||||
enabled: true
|
||||
when: wg_quick_iface.state | default(wg_quick_state) == 'present'
|
||||
|
||||
- name: Ensure systemd service is {{ wg_quick_iface.state | default(wg_quick_state) }}
|
||||
ansible.builtin.systemd_service:
|
||||
name: "wg-quick@{{ wg_quick_iface.name }}.service"
|
||||
state: >-2
|
||||
{{ (wg_quick_iface.state | default(wg_quick_state) == 'present')
|
||||
| ternary('started', 'absent') }}
|
||||
|
||||
- name: Ensure wg-quick configuration for interface '{{ wg_quick_iface.name }}' is absent
|
||||
ansible.builtin.file:
|
||||
path: "{{ wg_quick_configuration_dir }}/{{ wg_quick_face.name }}.conf"
|
||||
state: "absent"
|
||||
when: wg_quick_iface.state | default(wg_quick_state) == 'absent'
|
||||
27
roles/wg_quick/tasks/main.yml
Normal file
27
roles/wg_quick/tasks/main.yml
Normal file
@@ -0,0 +1,27 @@
|
||||
---
|
||||
- name: Ensure wg_quick_state is valid
|
||||
ansible.builtin.fail:
|
||||
msg: >-2
|
||||
Invalid state '{{ wg_quick_state }}'. Valid
|
||||
states are {{ wg_quick_states | join(', ') }}.
|
||||
when: wg_quick_state not in wg_quick_states
|
||||
|
||||
- name: Ensure system packages are available
|
||||
ansible.builtin.package:
|
||||
name: "{{ wg_quick_system_packages }}"
|
||||
state: "present"
|
||||
when: wg_quick_state == 'present'
|
||||
|
||||
- name: Ensure configuration folder is present
|
||||
ansible.builtin.file:
|
||||
name: "{{ wg_quick_configuration_dir }}"
|
||||
state: "directory"
|
||||
when: wg_quick_state == 'present'
|
||||
|
||||
- name: Ensure connections are in the configured state
|
||||
ansible.builtin.include_tasks:
|
||||
file: "configure-interface.yml"
|
||||
loop: "{{ wg_quick_interfaces }}"
|
||||
loop_control:
|
||||
loop_var: "wg_quick_iface"
|
||||
label: "{{ wg_quick_iface.name }}"
|
||||
32
roles/wg_quick/templates/wg-quick.conf.j2
Normal file
32
roles/wg_quick/templates/wg-quick.conf.j2
Normal file
@@ -0,0 +1,32 @@
|
||||
[Interface]
|
||||
Address = {{ wg_quick_iface.address | join(', ') }}
|
||||
ListenPort = {{ wg_quick_iface.listen_port }}
|
||||
|
||||
{% if wg_quick_iface.private_key %}
|
||||
PrivateKey = {{ wg_quick_iface.private_key }}
|
||||
{% elif wg_quick_iface.private_key_file %}
|
||||
PrivateKeyFile = {{ wg_quick_iface.private_key_file }}
|
||||
{% endif %}
|
||||
{% if wg_quick_iface.table is defined %}
|
||||
Table = {{ wg_quick_iface.table | ternary('on', 'off') }}
|
||||
{% endif %}
|
||||
{% if wg_quick_iface.post_up %}
|
||||
PostUp = /bin/bash -c "{{ wg_quick_iface.post_up | join('; ') }}"
|
||||
{% endif %}
|
||||
{% if wg_quick_iface.pre_down %}
|
||||
PreDown = /bin/bash -c "{{ wg_quick_iface.pre_down | join('; ') }}"
|
||||
{% endif %}
|
||||
|
||||
|
||||
{% for _peer in wg_quick_iface.peers %}
|
||||
[Peer]
|
||||
Endpoint = {{ _peer.endpoint }}
|
||||
PublicKey = {{ _peer.public_key }}
|
||||
AllowedIPs = {{ _peer.allowed_ips | join(', ') }}
|
||||
{% if _peer.persistent_keepalive %}
|
||||
PersistentKeepalive = {{ _peer.persistent_keepalive }}
|
||||
{% endif %}
|
||||
{% if 'psk' in _peer %}
|
||||
PresharedKey = {{ _peer.psk }}
|
||||
{% endif %}
|
||||
{% endfor %}
|
||||
4
roles/wg_quick/vars/main.yml
Normal file
4
roles/wg_quick/vars/main.yml
Normal file
@@ -0,0 +1,4 @@
|
||||
---
|
||||
wg_quick_states:
|
||||
- "present"
|
||||
- "absent"
|
||||
Reference in New Issue
Block a user