mirror of
https://github.com/kevinveenbirkenbach/computer-playbook.git
synced 2025-09-08 11:17:17 +02:00
Compare commits
5 Commits
072ad6f186
...
33276263b0
Author | SHA1 | Date | |
---|---|---|---|
33276263b0 | |||
168c5c0da6 | |||
aa61bf2a44 | |||
25cee9a4c7 | |||
6780950257 |
@@ -3,10 +3,16 @@ import os
|
||||
import argparse
|
||||
import yaml
|
||||
import json
|
||||
from collections import deque
|
||||
import re
|
||||
from typing import List, Dict, Any, Set
|
||||
|
||||
|
||||
JINJA_PATTERN = re.compile(r'{{.*}}')
|
||||
ALL_DEP_TYPES = ['run_after', 'dependencies', 'include_tasks', 'import_tasks', 'include_role', 'import_role']
|
||||
ALL_DIRECTIONS = ['to', 'from']
|
||||
ALL_KEYS = [f"{dep}_{dir}" for dep in ALL_DEP_TYPES for dir in ALL_DIRECTIONS]
|
||||
|
||||
|
||||
def find_role_meta(roles_dir: str, role: str) -> str:
|
||||
path = os.path.join(roles_dir, role, 'meta', 'main.yml')
|
||||
if not os.path.isfile(path):
|
||||
@@ -14,10 +20,14 @@ def find_role_meta(roles_dir: str, role: str) -> str:
|
||||
return path
|
||||
|
||||
|
||||
def find_role_tasks(roles_dir: str, role: str) -> str:
|
||||
path = os.path.join(roles_dir, role, 'tasks', 'main.yml')
|
||||
if not os.path.isfile(path):
|
||||
raise FileNotFoundError(f"Tasks not found for role: {role}")
|
||||
return path
|
||||
|
||||
|
||||
def load_meta(path: str) -> Dict[str, Any]:
|
||||
"""
|
||||
Load meta/main.yml → return galaxy_info + run_after + dependencies
|
||||
"""
|
||||
with open(path, 'r') as f:
|
||||
data = yaml.safe_load(f) or {}
|
||||
|
||||
@@ -28,6 +38,24 @@ def load_meta(path: str) -> Dict[str, Any]:
|
||||
'dependencies': data.get('dependencies', []) or []
|
||||
}
|
||||
|
||||
|
||||
def load_tasks(path: str, dep_type: str) -> List[str]:
|
||||
with open(path, 'r') as f:
|
||||
data = yaml.safe_load(f) or []
|
||||
|
||||
included_roles = []
|
||||
|
||||
for task in data:
|
||||
if dep_type in task:
|
||||
entry = task[dep_type]
|
||||
if isinstance(entry, dict):
|
||||
entry = entry.get('name', '')
|
||||
if entry and not JINJA_PATTERN.search(entry):
|
||||
included_roles.append(entry)
|
||||
|
||||
return included_roles
|
||||
|
||||
|
||||
def build_single_graph(
|
||||
start_role: str,
|
||||
dep_type: str,
|
||||
@@ -35,71 +63,73 @@ def build_single_graph(
|
||||
roles_dir: str,
|
||||
max_depth: int
|
||||
) -> Dict[str, Any]:
|
||||
"""
|
||||
Build one graph for one dependency type and direction:
|
||||
- 'to': follow edges source→target
|
||||
- 'from': reverse edges (find roles listing this role)
|
||||
- max_depth > 0: limit hops to max_depth
|
||||
- max_depth ≤ 0: stop when you’d revisit a node already on the path
|
||||
"""
|
||||
nodes: Dict[str, Dict[str, Any]] = {}
|
||||
links: List[Dict[str, str]] = []
|
||||
|
||||
def traverse(role: str, depth: int, path: Set[str]):
|
||||
# Register node once
|
||||
if role not in nodes:
|
||||
meta = load_meta(find_role_meta(roles_dir, role))
|
||||
node = {'id': role}
|
||||
node.update(meta['galaxy_info'])
|
||||
node['doc_url'] = f"https://docs.cymais.cloud/roles/{role}/README.html"
|
||||
node['source_url'] = (
|
||||
f"https://github.com/kevinveenbirkenbach/cymais/tree/master/roles/{role}"
|
||||
)
|
||||
node['source_url'] = f"https://github.com/kevinveenbirkenbach/cymais/tree/master/roles/{role}"
|
||||
nodes[role] = node
|
||||
|
||||
# Depth guard
|
||||
if max_depth > 0 and depth >= max_depth:
|
||||
return
|
||||
|
||||
# Determine neighbors according to direction
|
||||
neighbors = []
|
||||
if dep_type in ['run_after', 'dependencies']:
|
||||
meta = load_meta(find_role_meta(roles_dir, role))
|
||||
neighbors = meta.get(dep_type, [])
|
||||
else:
|
||||
try:
|
||||
neighbors = load_tasks(find_role_tasks(roles_dir, role), dep_type)
|
||||
except FileNotFoundError:
|
||||
neighbors = []
|
||||
|
||||
if direction == 'to':
|
||||
neighbors = load_meta(find_role_meta(roles_dir, role)).get(dep_type, [])
|
||||
for tgt in neighbors:
|
||||
links.append({'source': role, 'target': tgt, 'type': dep_type})
|
||||
# General cycle check
|
||||
if tgt in path:
|
||||
continue
|
||||
traverse(tgt, depth + 1, path | {tgt})
|
||||
|
||||
else: # direction == 'from'
|
||||
# Find all roles that list this role in their dep_type
|
||||
for other in os.listdir(roles_dir):
|
||||
try:
|
||||
meta_o = load_meta(find_role_meta(roles_dir, other))
|
||||
other_neighbors = []
|
||||
if dep_type in ['run_after', 'dependencies']:
|
||||
meta_o = load_meta(find_role_meta(roles_dir, other))
|
||||
other_neighbors = meta_o.get(dep_type, [])
|
||||
else:
|
||||
other_neighbors = load_tasks(find_role_tasks(roles_dir, other), dep_type)
|
||||
|
||||
if role in other_neighbors:
|
||||
links.append({'source': other, 'target': role, 'type': dep_type})
|
||||
if other in path:
|
||||
continue
|
||||
traverse(other, depth + 1, path | {other})
|
||||
|
||||
except FileNotFoundError:
|
||||
continue
|
||||
if role in meta_o.get(dep_type, []):
|
||||
links.append({'source': other, 'target': role, 'type': dep_type})
|
||||
if other in path:
|
||||
continue
|
||||
traverse(other, depth + 1, path | {other})
|
||||
|
||||
# Kick off recursion
|
||||
traverse(start_role, depth=0, path={start_role})
|
||||
return {'nodes': list(nodes.values()), 'links': links}
|
||||
|
||||
|
||||
def build_mappings(
|
||||
start_role: str,
|
||||
mappings: List[Dict[str, str]],
|
||||
roles_dir: str,
|
||||
max_depth: int
|
||||
) -> Dict[str, Any]:
|
||||
result: Dict[str, Any] = {}
|
||||
for mapping in mappings:
|
||||
for dep_type, direction in mapping.items():
|
||||
key = f"{dep_type}_{direction}"
|
||||
result[key] = build_single_graph(
|
||||
start_role, dep_type, direction, roles_dir, max_depth)
|
||||
for key in ALL_KEYS:
|
||||
dep_type, direction = key.rsplit('_', 1)
|
||||
try:
|
||||
result[key] = build_single_graph(start_role, dep_type, direction, roles_dir, max_depth)
|
||||
except Exception:
|
||||
result[key] = {'nodes': [], 'links': []}
|
||||
return result
|
||||
|
||||
|
||||
@@ -124,64 +154,20 @@ def main():
|
||||
script_dir = os.path.dirname(os.path.abspath(__file__))
|
||||
default_roles_dir = os.path.abspath(os.path.join(script_dir, '..', '..', 'roles'))
|
||||
|
||||
parser = argparse.ArgumentParser(
|
||||
description="Generate graphs based on dependency mappings"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-r', '--role',
|
||||
required=True,
|
||||
help="Starting role name"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-m', '--mapping',
|
||||
nargs='+',
|
||||
default=[
|
||||
'run_after:to',
|
||||
'run_after:from',
|
||||
'dependencies:to',
|
||||
'dependencies:from'
|
||||
],
|
||||
help="Mapping entries as type:direction (default all 4 combos)"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-D', '--depth',
|
||||
type=int,
|
||||
default=0,
|
||||
help="Max recursion depth (>0) or <=0 to stop on cycle"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-o', '--output',
|
||||
choices=['yaml', 'json', 'console'],
|
||||
default='console',
|
||||
help="Output format"
|
||||
)
|
||||
parser.add_argument(
|
||||
'--roles-dir',
|
||||
default=default_roles_dir,
|
||||
help="Roles directory"
|
||||
)
|
||||
parser = argparse.ArgumentParser(description="Generate dependency graphs")
|
||||
parser.add_argument('-r', '--role', required=True, help="Starting role name")
|
||||
parser.add_argument('-D', '--depth', type=int, default=0, help="Max recursion depth")
|
||||
parser.add_argument('-o', '--output', choices=['yaml', 'json', 'console'], default='console')
|
||||
parser.add_argument('--roles-dir', default=default_roles_dir, help="Roles directory")
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
mappings: List[Dict[str, str]] = []
|
||||
for entry in args.mapping:
|
||||
if ':' not in entry:
|
||||
parser.error(f"Invalid mapping '{entry}', must be type:direction")
|
||||
dep_type, direction = entry.split(':', 1)
|
||||
if dep_type not in ('run_after', 'dependencies'):
|
||||
parser.error(f"Unknown dependency type '{dep_type}'")
|
||||
if direction not in ('to', 'from'):
|
||||
parser.error(f"Unknown direction '{direction}'")
|
||||
mappings.append({dep_type: direction})
|
||||
graphs = build_mappings(args.role, args.roles_dir, args.depth)
|
||||
|
||||
graphs = build_mappings(
|
||||
start_role=args.role,
|
||||
mappings=mappings,
|
||||
roles_dir=args.roles_dir,
|
||||
max_depth=args.depth
|
||||
)
|
||||
|
||||
for key, graph_data in graphs.items():
|
||||
for key in ALL_KEYS:
|
||||
graph_data = graphs.get(key, {'nodes': [], 'links': []})
|
||||
output_graph(graph_data, args.output, args.role, key)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
|
@@ -2,7 +2,7 @@
|
||||
import os
|
||||
import argparse
|
||||
import json
|
||||
from typing import List, Dict, Any
|
||||
from typing import Dict, Any
|
||||
|
||||
from cli.generate.graph import build_mappings, output_graph
|
||||
|
||||
@@ -21,22 +21,13 @@ def main():
|
||||
default_roles_dir = os.path.abspath(os.path.join(script_dir, '..', '..', 'roles'))
|
||||
|
||||
parser = argparse.ArgumentParser(
|
||||
description="Generate mappings-based graphs for each role and write tree.json"
|
||||
description="Generate all graphs for each role and write meta/tree.json"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-d', '--role_dir',
|
||||
default=default_roles_dir,
|
||||
help=f"Path to roles directory (default: {default_roles_dir})"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-m', '--mapping',
|
||||
nargs='+',
|
||||
default=[
|
||||
'run_after:to', 'run_after:from',
|
||||
'dependencies:to', 'dependencies:from'
|
||||
],
|
||||
help="Mapping entries as type:direction (default all 4 combos)"
|
||||
)
|
||||
parser.add_argument(
|
||||
'-D', '--depth',
|
||||
type=int,
|
||||
@@ -61,17 +52,8 @@ def main():
|
||||
)
|
||||
args = parser.parse_args()
|
||||
|
||||
# parse mappings
|
||||
mappings: List[Dict[str, str]] = []
|
||||
for entry in args.mapping:
|
||||
if ':' not in entry:
|
||||
parser.error(f"Invalid mapping '{entry}', must be type:direction")
|
||||
dep_type, direction = entry.split(':', 1)
|
||||
mappings.append({dep_type: direction})
|
||||
|
||||
if args.verbose:
|
||||
print(f"Roles directory: {args.role_dir}")
|
||||
print(f"Mappings: {mappings}")
|
||||
print(f"Max depth: {args.depth}")
|
||||
print(f"Output format: {args.output}")
|
||||
print(f"Preview mode: {args.preview}")
|
||||
@@ -80,15 +62,12 @@ def main():
|
||||
if args.verbose:
|
||||
print(f"Processing role: {role_name}")
|
||||
|
||||
# Build graphs for each mapping
|
||||
graphs = build_mappings(
|
||||
graphs: Dict[str, Any] = build_mappings(
|
||||
start_role=role_name,
|
||||
mappings=mappings,
|
||||
roles_dir=args.role_dir,
|
||||
max_depth=args.depth
|
||||
)
|
||||
|
||||
# Prepare output file or previews
|
||||
if args.preview:
|
||||
for key, data in graphs.items():
|
||||
if args.verbose:
|
||||
@@ -97,13 +76,10 @@ def main():
|
||||
else:
|
||||
tree_file = os.path.join(role_path, 'meta', 'tree.json')
|
||||
os.makedirs(os.path.dirname(tree_file), exist_ok=True)
|
||||
# write combined JSON
|
||||
with open(tree_file, 'w') as f:
|
||||
json.dump(graphs, f, indent=2)
|
||||
if args.verbose:
|
||||
print(f"Wrote {tree_file}")
|
||||
else:
|
||||
print(f"Wrote {tree_file}")
|
||||
print(f"Wrote {tree_file}")
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
|
@@ -1,6 +1,6 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
CLI for extracting invokable role paths from a nested roles YAML file using argparse.
|
||||
CLI for extracting invokable or non-invokable role paths from a nested roles YAML file using argparse.
|
||||
Assumes a default roles file at the project root if none is provided.
|
||||
"""
|
||||
|
||||
@@ -18,12 +18,12 @@ sys.path.insert(0, project_root)
|
||||
|
||||
import argparse
|
||||
import yaml
|
||||
from filter_plugins.invokable_paths import get_invokable_paths
|
||||
from filter_plugins.invokable_paths import get_invokable_paths, get_non_invokable_paths
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(
|
||||
description="Extract invokable role paths from a nested roles YAML file."
|
||||
description="Extract invokable or non-invokable role paths from a nested roles YAML file."
|
||||
)
|
||||
parser.add_argument(
|
||||
"roles_file",
|
||||
@@ -33,13 +33,33 @@ def main():
|
||||
)
|
||||
parser.add_argument(
|
||||
"--suffix", "-s",
|
||||
help="Optional suffix to append to each invokable path.",
|
||||
help="Optional suffix to append to each path.",
|
||||
default=None
|
||||
)
|
||||
|
||||
mode_group = parser.add_mutually_exclusive_group()
|
||||
mode_group.add_argument(
|
||||
"--non-invokable", "-n",
|
||||
action='store_true',
|
||||
help="List paths where 'invokable' is False or not set."
|
||||
)
|
||||
mode_group.add_argument(
|
||||
"--invokable", "-i",
|
||||
action='store_true',
|
||||
help="List paths where 'invokable' is True. (default behavior)"
|
||||
)
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
# Default to invokable if neither flag is provided
|
||||
list_non = args.non_invokable
|
||||
list_inv = args.invokable or not (args.non_invokable or args.invokable)
|
||||
|
||||
try:
|
||||
paths = get_invokable_paths(args.roles_file, args.suffix)
|
||||
if list_non:
|
||||
paths = get_non_invokable_paths(args.roles_file, args.suffix)
|
||||
else:
|
||||
paths = get_invokable_paths(args.roles_file, args.suffix)
|
||||
except FileNotFoundError as e:
|
||||
print(f"Error: {e}", file=sys.stderr)
|
||||
sys.exit(1)
|
||||
@@ -55,4 +75,4 @@ def main():
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
main()
|
||||
|
76
cli/meta/role_folder.py
Normal file
76
cli/meta/role_folder.py
Normal file
@@ -0,0 +1,76 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
CLI Script: get_role_folder_cli.py
|
||||
|
||||
This script determines the appropriate Ansible role folder based on the provided application_id
|
||||
by inspecting each role's vars/main.yml within the roles directory. By default, it assumes the
|
||||
roles directory is located at the project root, relative to this script's location.
|
||||
|
||||
Example:
|
||||
./get_role_folder_cli.py --application-id my-app-id
|
||||
"""
|
||||
import os
|
||||
import sys
|
||||
import argparse
|
||||
import yaml
|
||||
|
||||
|
||||
def get_role_folder(application_id, roles_path):
|
||||
"""
|
||||
Find the role directory under `roles_path` whose vars/main.yml contains the specified application_id.
|
||||
|
||||
:param application_id: The application_id to match.
|
||||
:param roles_path: Path to the roles directory.
|
||||
:return: The name of the matching role directory.
|
||||
:raises RuntimeError: If no match is found or if an error occurs while reading files.
|
||||
"""
|
||||
if not os.path.isdir(roles_path):
|
||||
raise RuntimeError(f"Roles path not found: {roles_path}")
|
||||
|
||||
for role in sorted(os.listdir(roles_path)):
|
||||
role_dir = os.path.join(roles_path, role)
|
||||
vars_file = os.path.join(role_dir, 'vars', 'main.yml')
|
||||
if os.path.isfile(vars_file):
|
||||
try:
|
||||
with open(vars_file, 'r') as f:
|
||||
data = yaml.safe_load(f) or {}
|
||||
except Exception as e:
|
||||
raise RuntimeError(f"Failed to load {vars_file}: {e}")
|
||||
|
||||
if data.get('application_id') == application_id:
|
||||
return role
|
||||
|
||||
raise RuntimeError(f"No role found with application_id '{application_id}' in {roles_path}")
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(
|
||||
description='Determine the Ansible role folder by application_id'
|
||||
)
|
||||
parser.add_argument(
|
||||
'application_id',
|
||||
help='The application_id defined in vars/main.yml to search for'
|
||||
)
|
||||
parser.add_argument(
|
||||
'-r', '--roles-path',
|
||||
default=os.path.join(
|
||||
os.path.dirname(os.path.realpath(__file__)),
|
||||
os.pardir, os.pardir,
|
||||
'roles'
|
||||
),
|
||||
help='Path to the roles directory (default: roles/ at project root)'
|
||||
)
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
try:
|
||||
folder = get_role_folder(args.application_id, args.roles_path)
|
||||
print(folder)
|
||||
sys.exit(0)
|
||||
except RuntimeError as err:
|
||||
print(f"Error: {err}", file=sys.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
@@ -12,10 +12,10 @@ CyMaIS covers everything from essential system setups to advanced configurations
|
||||
Every business is unique, and so is CyMaIS! With a modular architecture, it adapts to specific requirements, whether for startups, growing businesses, NGOs, or large enterprises.
|
||||
|
||||
## Proactive Monitoring & Maintenance 🔍
|
||||
With automated updates, system health checks, and security audits, CyMaIS ensures your infrastructure is always up-to-date and running smoothly. Roles such as `mon-bot-docker-container`, `mon-bot-btrfs`, and `mon-bot-webserver` help monitor system integrity.
|
||||
With automated updates, system health checks, and security audits, CyMaIS ensures your infrastructure is always up-to-date and running smoothly. Roles such as `sys-hlth-docker-container`, `sys-hlth-btrfs`, and `sys-hlth-webserver` help monitor system integrity.
|
||||
|
||||
## Uncompromised Security 🔒
|
||||
Security is a top priority! CyMaIS includes robust security features like full-disk encryption recommendations, 2FA enforcement, encrypted server deployments (`web-app-keycloak`, `svc-openldap`), and secure backup solutions (`bkp-remote-to-local`, `bkp-data-to-usb`).
|
||||
Security is a top priority! CyMaIS includes robust security features like full-disk encryption recommendations, 2FA enforcement, encrypted server deployments (`web-app-keycloak`, `svc-openldap`), and secure backup solutions (`sys-bkp-remote-to-local`, `svc-sys-bkp-data-to-usb`).
|
||||
|
||||
## User-Friendly with Expert Support 👩💻
|
||||
No need to be a Linux or Docker expert! CyMaIS simplifies deployment with intuitive role-based automation. Documentation and community support make IT administration accessible to all experience levels.
|
||||
|
@@ -10,13 +10,13 @@ Follow these guides to install and configure CyMaIS:
|
||||
|
||||
## Key Responsibilities 🔧
|
||||
- **User Management** - Configure LDAP, Keycloak, and user permissions.
|
||||
- **Security & Backups** - Set up `bkp-remote-to-local`, `bkp-data-to-usb`, and `core-security` roles.
|
||||
- **Security & Backups** - Set up `sys-bkp-remote-to-local`, `svc-sys-bkp-data-to-usb`, and `core-security` roles.
|
||||
- **Application Hosting** - Deploy services like `Nextcloud`, `Matrix`, `Gitea`, and more.
|
||||
- **Networking & VPN** - Configure `WireGuard`, `OpenVPN`, and `Nginx Reverse Proxy`.
|
||||
|
||||
## Managing & Updating CyMaIS 🔄
|
||||
- Regularly update services using `update-docker`, `update-pacman`, or `update-apt`.
|
||||
- Monitor system health with `mon-bot-btrfs`, `mon-bot-webserver`, and `mon-bot-docker-container`.
|
||||
- Automate system maintenance with `maint-lock`, `cln-backups-service`, and `maint-docker-restart`.
|
||||
- Monitor system health with `sys-hlth-btrfs`, `sys-hlth-webserver`, and `sys-hlth-docker-container`.
|
||||
- Automate system maintenance with `sys-lock`, `sys-cln-bkps-service`, and `sys-rpr-docker-hard`.
|
||||
|
||||
For more details, refer to the specific guides above.
|
@@ -5,14 +5,12 @@ from ansible.errors import AnsibleFilterError
|
||||
|
||||
class FilterModule(object):
|
||||
def filters(self):
|
||||
# module_util-Verzeichnis ermitteln und zum Import-Pfad hinzufügen
|
||||
plugin_dir = os.path.dirname(__file__)
|
||||
project_root = os.path.dirname(plugin_dir)
|
||||
module_utils = os.path.join(project_root, 'module_utils')
|
||||
if module_utils not in sys.path:
|
||||
sys.path.append(module_utils)
|
||||
|
||||
# jetzt kannst Du domain_utils importieren
|
||||
try:
|
||||
from domain_utils import get_domain
|
||||
except ImportError as e:
|
||||
|
48
filter_plugins/get_role_folder.py
Normal file
48
filter_plugins/get_role_folder.py
Normal file
@@ -0,0 +1,48 @@
|
||||
'''
|
||||
Ansible filter plugin: get_role_folder
|
||||
|
||||
This filter inspects each role under the given roles directory, loads its vars/main.yml,
|
||||
and returns the role folder name whose application_id matches the provided value.
|
||||
'''
|
||||
|
||||
from ansible.errors import AnsibleFilterError
|
||||
import os
|
||||
import yaml
|
||||
|
||||
|
||||
def get_role_folder(application_id, roles_path='roles'):
|
||||
"""
|
||||
Find the role directory under `roles_path` whose vars/main.yml contains the given application_id.
|
||||
|
||||
:param application_id: The application_id to match.
|
||||
:param roles_path: Path to the roles directory (default: 'roles').
|
||||
:return: The name of the matching role directory.
|
||||
:raises AnsibleFilterError: If vars file is unreadable or no match is found.
|
||||
"""
|
||||
if not os.path.isdir(roles_path):
|
||||
raise AnsibleFilterError(f"Roles path not found: {roles_path}")
|
||||
|
||||
for role in os.listdir(roles_path):
|
||||
role_dir = os.path.join(roles_path, role)
|
||||
vars_file = os.path.join(role_dir, 'vars', 'main.yml')
|
||||
if os.path.isfile(vars_file):
|
||||
try:
|
||||
with open(vars_file, 'r') as f:
|
||||
data = yaml.safe_load(f) or {}
|
||||
except Exception as e:
|
||||
raise AnsibleFilterError(f"Failed to load {vars_file}: {e}")
|
||||
|
||||
if data.get('application_id') == application_id:
|
||||
return role
|
||||
|
||||
raise AnsibleFilterError(f"No role found with application_id '{application_id}' in {roles_path}")
|
||||
|
||||
|
||||
class FilterModule(object):
|
||||
"""
|
||||
Register the get_role_folder filter
|
||||
"""
|
||||
def filters(self):
|
||||
return {
|
||||
'get_role_folder': get_role_folder,
|
||||
}
|
@@ -2,29 +2,18 @@ import os
|
||||
import yaml
|
||||
from typing import Dict, List, Optional
|
||||
|
||||
|
||||
def get_invokable_paths(
|
||||
roles_file: Optional[str] = None,
|
||||
suffix: Optional[str] = None
|
||||
) -> List[str]:
|
||||
"""
|
||||
Load nested roles YAML from the given file (or default at project root) and return
|
||||
dash-joined paths where 'invokable' is True. Appends suffix if provided.
|
||||
|
||||
:param roles_file: Optional path to YAML file. Defaults to '<project_root>/roles/categories.yml'.
|
||||
:param suffix: Optional suffix to append to each invokable path.
|
||||
:return: List of invokable paths.
|
||||
:raises FileNotFoundError: If the YAML file cannot be found.
|
||||
:raises yaml.YAMLError: If the YAML file cannot be parsed.
|
||||
:raises ValueError: If the root of the YAML is not a dictionary.
|
||||
Load nested roles YAML and return dash-joined paths where 'invokable' is True. Appends suffix if provided.
|
||||
"""
|
||||
# Determine default roles_file if not provided
|
||||
if not roles_file:
|
||||
script_dir = os.path.dirname(os.path.abspath(__file__))
|
||||
project_root = os.path.dirname(script_dir)
|
||||
roles_file = os.path.join(project_root, 'roles', 'categories.yml')
|
||||
|
||||
# Load and validate YAML
|
||||
try:
|
||||
with open(roles_file, 'r') as f:
|
||||
data = yaml.safe_load(f) or {}
|
||||
@@ -36,7 +25,6 @@ def get_invokable_paths(
|
||||
if not isinstance(data, dict):
|
||||
raise ValueError("YAML root is not a dictionary")
|
||||
|
||||
# Unwrap if single 'roles' key
|
||||
roles = data
|
||||
if 'roles' in roles and isinstance(roles['roles'], dict) and len(roles) == 1:
|
||||
roles = roles['roles']
|
||||
@@ -54,7 +42,6 @@ def get_invokable_paths(
|
||||
p += suffix
|
||||
found.append(p)
|
||||
|
||||
# Recurse into non-metadata child dicts
|
||||
children = {
|
||||
ck: cv for ck, cv in cfg.items()
|
||||
if ck not in METADATA and isinstance(cv, dict)
|
||||
@@ -66,6 +53,61 @@ def get_invokable_paths(
|
||||
return _recurse(roles)
|
||||
|
||||
|
||||
def get_non_invokable_paths(
|
||||
roles_file: Optional[str] = None,
|
||||
suffix: Optional[str] = None
|
||||
) -> List[str]:
|
||||
"""
|
||||
Load nested roles YAML and return dash-joined paths where 'invokable' is False or missing.
|
||||
Appends suffix if provided.
|
||||
"""
|
||||
if not roles_file:
|
||||
script_dir = os.path.dirname(os.path.abspath(__file__))
|
||||
project_root = os.path.dirname(script_dir)
|
||||
roles_file = os.path.join(project_root, 'roles', 'categories.yml')
|
||||
|
||||
try:
|
||||
with open(roles_file, 'r') as f:
|
||||
data = yaml.safe_load(f) or {}
|
||||
except FileNotFoundError:
|
||||
raise FileNotFoundError(f"Roles file not found: {roles_file}")
|
||||
except yaml.YAMLError as e:
|
||||
raise yaml.YAMLError(f"Error parsing YAML {roles_file}: {e}")
|
||||
|
||||
if not isinstance(data, dict):
|
||||
raise ValueError("YAML root is not a dictionary")
|
||||
|
||||
roles = data
|
||||
if 'roles' in roles and isinstance(roles['roles'], dict) and len(roles) == 1:
|
||||
roles = roles['roles']
|
||||
|
||||
def _recurse_non(subroles: Dict[str, dict], parent: List[str] = None) -> List[str]:
|
||||
parent = parent or []
|
||||
found: List[str] = []
|
||||
METADATA = {'title', 'description', 'icon', 'invokable'}
|
||||
|
||||
for key, cfg in subroles.items():
|
||||
path = parent + [key]
|
||||
p = '-'.join(path)
|
||||
inv = cfg.get('invokable', False)
|
||||
if not inv:
|
||||
entry = p + (suffix or "")
|
||||
found.append(entry)
|
||||
|
||||
children = {
|
||||
ck: cv for ck, cv in cfg.items()
|
||||
if ck not in METADATA and isinstance(cv, dict)
|
||||
}
|
||||
if children:
|
||||
found.extend(_recurse_non(children, path))
|
||||
return found
|
||||
|
||||
return _recurse_non(roles)
|
||||
|
||||
|
||||
class FilterModule:
|
||||
def filters(self):
|
||||
return {'invokable_paths': get_invokable_paths}
|
||||
return {
|
||||
'invokable_paths': get_invokable_paths,
|
||||
'non_invokable_paths': get_non_invokable_paths
|
||||
}
|
||||
|
@@ -13,23 +13,23 @@ system_maintenance_lock_timeout_restart_docker: "{{system_maintenance_lock_ti
|
||||
|
||||
### Defined Services for Backup Tasks
|
||||
system_maintenance_backup_services:
|
||||
- "bkp-docker-to-local"
|
||||
- "bkp-remote-to-local"
|
||||
- "bkp-data-to-usb"
|
||||
- "bkp-docker-to-local-everything"
|
||||
- "sys-bkp-docker-to-local"
|
||||
- "svc-sys-bkp-rmt-2-loc"
|
||||
- "svc-sys-bkp-data-to-usb"
|
||||
- "sys-bkp-docker-to-local-everything"
|
||||
|
||||
### Defined Services for System Cleanup
|
||||
system_maintenance_cleanup_services:
|
||||
- "cln-backups"
|
||||
- "cln-disc-space"
|
||||
- "cln-failed-docker-backups"
|
||||
- "sys-cln-backups"
|
||||
- "sys-cln-disc-space"
|
||||
- "sys-cln-faild-bkps"
|
||||
|
||||
### Services that Manipulate the System
|
||||
system_maintenance_manipulation_services:
|
||||
- "maint-docker-heal"
|
||||
- "sys-rpr-docker-soft"
|
||||
- "update-docker"
|
||||
- "cln-docker-storage-optimizer"
|
||||
- "maint-docker-restart"
|
||||
- "sys-opt-ssd-hdd"
|
||||
- "sys-rpr-docker-hard"
|
||||
|
||||
## Total System Maintenance Services
|
||||
system_maintenance_services: "{{ system_maintenance_backup_services + system_maintenance_cleanup_services + system_maintenance_manipulation_services }}"
|
||||
|
@@ -3,4 +3,4 @@
|
||||
path_administrator_home: "/home/administrator/"
|
||||
path_administrator_scripts: "/opt/scripts/"
|
||||
path_docker_compose_instances: "/opt/docker/"
|
||||
path_system_lock_script: "/opt/scripts/maint-lock.py"
|
||||
path_system_lock_script: "/opt/scripts/sys-lock.py"
|
@@ -15,7 +15,7 @@ For a complete list of role categories and detailed definitions, see:
|
||||
Fundamental system configuration (SSH, journald, sudo, etc.)
|
||||
|
||||
- **gen-***
|
||||
Generic helpers and language/tool installers (e.g. `gen-git`, `gen-locales`, `gen-timer`)
|
||||
Generic helpers and language/tool installers (e.g. `gen-git`, `gen-locales`, `sys-timer`)
|
||||
|
||||
- **desk-***
|
||||
Desktop environment and application roles (e.g. `desk-gnome`, `desk-browser`, `desk-libreoffice`)
|
||||
@@ -59,13 +59,13 @@ For a complete list of role categories and detailed definitions, see:
|
||||
|
||||
## Monitoring & Alerting
|
||||
|
||||
- **mon-bot-***
|
||||
- **sys-hlth-***
|
||||
“Bot”-style health checks (Btrfs, disk‐space, Docker, journalctl, CSP crawler, webserver) with alerts.
|
||||
|
||||
- **monitor-core-***
|
||||
Low-level system monitors (journalctl, Docker containers, disk space, etc.)
|
||||
|
||||
- **alert-***
|
||||
- **sys-alm-***
|
||||
Notification handlers for failures (core, email, Telegram).
|
||||
|
||||
---
|
||||
@@ -78,14 +78,14 @@ For a complete list of role categories and detailed definitions, see:
|
||||
- **maint-docker-***
|
||||
Automated recovery and restarts for Docker Compose workloads.
|
||||
|
||||
- **cln-***
|
||||
- **sys-cln-***
|
||||
Housekeeping tasks (old backups, expired certs, log rotation).
|
||||
|
||||
---
|
||||
|
||||
## Backup & Restore
|
||||
|
||||
- **bkp-***
|
||||
- **sys-bkp-***
|
||||
Local and remote backup strategies for files, Docker volumes, databases.
|
||||
|
||||
---
|
||||
@@ -112,8 +112,8 @@ For a complete list of role categories and detailed definitions, see:
|
||||
|
||||
> **Tip:** To find a role quickly, search for its prefix:
|
||||
> `core-`, `gen-`, `desk-`, `srv-web-`, `web-svc-`, `web-app-`,
|
||||
> `net-`, `svc-`, `mon-bot-`, `monitor-core-`, `alert-`,
|
||||
> `maint-`, `maint-docker-`, `cln-`, `bkp-`, `update-`,
|
||||
> `net-`, `svc-`, `sys-hlth-`, `monitor-core-`, `sys-alm-`,
|
||||
> `maint-`, `maint-docker-`, `sys-cln-`, `sys-bkp-`, `update-`,
|
||||
> `pkgmgr-`, `user-`.
|
||||
|
||||
---
|
||||
|
@@ -1,8 +0,0 @@
|
||||
[Unit]
|
||||
Description=Notifier for %i
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/usr/bin/systemctl start alert-telegram.cymais@%i.service alert-email.cymais@%i.service
|
||||
User=root
|
||||
Group=systemd-journal
|
@@ -1 +0,0 @@
|
||||
application_id: compose
|
@@ -1,4 +0,0 @@
|
||||
- name: "restart alert-email service"
|
||||
systemd:
|
||||
name: alert-email.cymais.service
|
||||
daemon_reload: yes
|
@@ -1,2 +0,0 @@
|
||||
systemd_notifier_email_folder: '{{path_administrator_scripts}}alert-email/'
|
||||
application_id: email
|
@@ -1,4 +0,0 @@
|
||||
- name: "restart alert-telegram service"
|
||||
systemd:
|
||||
name: alert-telegram.cymais.service
|
||||
daemon_reload: yes
|
@@ -1,3 +0,0 @@
|
||||
systemd_telegram_folder: /opt/ansible-roles/alert-telegram/
|
||||
systemd_telegram_script: '{{systemd_telegram_folder}}alert-telegram.sh'
|
||||
application_id: telegram
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload bkp-data-to-usb.cymais.service"
|
||||
systemd:
|
||||
name: bkp-data-to-usb.cymais.service
|
||||
state: reloaded
|
||||
daemon_reload: yes
|
@@ -1,2 +0,0 @@
|
||||
backup_directory_validator_folder: '{{path_administrator_scripts}}directory-validator/'
|
||||
application_id: directory-validator
|
@@ -1,9 +0,0 @@
|
||||
- name: "reload bkp-docker-to-local-everything.cymais.service"
|
||||
systemd:
|
||||
name: bkp-docker-to-local-everything.cymais.service
|
||||
daemon_reload: yes
|
||||
|
||||
- name: "reload bkp-docker-to-local.cymais.service"
|
||||
systemd:
|
||||
name: bkp-docker-to-local.cymais.service
|
||||
daemon_reload: yes
|
@@ -1 +0,0 @@
|
||||
application_id: provider
|
@@ -1,4 +0,0 @@
|
||||
- name: "reload bkp-remote-to-local service"
|
||||
systemd:
|
||||
name: bkp-remote-to-local.cymais.service
|
||||
daemon_reload: yes
|
@@ -1,2 +0,0 @@
|
||||
docker_backup_remote_to_local_folder: '{{path_administrator_scripts}}bkp-remote-to-local/'
|
||||
application_id: bkp-remote-to-local
|
@@ -1,14 +1,42 @@
|
||||
roles:
|
||||
core:
|
||||
title: "Core & System"
|
||||
description: "Fundamental system configuration"
|
||||
icon: "fas fa-cogs"
|
||||
invokable: true
|
||||
sys:
|
||||
title: "System"
|
||||
invokable: false
|
||||
alm:
|
||||
title: "Alerting"
|
||||
description: "Notification handlers for system events"
|
||||
icon: "fas fa-bell"
|
||||
invokable: false
|
||||
cln:
|
||||
title: "Cleanup"
|
||||
description: "Roles for cleaning up various system resources—old backups, unused certificates, temporary files, Docker volumes, disk caches, deprecated domains, and more."
|
||||
icon: "fas fa-trash-alt"
|
||||
invokable: false
|
||||
hlth:
|
||||
title: "Monitoring"
|
||||
description: "Roles for system monitoring and health checks—encompassing bot-style automated checks and core low-level monitors for logs, containers, disk usage, and more."
|
||||
icon: "fas fa-chart-area"
|
||||
invokable: false
|
||||
bkp:
|
||||
title: "Backup & Restore"
|
||||
description: "Backup strategies & restore procedures"
|
||||
icon: "fas fa-hdd"
|
||||
invokable: false
|
||||
update:
|
||||
title: "Updates & Package Management"
|
||||
description: "OS & package updates"
|
||||
icon: "fas fa-sync"
|
||||
invokable: false
|
||||
drv:
|
||||
title: "Drivers"
|
||||
description: "Roles for installing and configuring hardware drivers—covering printers, graphics, input devices, and other peripheral support."
|
||||
icon: "fas fa-microchip"
|
||||
invokable: true
|
||||
# core:
|
||||
# title: "Core & System"
|
||||
# description: "Fundamental system configuration"
|
||||
# icon: "fas fa-cogs"
|
||||
# invokable: true
|
||||
gen:
|
||||
title: "Generic"
|
||||
description: "Helper roles & installers (git, locales, timer, etc.)"
|
||||
@@ -73,38 +101,8 @@ roles:
|
||||
description: "Docker infrastructure services (DBMS, LDAP, Redis, etc.)"
|
||||
icon: "fas fa-database"
|
||||
invokable: true
|
||||
mon:
|
||||
title: "Monitoring"
|
||||
description: "Roles for system monitoring and health checks—encompassing bot-style automated checks and core low-level monitors for logs, containers, disk usage, and more."
|
||||
icon: "fas fa-chart-area"
|
||||
invokable: true
|
||||
alert:
|
||||
title: "Alerting"
|
||||
description: "Notification handlers for system events"
|
||||
icon: "fas fa-bell"
|
||||
invokable: true
|
||||
maint:
|
||||
title: "Maintenance & Healing"
|
||||
description: "Periodic maintenance & auto-recovery"
|
||||
icon: "fas fa-tools"
|
||||
invokable: true
|
||||
bkp:
|
||||
title: "Backup & Restore"
|
||||
description: "Backup strategies & restore procedures"
|
||||
icon: "fas fa-hdd"
|
||||
invokable: true
|
||||
update:
|
||||
title: "Updates & Package Management"
|
||||
description: "OS & package updates"
|
||||
icon: "fas fa-sync"
|
||||
invokable: true
|
||||
user:
|
||||
title: "Users & Access"
|
||||
description: "User accounts & access control"
|
||||
icon: "fas fa-users"
|
||||
invokable: false
|
||||
cln:
|
||||
title: "Cleanup"
|
||||
description: "Roles for cleaning up various system resources—old backups, unused certificates, temporary files, Docker volumes, disk caches, deprecated domains, and more."
|
||||
icon: "fas fa-trash-alt"
|
||||
invokable: true
|
||||
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload cln-backups.cymais.service"
|
||||
systemd:
|
||||
name: cln-backups.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,2 +0,0 @@
|
||||
cleanup_backups_directory: '{{path_administrator_scripts}}cln-backups/'
|
||||
application_id: backups-service
|
@@ -1 +0,0 @@
|
||||
application_id: backups-timer
|
@@ -1,6 +0,0 @@
|
||||
- name: "Reload and restart cln-certs.cymais.service"
|
||||
systemd:
|
||||
name: cln-certs.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
||||
state: restarted
|
@@ -1 +0,0 @@
|
||||
application_id: certs
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload cln-disc-space.cymais.service"
|
||||
systemd:
|
||||
name: cln-disc-space.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,26 +0,0 @@
|
||||
- name: "create {{cleanup_disc_space_folder}}"
|
||||
file:
|
||||
path: "{{cleanup_disc_space_folder}}"
|
||||
state: directory
|
||||
mode: 0755
|
||||
|
||||
- name: create cln-disc-space.sh
|
||||
template:
|
||||
src: cln-disc-space.sh.j2
|
||||
dest: "{{cleanup_disc_space_folder}}cln-disc-space.sh"
|
||||
|
||||
- name: create cln-disc-space.cymais.service
|
||||
template:
|
||||
src: cln-disc-space.service.j2
|
||||
dest: /etc/systemd/system/cln-disc-space.cymais.service
|
||||
notify: reload cln-disc-space.cymais.service
|
||||
|
||||
- name: "set 'service_name' to '{{ role_name }}'"
|
||||
set_fact:
|
||||
service_name: "{{ role_name }}"
|
||||
|
||||
- name: "include role for gen-timer for {{service_name}}"
|
||||
include_role:
|
||||
name: gen-timer
|
||||
vars:
|
||||
on_calendar: "{{on_calendar_cleanup_disc_space}}"
|
@@ -1,2 +0,0 @@
|
||||
cleanup_disc_space_folder: '{{path_administrator_scripts}}cln-disc-space/'
|
||||
application_id: disc-space
|
@@ -1 +0,0 @@
|
||||
application_id: docker-anonymous-volumes
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload cln-docker-storage-optimizer.cymais.service"
|
||||
systemd:
|
||||
name: cln-docker-storage-optimizer.cymais.service
|
||||
state: reloaded
|
||||
daemon_reload: yes
|
@@ -1,22 +0,0 @@
|
||||
- name: "create {{storage_optimizer_directory}}"
|
||||
file:
|
||||
path: "{{storage_optimizer_directory}}"
|
||||
state: directory
|
||||
mode: 0755
|
||||
|
||||
- name: create cln-docker-storage-optimizer.cymais.service
|
||||
template:
|
||||
src: cln-docker-storage-optimizer.service.j2
|
||||
dest: /etc/systemd/system/cln-docker-storage-optimizer.cymais.service
|
||||
notify: reload cln-docker-storage-optimizer.cymais.service
|
||||
|
||||
- name: create cln-docker-storage-optimizer.py
|
||||
copy:
|
||||
src: cln-docker-storage-optimizer.py
|
||||
dest: "{{storage_optimizer_script}}"
|
||||
mode: 0755
|
||||
|
||||
- name: "optimize storage performance"
|
||||
systemd:
|
||||
name: cln-docker-storage-optimizer.cymais.service
|
||||
state: started
|
@@ -1,3 +0,0 @@
|
||||
storage_optimizer_directory: '{{path_administrator_scripts}}cln-docker-storage-optimizer/'
|
||||
storage_optimizer_script: '{{storage_optimizer_directory}}cln-docker-storage-optimizer.py'
|
||||
application_id: docker-storage-optimizer
|
@@ -1 +0,0 @@
|
||||
application_id: domains
|
@@ -1,5 +0,0 @@
|
||||
- name: "Reload cln-failed-docker-backups.cymais.service"
|
||||
systemd:
|
||||
name: cln-failed-docker-backups.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -13,4 +13,4 @@
|
||||
when: applications | is_feature_enabled('central_database',application_id)
|
||||
|
||||
- name: "Add database to backup"
|
||||
include_tasks: "{{ playbook_dir }}/roles/bkp-docker-to-local/tasks/seed-database-to-backup.yml"
|
||||
include_tasks: "{{ playbook_dir }}/roles/sys-bkp-docker-to-local/tasks/seed-database-to-backup.yml"
|
@@ -1 +0,0 @@
|
||||
application_id: daemon
|
@@ -1 +0,0 @@
|
||||
application_id: journalctl
|
@@ -1,21 +0,0 @@
|
||||
# System Security
|
||||
|
||||
## Description
|
||||
|
||||
This role configures security-related settings on the target system to ensure that essential security services are properly managed. It reinforces system security by integrating with related roles and enforcing best practices.
|
||||
|
||||
## Overview
|
||||
|
||||
The role focuses on:
|
||||
- Configuring key security services such as [SSH](https://en.wikipedia.org/wiki/Secure_Shell).
|
||||
- Integrating with other security roles to maintain a robust system security posture.
|
||||
- Ensuring that security configurations are applied consistently across the system.
|
||||
|
||||
## Purpose
|
||||
|
||||
The primary purpose of this role is to enhance overall system security by managing and configuring essential security services and settings.
|
||||
|
||||
## Features
|
||||
|
||||
- **Security Configuration:** Applies recommended security settings for core services.
|
||||
- **Integration:** Works alongside related roles such as [sshd](https://en.wikipedia.org/wiki/SSH_daemon) to provide comprehensive security management.
|
@@ -1,24 +0,0 @@
|
||||
---
|
||||
galaxy_info:
|
||||
author: "Kevin Veen-Birkenbach"
|
||||
description: "Configures security-related settings on the target system and ensures essential security services are properly managed."
|
||||
license: "CyMaIS NonCommercial License (CNCL)"
|
||||
license_url: "https://s.veen.world/cncl"
|
||||
company: |
|
||||
Kevin Veen-Birkenbach
|
||||
Consulting & Coaching Solutions
|
||||
https://www.veen.world
|
||||
min_ansible_version: "2.9"
|
||||
platforms:
|
||||
- name: Linux
|
||||
versions:
|
||||
- all
|
||||
galaxy_tags:
|
||||
- security
|
||||
- ssh
|
||||
- system
|
||||
repository: "https://s.veen.world/cymais"
|
||||
issue_tracker_url: "https://s.veen.world/cymaisissues"
|
||||
documentation: "https://s.veen.world/cymais"
|
||||
dependencies:
|
||||
- core-sshd
|
@@ -1 +0,0 @@
|
||||
application_id: security
|
@@ -1 +0,0 @@
|
||||
application_id: sshd
|
@@ -1 +0,0 @@
|
||||
application_id: sudo
|
@@ -10,7 +10,7 @@ Targeting user environments on Arch Linux (e.g., Manjaro), this role sets up the
|
||||
|
||||
## Purpose
|
||||
|
||||
The purpose of this role is to automate the configuration of cloud-integrated user directories by ensuring that common folders like `Downloads`, `Music`, and `Workspaces` are transparently redirected into a centralized cloud structure. This makes it easier to maintain bkp-friendly, cloud-ready setups for homelab and professional workflows.
|
||||
The purpose of this role is to automate the configuration of cloud-integrated user directories by ensuring that common folders like `Downloads`, `Music`, and `Workspaces` are transparently redirected into a centralized cloud structure. This makes it easier to maintain sys-bkp-friendly, cloud-ready setups for homelab and professional workflows.
|
||||
|
||||
## Features
|
||||
|
||||
|
@@ -27,9 +27,9 @@ galaxy_info:
|
||||
documentation: "https://s.veen.world/cymais/docker"
|
||||
|
||||
dependencies:
|
||||
- bkp-docker-to-local
|
||||
- sys-bkp-docker-to-local
|
||||
- user-administrator
|
||||
- mon-bot-docker-container
|
||||
- mon-bot-docker-volumes
|
||||
- maint-docker-heal
|
||||
- maint-docker-restart
|
||||
- sys-hlth-docker-container
|
||||
- sys-hlth-docker-volumes
|
||||
- sys-rpr-docker-soft
|
||||
- sys-rpr-docker-hard
|
||||
|
@@ -16,7 +16,7 @@
|
||||
group: administrator
|
||||
when: run_once_docker is not defined
|
||||
|
||||
- name: Set docker_enabled to true, to activate cln-docker-storage-optimizer
|
||||
- name: Set docker_enabled to true, to activate sys-opt-ssd-hdd
|
||||
set_fact:
|
||||
docker_enabled: true
|
||||
when: run_once_docker is not defined
|
||||
|
@@ -9,4 +9,4 @@ galaxy_info:
|
||||
- all
|
||||
dependencies:
|
||||
- gen-aur-helper
|
||||
- alert-compose
|
||||
- sys-alm-compose
|
||||
|
@@ -25,9 +25,9 @@
|
||||
set_fact:
|
||||
service_name: "{{ role_name }}"
|
||||
|
||||
- name: "include role for gen-timer for {{service_name}}"
|
||||
- name: "include role for sys-timer for {{service_name}}"
|
||||
include_role:
|
||||
name: gen-timer
|
||||
name: sys-timer
|
||||
vars:
|
||||
on_calendar: "{{on_calendar_msi_keyboard_color}}"
|
||||
persistent: "true"
|
||||
|
@@ -1,6 +1,6 @@
|
||||
[Unit]
|
||||
Description=Keyboard Color Service
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
OnFailure=sys-alm-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
|
@@ -1 +0,0 @@
|
||||
application_id: msmtp
|
@@ -1 +0,0 @@
|
||||
application_id: timer
|
@@ -1,4 +0,0 @@
|
||||
- name: "reload maint-btrfs-auto-balancer.cymais.service"
|
||||
systemd:
|
||||
name: maint-btrfs-auto-balancer.cymais.service
|
||||
daemon_reload: yes
|
@@ -1,5 +0,0 @@
|
||||
- name: restart maint-docker-heal.cymais.service
|
||||
systemd:
|
||||
name: maint-docker-heal.cymais.service
|
||||
state: restarted
|
||||
daemon_reload: yes
|
@@ -1,8 +0,0 @@
|
||||
[Unit]
|
||||
Description=restart unhealthy docker containers
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStartPre=/bin/sh -c '/usr/bin/python {{ path_system_lock_script }} {{ system_maintenance_services | join(' ') }} --ignore {{system_maintenance_cleanup_services| join(' ') }} maint-docker-heal --timeout "{{system_maintenance_lock_timeout_heal_docker}}"'
|
||||
ExecStart=/bin/sh -c '/bin/python {{heal_docker}}maint-docker-heal.py {{path_docker_compose_instances}}'
|
@@ -1,2 +0,0 @@
|
||||
heal_docker: '{{path_administrator_scripts}}maint-docker-heal/'
|
||||
application_id: docker-heal
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload maint-docker-restart.cymais.service"
|
||||
systemd:
|
||||
name: maint-docker-restart.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,3 +0,0 @@
|
||||
restart_docker_folder: '{{path_administrator_scripts}}maint-docker-restart/'
|
||||
restart_docker_script: '{{restart_docker_folder}}maint-docker-restart.py'
|
||||
application_id: docker-restart
|
@@ -1 +0,0 @@
|
||||
application_id: lock
|
@@ -1 +0,0 @@
|
||||
application_id: swapfile
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-btrfs.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-btrfs.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,26 +0,0 @@
|
||||
- name: "create {{docker_health_btrfs_folder}}"
|
||||
file:
|
||||
path: "{{docker_health_btrfs_folder}}"
|
||||
state: directory
|
||||
mode: 0755
|
||||
|
||||
- name: create mon-bot-btrfs.sh
|
||||
copy:
|
||||
src: mon-bot-btrfs.sh
|
||||
dest: "{{docker_health_btrfs_folder}}mon-bot-btrfs.sh"
|
||||
|
||||
- name: create mon-bot-btrfs.cymais.service
|
||||
template:
|
||||
src: mon-bot-btrfs.service.j2
|
||||
dest: /etc/systemd/system/mon-bot-btrfs.cymais.service
|
||||
notify: reload mon-bot-btrfs.cymais.service
|
||||
|
||||
- name: "set 'service_name' to '{{ role_name }}'"
|
||||
set_fact:
|
||||
service_name: "{{ role_name }}"
|
||||
|
||||
- name: "include role for gen-timer for {{service_name}}"
|
||||
include_role:
|
||||
name: gen-timer
|
||||
vars:
|
||||
on_calendar: "{{on_calendar_health_btrfs}}"
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=Check btrfs status
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/bin/bash {{docker_health_btrfs_folder}}mon-bot-btrfs.sh
|
@@ -1,2 +0,0 @@
|
||||
docker_health_btrfs_folder: '{{path_administrator_scripts}}mon-bot-btrfs/'
|
||||
application_id: bot-btrfs
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-csp.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-csp.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,3 +0,0 @@
|
||||
health_csp_crawler_folder: '{{ path_administrator_scripts }}mon-bot-csp/'
|
||||
health_csp_crawler_script: '{{ health_csp_crawler_folder }}mon-bot-csp.py'
|
||||
application_id: bot-csp
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-disc-space.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-disc-space.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,26 +0,0 @@
|
||||
- name: "create {{health_disc_space_folder}}"
|
||||
file:
|
||||
path: "{{health_disc_space_folder}}"
|
||||
state: directory
|
||||
mode: 0755
|
||||
|
||||
- name: create mon-bot-disc-space.sh
|
||||
copy:
|
||||
src: mon-bot-disc-space.sh
|
||||
dest: "{{health_disc_space_folder}}mon-bot-disc-space.sh"
|
||||
|
||||
- name: create mon-bot-disc-space.cymais.service
|
||||
template:
|
||||
src: mon-bot-disc-space.service.j2
|
||||
dest: /etc/systemd/system/mon-bot-disc-space.cymais.service
|
||||
notify: reload mon-bot-disc-space.cymais.service
|
||||
|
||||
- name: "set 'service_name' to '{{ role_name }}'"
|
||||
set_fact:
|
||||
service_name: "{{ role_name }}"
|
||||
|
||||
- name: "include role for gen-timer for {{service_name}}"
|
||||
include_role:
|
||||
name: gen-timer
|
||||
vars:
|
||||
on_calendar: "{{on_calendar_health_disc_space}}"
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=checking disc space
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/bin/bash {{health_disc_space_folder}}mon-bot-disc-space.sh {{size_percent_disc_space_warning}}
|
@@ -1,2 +0,0 @@
|
||||
health_disc_space_folder: '{{path_administrator_scripts}}mon-bot-disc-space/'
|
||||
application_id: bot-disc-space
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-docker-container.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-docker-container.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=Checking docker health
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/bin/bash {{health_docker_container_folder}}mon-bot-docker-container.sh
|
@@ -1,2 +0,0 @@
|
||||
health_docker_container_folder: '{{path_administrator_scripts}}mon-bot-docker-container/'
|
||||
application_id: bot-docker-container
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-docker-volumes.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-docker-volumes.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=Checking docker health
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/bin/bash {{ health_docker_volumes_folder }}mon-bot-docker-volumes.sh "{{ whitelisted_anonymous_docker_volumes | join(' ') }}"
|
@@ -1,2 +0,0 @@
|
||||
health_docker_volumes_folder: '{{path_administrator_scripts}}mon-bot-docker-volumes/'
|
||||
application_id: bot-docker-volumes
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-journalctl.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-journalctl.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=checking journalctl health
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/bin/bash {{health_journalctl_folder}}mon-bot-journalctl.sh
|
@@ -1,2 +0,0 @@
|
||||
health_journalctl_folder: '{{path_administrator_scripts}}mon-bot-journalctl/'
|
||||
application_id: bot-journalctl
|
@@ -1,21 +0,0 @@
|
||||
# mon-bot-msmtp
|
||||
|
||||
## Description
|
||||
|
||||
This Ansible role sends periodic health check emails using **msmtp** to verify that your mail transport agent is operational. It deploys a simple script and hooks it into a systemd service and timer, with failure notifications sent via Telegram.
|
||||
|
||||
## Overview
|
||||
|
||||
Optimized for Archlinux, this role creates the required directory structure, installs and configures the mon-bot-check script, and integrates with the **alert-telegram** role. It uses the **gen-timer** role to schedule regular checks based on your customizable `OnCalendar` setting.
|
||||
|
||||
## Purpose
|
||||
|
||||
The **mon-bot-msmtp** role ensures that your mail transport system stays available by sending a test email at defined intervals. If the email fails, a Telegram alert is triggered, allowing you to detect and address issues before they impact users.
|
||||
|
||||
## Features
|
||||
|
||||
- **Directory & Script Deployment:** Sets up `mon-bot-msmtp/` and deploys a templated Bash script to send test emails via msmtp.
|
||||
- **Systemd Service & Timer:** Provides `.service` and `.timer` units to run the check and schedule it automatically.
|
||||
- **Failure Notifications:** Leverages **alert-telegram** to push alerts when the script exits with an error.
|
||||
- **Configurable Schedule:** Define your desired check frequency using the `on_calendar_health_msmtp` variable.
|
||||
- **Email Destination:** Specify the recipient via the `users.administrator.email` variable.
|
@@ -1,5 +0,0 @@
|
||||
- name: reload mon-bot-msmtp.cymais.service
|
||||
systemd:
|
||||
name: mon-bot-msmtp.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,27 +0,0 @@
|
||||
- name: "create {{ health_msmtp_folder }}"
|
||||
file:
|
||||
path: "{{ health_msmtp_folder }}"
|
||||
state: directory
|
||||
mode: 0755
|
||||
|
||||
- name: create mon-bot-msmtp.sh
|
||||
template:
|
||||
src: mon-bot-msmtp.sh.j2
|
||||
dest: "{{ health_msmtp_folder }}mon-bot-msmtp.sh"
|
||||
mode: '0755'
|
||||
|
||||
- name: create mon-bot-msmtp.cymais.service
|
||||
template:
|
||||
src: mon-bot-msmtp.service.j2
|
||||
dest: /etc/systemd/system/mon-bot-msmtp.cymais.service
|
||||
notify: reload mon-bot-msmtp.cymais.service
|
||||
|
||||
- name: "set 'service_name' to '{{ role_name }}'"
|
||||
set_fact:
|
||||
service_name: "{{ role_name }}"
|
||||
|
||||
- name: include role for gen-timer for {{ service_name }}
|
||||
include_role:
|
||||
name: gen-timer
|
||||
vars:
|
||||
on_calendar: "{{ on_calendar_health_msmtp }}"
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=Check msmtp liveliness
|
||||
OnFailure=alert-telegram.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/bin/bash {{ health_msmtp_folder }}mon-bot-msmtp.sh
|
@@ -1,2 +0,0 @@
|
||||
health_msmtp_folder: '{{ path_administrator_scripts }}mon-bot-msmtp/'
|
||||
application_id: bot-msmtp
|
@@ -1,5 +0,0 @@
|
||||
- name: "reload mon-bot-webserver.cymais.service"
|
||||
systemd:
|
||||
name: mon-bot-webserver.cymais.service
|
||||
enabled: yes
|
||||
daemon_reload: yes
|
@@ -1,7 +0,0 @@
|
||||
[Unit]
|
||||
Description=Check nginx configuration status
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
ExecStart=/usr/bin/python3 {{ health_nginx_folder }}mon-bot-webserver.py
|
@@ -1,2 +0,0 @@
|
||||
health_nginx_folder: '{{ path_administrator_scripts }}mon-bot-webserver/'
|
||||
application_id: bot-webserver
|
@@ -1 +0,0 @@
|
||||
application_id: dns-records
|
@@ -13,7 +13,7 @@ This Ansible role simplifies the deployment of **Let's Encrypt certificates** in
|
||||
- Automatically **reloads or restarts Nginx services** when certificates are updated.
|
||||
- **Configures and manages a `systemd` service** for automated certificate deployment.
|
||||
- **Includes a `systemd` timer** for scheduled renewals.
|
||||
- **Handles dependent services** like `alert-compose`.
|
||||
- **Handles dependent services** like `sys-alm-compose`.
|
||||
|
||||
---
|
||||
|
||||
@@ -29,7 +29,7 @@ This Ansible role simplifies the deployment of **Let's Encrypt certificates** in
|
||||
3. **Configure `systemd` Service**
|
||||
- Deploys a `systemd` service file for the deployment process.
|
||||
|
||||
4. **Include `gen-timer` Role**
|
||||
4. **Include `sys-timer` Role**
|
||||
- Schedules automatic certificate deployment using a `systemd` timer.
|
||||
|
||||
### **2️⃣ Handlers**
|
||||
|
@@ -21,5 +21,5 @@ galaxy_info:
|
||||
- name: Any
|
||||
versions: [ all ]
|
||||
dependencies:
|
||||
- alert-compose
|
||||
- sys-alm-compose
|
||||
|
||||
|
@@ -23,9 +23,9 @@
|
||||
dest: "/etc/systemd/system/srv-proxy-6-6-tls-deploy.{{application_id}}.cymais.service"
|
||||
notify: restart srv-proxy-6-6-tls-deploy.cymais.service
|
||||
|
||||
- name: "include role for gen-timer for {{service_name}}"
|
||||
- name: "include role for sys-timer for {{service_name}}"
|
||||
include_role:
|
||||
name: gen-timer
|
||||
name: sys-timer
|
||||
vars:
|
||||
on_calendar: "{{on_calendar_deploy_certificates}}"
|
||||
service_name: "srv-proxy-6-6-tls-deploy.{{application_id}}"
|
||||
|
@@ -1,6 +1,6 @@
|
||||
[Unit]
|
||||
Description=Let's Encrypt deploy to {{docker_compose.directories.instance}}
|
||||
OnFailure=alert-compose.cymais@%n.service
|
||||
OnFailure=sys-alm-compose.cymais@%n.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
|
@@ -13,7 +13,7 @@ server
|
||||
{{ proxy_extra_configuration }}
|
||||
{% endif %}
|
||||
|
||||
{% include 'roles/net-letsencrypt/templates/ssl_header.j2' %}
|
||||
{% include 'roles/srv-web-7-7-letsencrypt/templates/ssl_header.j2' %}
|
||||
|
||||
{% if applications | is_feature_enabled('oauth2', application_id) %}
|
||||
{% set acl = applications[application_id].oauth2_proxy.acl | default({}) %}
|
||||
|
@@ -6,7 +6,7 @@ map $http_upgrade $connection_upgrade {
|
||||
server {
|
||||
server_name {{ domain }};
|
||||
|
||||
{% include 'roles/net-letsencrypt/templates/ssl_header.j2' %}
|
||||
{% include 'roles/srv-web-7-7-letsencrypt/templates/ssl_header.j2' %}
|
||||
{% include 'roles/srv-web-7-7-inj-compose/templates/global.includes.conf.j2' %}
|
||||
|
||||
client_max_body_size {{ client_max_body_size | default('100m') }};
|
||||
|
@@ -6,12 +6,12 @@ This role automates the setup of an automatic [Let's Encrypt](https://letsencryp
|
||||
|
||||
## 📖 Overview
|
||||
|
||||
Optimized for Archlinux systems, this role installs the `certbot-nginx` package, configures a dedicated `systemd` service for certificate renewal, and integrates with a `gen-timer` to schedule periodic renewals. After a renewal, Nginx is reloaded to apply the updated certificates immediately.
|
||||
Optimized for Archlinux systems, this role installs the `certbot-nginx` package, configures a dedicated `systemd` service for certificate renewal, and integrates with a `sys-timer` to schedule periodic renewals. After a renewal, Nginx is reloaded to apply the updated certificates immediately.
|
||||
|
||||
### Key Features
|
||||
- **Automatic Renewal:** Schedules unattended certificate renewals using gen-timers.
|
||||
- **Automatic Renewal:** Schedules unattended certificate renewals using sys-timers.
|
||||
- **Seamless Nginx Reload:** Reloads the Nginx service automatically after successful renewals.
|
||||
- **Systemd Integration:** Manages renewal operations reliably with `systemd` and `alert-compose`.
|
||||
- **Systemd Integration:** Manages renewal operations reliably with `systemd` and `sys-alm-compose`.
|
||||
- **Quiet and Safe Operation:** Uses `--quiet` and `--agree-tos` flags to ensure non-interactive renewals.
|
||||
|
||||
## 🎯 Purpose
|
||||
@@ -22,8 +22,8 @@ The Nginx Certbot Automation role ensures that Let's Encrypt SSL/TLS certificate
|
||||
|
||||
- **Certbot-Nginx Package Installation:** Installs required certbot plugins for Nginx.
|
||||
- **Custom Systemd Service:** Configures a lightweight, dedicated renewal service.
|
||||
- **Timer Setup:** Uses gen-timer to run certbot renewals periodically.
|
||||
- **Failure Notification:** Integrated with `alert-compose` for alerting on failures.
|
||||
- **Timer Setup:** Uses sys-timer to run certbot renewals periodically.
|
||||
- **Failure Notification:** Integrated with `sys-alm-compose` for alerting on failures.
|
||||
|
||||
## 🔗 Learn More
|
||||
|
||||
|
@@ -27,7 +27,7 @@ galaxy_info:
|
||||
issue_tracker_url: "https://s.veen.world/cymaisissues"
|
||||
documentation: "https://s.veen.world/cymais"
|
||||
dependencies:
|
||||
- gen-certbot
|
||||
- srv-web-7-7-certbot
|
||||
- srv-web-7-4-core
|
||||
- alert-compose
|
||||
- cln-certs
|
||||
- sys-alm-compose
|
||||
- sys-cln-certs
|
||||
|
@@ -17,9 +17,9 @@
|
||||
service_name: "{{ role_name }}"
|
||||
when: run_once_nginx_certbot is not defined
|
||||
|
||||
- name: "include role for gen-timer for {{service_name}}"
|
||||
- name: "include role for sys-timer for {{service_name}}"
|
||||
include_role:
|
||||
name: gen-timer
|
||||
name: sys-timer
|
||||
vars:
|
||||
on_calendar: "{{on_calendar_renew_lets_encrypt_certificates}}"
|
||||
persistent: "true"
|
||||
|
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user