2026-04-16 13:45:04 +00:00
|
|
|
terraform {
|
|
|
|
|
required_providers {
|
|
|
|
|
cloudflare = {
|
|
|
|
|
source = "cloudflare/cloudflare"
|
|
|
|
|
version = "~> 4"
|
|
|
|
|
}
|
|
|
|
|
kubernetes = {
|
|
|
|
|
source = "hashicorp/kubernetes"
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2026-03-17 21:42:16 +00:00
|
|
|
variable "name" {}
|
|
|
|
|
variable "namespace" {
|
|
|
|
|
default = "reverse-proxy"
|
|
|
|
|
}
|
|
|
|
|
variable "external_name" {}
|
|
|
|
|
variable "port" {
|
|
|
|
|
default = "80"
|
|
|
|
|
}
|
|
|
|
|
variable "tls_secret_name" {}
|
|
|
|
|
variable "backend_protocol" {
|
|
|
|
|
default = "HTTP"
|
|
|
|
|
}
|
|
|
|
|
variable "protected" {
|
|
|
|
|
type = bool
|
|
|
|
|
default = true
|
|
|
|
|
}
|
|
|
|
|
variable "ingress_path" {
|
|
|
|
|
type = list(string)
|
|
|
|
|
default = ["/"]
|
|
|
|
|
}
|
|
|
|
|
variable "max_body_size" {
|
|
|
|
|
type = string
|
|
|
|
|
default = "50m"
|
|
|
|
|
}
|
|
|
|
|
variable "extra_annotations" {
|
|
|
|
|
default = {}
|
|
|
|
|
}
|
|
|
|
|
variable "custom_content_security_policy" {
|
|
|
|
|
default = null
|
|
|
|
|
type = string
|
|
|
|
|
}
|
|
|
|
|
variable "strip_auth_headers" {
|
|
|
|
|
type = bool
|
|
|
|
|
default = false
|
|
|
|
|
}
|
|
|
|
|
variable "extra_middlewares" {
|
|
|
|
|
type = list(string)
|
|
|
|
|
default = []
|
|
|
|
|
}
|
2026-04-05 20:47:58 +03:00
|
|
|
variable "skip_global_rate_limit" {
|
|
|
|
|
type = bool
|
|
|
|
|
default = false
|
|
|
|
|
}
|
2026-04-16 13:45:04 +00:00
|
|
|
variable "dns_type" {
|
|
|
|
|
type = string
|
|
|
|
|
default = "none"
|
|
|
|
|
description = "Cloudflare DNS: 'proxied' (CNAME to tunnel), 'non-proxied' (A/AAAA to public IP), or 'none'"
|
|
|
|
|
validation {
|
|
|
|
|
condition = contains(["proxied", "non-proxied", "none"], var.dns_type)
|
|
|
|
|
error_message = "dns_type must be 'proxied', 'non-proxied', or 'none'."
|
|
|
|
|
}
|
|
|
|
|
}
|
[infra] Per-ingress external-monitor annotation + actualbudget plan-time fix [ci skip]
## Context
Two operational gaps surfaced during a healthcheck sweep today:
1. **External monitoring coverage**: Only ~13 hostnames (via `cloudflare_proxied_names`
in `config.tfvars`) had `[External]` monitors in Uptime Kuma. Any service deployed via
`ingress_factory` with `dns_type = "proxied"` auto-created its DNS record but was NOT
registered for external probing — so outages like Immich going down externally were
invisible until a user complained. 99 of ~125 public ingresses had no external
monitor.
2. **actualbudget stack unplannable**: `count = var.budget_encryption_password != null
? 1 : 0` in `factory/main.tf:152` failed with "Invalid count argument" because the
value flows from a `data.kubernetes_secret` whose contents are `(known after apply)`
at plan time. Blocked CI applies and drift reconciliation.
## This change
### Per-ingress external-monitor annotation (ingress_factory + reverse_proxy/factory)
- New variables `external_monitor` (bool, nullable) + `external_monitor_name` (string,
nullable). Default is "follow dns_type" — enabled for any public DNS record
(`dns_type != "none"`, covers both proxied and non-proxied so Immich and other
direct-A records are also monitored).
- Emits two annotations on the Ingress:
- `uptime.viktorbarzin.me/external-monitor = "true"`
- `uptime.viktorbarzin.me/external-monitor-name = "<label>"` (optional override)
### external-monitor-sync CronJob (uptime-kuma stack)
- Discovers targets from live Ingress objects via the K8s API first (filter by
annotation), falls back to the legacy `external-monitor-targets` ConfigMap on any
API error (zero rollout risk).
- New `ServiceAccount` + cluster-wide `ClusterRole`/`ClusterRoleBinding` giving
`list`/`get` on `networking.k8s.io/ingresses`.
- `API_SERVER` now uses the `KUBERNETES_SERVICE_HOST` env var (always injected by K8s)
instead of `kubernetes.default.svc` — the search-domain expansion failed in the
CronJob pod's DNS config. Verified working: CronJob now logs
`Loaded N external monitor targets (source=k8s-api)`.
### actualbudget count-on-unknown refactor
- Replaced `count = var.budget_encryption_password != null ? 1 : 0` with two explicit
plan-time booleans: `enable_http_api` and `enable_bank_sync`. Values are known at
plan; no `-target` workaround needed.
- Callers (`stacks/actualbudget/main.tf`) pass `true` explicitly. Runtime behaviour is
unchanged — the secret is still consumed via env var.
- Also aligned the factory with live state (the 3 budget-* PVCs had been migrated
`proxmox-lvm` → `proxmox-lvm-encrypted` outside Terraform): PVC resource renamed
`data_proxmox` → `data_encrypted`, storage class updated, orphaned `nfs_data` module
removed. State was rm'd + re-imported with matching UIDs, so no data was moved.
## Rollout status (already partially applied in this session)
- `stacks/uptime-kuma` applied — SA + RBAC + CronJob changes live; FQDN fix verified
- `stacks/actualbudget` applied — budget-{viktor,anca,emo} all 200 OK externally
- `stacks/mailserver` + 21 other ingress_factory consumers applied — annotations live
- CronJob `external-monitor-sync` latest run: `source=k8s-api`, 26 monitors active
(was 13 on the central list)
## Deferred (separate work)
- 4 stacks show pre-existing DESTRUCTIVE drift in plan (metallb namespace, claude-memory,
rbac, redis) — NOT triggered by this commit but will be by CI's global-file cascade.
`[ci skip]` here so those don't auto-apply; they will be fixed manually before the
next CI push.
- Cleanup of `cloudflare_proxied_names` list once Helm-managed ingresses (authentik,
grafana, vault, forgejo) are annotated — separate PR.
## Test plan
### Automated
\`\`\`
\$ kubectl -n uptime-kuma logs \$(kubectl -n uptime-kuma get pods -l job-name -o name | tail -1)
Loaded 26 external monitor targets (source=k8s-api)
Sync complete: 7 created, 0 deleted, 17 unchanged
\$ curl -sk -o /dev/null -w "%{http_code}\n" -H "Accept: text/html" \\
https://dawarich.viktorbarzin.me/ https://nextcloud.viktorbarzin.me/ \\
https://budget-viktor.viktorbarzin.me/
200 302 200
\$ kubectl -n actualbudget get deploy,pvc -l app=budget-viktor
deployment.apps/budget-viktor 1/1 1 1 Ready
persistentvolumeclaim/budget-viktor-data-encrypted Bound 10Gi RWO proxmox-lvm-encrypted
\`\`\`
### Manual Verification
1. Confirm the annotation is present on an ingress_factory ingress:
\`\`\`
kubectl -n dawarich get ingress dawarich -o \\
jsonpath='{.metadata.annotations.uptime\.viktorbarzin\.me/external-monitor}'
# Expected: "true"
\`\`\`
2. Confirm the new `[External] <name>` monitor appears in Uptime Kuma within 10 min
(CronJob interval). For Immich specifically, it will appear after the immich stack
is re-applied.
3. Verify actualbudget plan is clean:
\`\`\`
cd stacks/actualbudget && scripts/tg plan --non-interactive
# Expected: no "Invalid count argument" errors
\`\`\`
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-17 10:34:32 +00:00
|
|
|
|
|
|
|
|
# Uptime Kuma external monitor: when true, annotate the ingress so the
|
|
|
|
|
# external-monitor-sync CronJob creates a `[External] <name>` monitor pointing
|
|
|
|
|
# at https://<host>. Null means "follow dns_type" — enabled when proxied.
|
|
|
|
|
variable "external_monitor" {
|
|
|
|
|
type = bool
|
|
|
|
|
default = null
|
|
|
|
|
description = "Enable Uptime Kuma external monitor. null = auto (enabled when dns_type == 'proxied')."
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
variable "external_monitor_name" {
|
|
|
|
|
type = string
|
|
|
|
|
default = null
|
|
|
|
|
description = "Override the monitor label. Defaults to the ingress hostname label."
|
|
|
|
|
}
|
2026-04-16 13:45:04 +00:00
|
|
|
variable "cloudflare_zone_id" {
|
|
|
|
|
type = string
|
|
|
|
|
default = "fd2c5dd4efe8fe38958944e74d0ced6d"
|
|
|
|
|
}
|
|
|
|
|
variable "cloudflare_tunnel_id" {
|
|
|
|
|
type = string
|
|
|
|
|
default = "75182cd7-bb91-4310-b961-5d8967da8b41"
|
|
|
|
|
}
|
|
|
|
|
variable "public_ip" {
|
|
|
|
|
type = string
|
|
|
|
|
default = "176.12.22.76"
|
|
|
|
|
}
|
|
|
|
|
variable "public_ipv6" {
|
|
|
|
|
type = string
|
|
|
|
|
default = "2001:470:6e:43d::2"
|
|
|
|
|
}
|
2026-03-17 21:42:16 +00:00
|
|
|
|
|
|
|
|
|
|
|
|
|
resource "kubernetes_service" "proxied-service" {
|
|
|
|
|
metadata {
|
|
|
|
|
name = var.name
|
|
|
|
|
namespace = var.namespace
|
|
|
|
|
labels = {
|
|
|
|
|
"app" = var.name
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
spec {
|
|
|
|
|
type = "ExternalName"
|
|
|
|
|
external_name = var.external_name
|
|
|
|
|
|
|
|
|
|
port {
|
|
|
|
|
name = var.backend_protocol == "HTTPS" ? "https-${var.name}" : "${var.name}-web"
|
|
|
|
|
port = var.port
|
|
|
|
|
protocol = "TCP"
|
|
|
|
|
target_port = var.port
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
[infra] Per-ingress external-monitor annotation + actualbudget plan-time fix [ci skip]
## Context
Two operational gaps surfaced during a healthcheck sweep today:
1. **External monitoring coverage**: Only ~13 hostnames (via `cloudflare_proxied_names`
in `config.tfvars`) had `[External]` monitors in Uptime Kuma. Any service deployed via
`ingress_factory` with `dns_type = "proxied"` auto-created its DNS record but was NOT
registered for external probing — so outages like Immich going down externally were
invisible until a user complained. 99 of ~125 public ingresses had no external
monitor.
2. **actualbudget stack unplannable**: `count = var.budget_encryption_password != null
? 1 : 0` in `factory/main.tf:152` failed with "Invalid count argument" because the
value flows from a `data.kubernetes_secret` whose contents are `(known after apply)`
at plan time. Blocked CI applies and drift reconciliation.
## This change
### Per-ingress external-monitor annotation (ingress_factory + reverse_proxy/factory)
- New variables `external_monitor` (bool, nullable) + `external_monitor_name` (string,
nullable). Default is "follow dns_type" — enabled for any public DNS record
(`dns_type != "none"`, covers both proxied and non-proxied so Immich and other
direct-A records are also monitored).
- Emits two annotations on the Ingress:
- `uptime.viktorbarzin.me/external-monitor = "true"`
- `uptime.viktorbarzin.me/external-monitor-name = "<label>"` (optional override)
### external-monitor-sync CronJob (uptime-kuma stack)
- Discovers targets from live Ingress objects via the K8s API first (filter by
annotation), falls back to the legacy `external-monitor-targets` ConfigMap on any
API error (zero rollout risk).
- New `ServiceAccount` + cluster-wide `ClusterRole`/`ClusterRoleBinding` giving
`list`/`get` on `networking.k8s.io/ingresses`.
- `API_SERVER` now uses the `KUBERNETES_SERVICE_HOST` env var (always injected by K8s)
instead of `kubernetes.default.svc` — the search-domain expansion failed in the
CronJob pod's DNS config. Verified working: CronJob now logs
`Loaded N external monitor targets (source=k8s-api)`.
### actualbudget count-on-unknown refactor
- Replaced `count = var.budget_encryption_password != null ? 1 : 0` with two explicit
plan-time booleans: `enable_http_api` and `enable_bank_sync`. Values are known at
plan; no `-target` workaround needed.
- Callers (`stacks/actualbudget/main.tf`) pass `true` explicitly. Runtime behaviour is
unchanged — the secret is still consumed via env var.
- Also aligned the factory with live state (the 3 budget-* PVCs had been migrated
`proxmox-lvm` → `proxmox-lvm-encrypted` outside Terraform): PVC resource renamed
`data_proxmox` → `data_encrypted`, storage class updated, orphaned `nfs_data` module
removed. State was rm'd + re-imported with matching UIDs, so no data was moved.
## Rollout status (already partially applied in this session)
- `stacks/uptime-kuma` applied — SA + RBAC + CronJob changes live; FQDN fix verified
- `stacks/actualbudget` applied — budget-{viktor,anca,emo} all 200 OK externally
- `stacks/mailserver` + 21 other ingress_factory consumers applied — annotations live
- CronJob `external-monitor-sync` latest run: `source=k8s-api`, 26 monitors active
(was 13 on the central list)
## Deferred (separate work)
- 4 stacks show pre-existing DESTRUCTIVE drift in plan (metallb namespace, claude-memory,
rbac, redis) — NOT triggered by this commit but will be by CI's global-file cascade.
`[ci skip]` here so those don't auto-apply; they will be fixed manually before the
next CI push.
- Cleanup of `cloudflare_proxied_names` list once Helm-managed ingresses (authentik,
grafana, vault, forgejo) are annotated — separate PR.
## Test plan
### Automated
\`\`\`
\$ kubectl -n uptime-kuma logs \$(kubectl -n uptime-kuma get pods -l job-name -o name | tail -1)
Loaded 26 external monitor targets (source=k8s-api)
Sync complete: 7 created, 0 deleted, 17 unchanged
\$ curl -sk -o /dev/null -w "%{http_code}\n" -H "Accept: text/html" \\
https://dawarich.viktorbarzin.me/ https://nextcloud.viktorbarzin.me/ \\
https://budget-viktor.viktorbarzin.me/
200 302 200
\$ kubectl -n actualbudget get deploy,pvc -l app=budget-viktor
deployment.apps/budget-viktor 1/1 1 1 Ready
persistentvolumeclaim/budget-viktor-data-encrypted Bound 10Gi RWO proxmox-lvm-encrypted
\`\`\`
### Manual Verification
1. Confirm the annotation is present on an ingress_factory ingress:
\`\`\`
kubectl -n dawarich get ingress dawarich -o \\
jsonpath='{.metadata.annotations.uptime\.viktorbarzin\.me/external-monitor}'
# Expected: "true"
\`\`\`
2. Confirm the new `[External] <name>` monitor appears in Uptime Kuma within 10 min
(CronJob interval). For Immich specifically, it will appear after the immich stack
is re-applied.
3. Verify actualbudget plan is clean:
\`\`\`
cd stacks/actualbudget && scripts/tg plan --non-interactive
# Expected: no "Invalid count argument" errors
\`\`\`
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-17 10:34:32 +00:00
|
|
|
locals {
|
|
|
|
|
# External monitor defaults: on when proxied, off otherwise. Explicit bool overrides.
|
|
|
|
|
effective_external_monitor = var.external_monitor != null ? var.external_monitor : (var.dns_type == "proxied")
|
|
|
|
|
|
|
|
|
|
external_monitor_annotations = local.effective_external_monitor ? merge(
|
|
|
|
|
{ "uptime.viktorbarzin.me/external-monitor" = "true" },
|
|
|
|
|
var.external_monitor_name != null ? { "uptime.viktorbarzin.me/external-monitor-name" = var.external_monitor_name } : {},
|
|
|
|
|
) : {}
|
|
|
|
|
}
|
|
|
|
|
|
2026-03-17 21:42:16 +00:00
|
|
|
resource "kubernetes_ingress_v1" "proxied-ingress" {
|
|
|
|
|
metadata {
|
|
|
|
|
name = var.name
|
|
|
|
|
namespace = var.namespace
|
|
|
|
|
annotations = merge({
|
|
|
|
|
"traefik.ingress.kubernetes.io/router.middlewares" = join(",", compact(concat([
|
2026-04-05 20:47:58 +03:00
|
|
|
"traefik-retry@kubernetescrd",
|
|
|
|
|
var.skip_global_rate_limit ? null : "traefik-rate-limit@kubernetescrd",
|
2026-03-17 21:42:16 +00:00
|
|
|
var.custom_content_security_policy == null ? "traefik-csp-headers@kubernetescrd" : null,
|
|
|
|
|
"traefik-crowdsec@kubernetescrd",
|
|
|
|
|
var.protected ? "traefik-authentik-forward-auth@kubernetescrd" : null,
|
|
|
|
|
var.strip_auth_headers ? "traefik-strip-auth-headers@kubernetescrd" : null,
|
|
|
|
|
var.custom_content_security_policy != null ? "${var.namespace}-custom-csp-${var.name}@kubernetescrd" : null,
|
|
|
|
|
], var.extra_middlewares)))
|
|
|
|
|
"traefik.ingress.kubernetes.io/router.entrypoints" = "websecure"
|
|
|
|
|
"traefik.ingress.kubernetes.io/service.serversscheme" = var.backend_protocol == "HTTPS" ? "https" : null
|
|
|
|
|
"traefik.ingress.kubernetes.io/service.serverstransport" = var.backend_protocol == "HTTPS" ? "traefik-insecure-skip-verify@kubernetescrd" : null
|
[infra] Per-ingress external-monitor annotation + actualbudget plan-time fix [ci skip]
## Context
Two operational gaps surfaced during a healthcheck sweep today:
1. **External monitoring coverage**: Only ~13 hostnames (via `cloudflare_proxied_names`
in `config.tfvars`) had `[External]` monitors in Uptime Kuma. Any service deployed via
`ingress_factory` with `dns_type = "proxied"` auto-created its DNS record but was NOT
registered for external probing — so outages like Immich going down externally were
invisible until a user complained. 99 of ~125 public ingresses had no external
monitor.
2. **actualbudget stack unplannable**: `count = var.budget_encryption_password != null
? 1 : 0` in `factory/main.tf:152` failed with "Invalid count argument" because the
value flows from a `data.kubernetes_secret` whose contents are `(known after apply)`
at plan time. Blocked CI applies and drift reconciliation.
## This change
### Per-ingress external-monitor annotation (ingress_factory + reverse_proxy/factory)
- New variables `external_monitor` (bool, nullable) + `external_monitor_name` (string,
nullable). Default is "follow dns_type" — enabled for any public DNS record
(`dns_type != "none"`, covers both proxied and non-proxied so Immich and other
direct-A records are also monitored).
- Emits two annotations on the Ingress:
- `uptime.viktorbarzin.me/external-monitor = "true"`
- `uptime.viktorbarzin.me/external-monitor-name = "<label>"` (optional override)
### external-monitor-sync CronJob (uptime-kuma stack)
- Discovers targets from live Ingress objects via the K8s API first (filter by
annotation), falls back to the legacy `external-monitor-targets` ConfigMap on any
API error (zero rollout risk).
- New `ServiceAccount` + cluster-wide `ClusterRole`/`ClusterRoleBinding` giving
`list`/`get` on `networking.k8s.io/ingresses`.
- `API_SERVER` now uses the `KUBERNETES_SERVICE_HOST` env var (always injected by K8s)
instead of `kubernetes.default.svc` — the search-domain expansion failed in the
CronJob pod's DNS config. Verified working: CronJob now logs
`Loaded N external monitor targets (source=k8s-api)`.
### actualbudget count-on-unknown refactor
- Replaced `count = var.budget_encryption_password != null ? 1 : 0` with two explicit
plan-time booleans: `enable_http_api` and `enable_bank_sync`. Values are known at
plan; no `-target` workaround needed.
- Callers (`stacks/actualbudget/main.tf`) pass `true` explicitly. Runtime behaviour is
unchanged — the secret is still consumed via env var.
- Also aligned the factory with live state (the 3 budget-* PVCs had been migrated
`proxmox-lvm` → `proxmox-lvm-encrypted` outside Terraform): PVC resource renamed
`data_proxmox` → `data_encrypted`, storage class updated, orphaned `nfs_data` module
removed. State was rm'd + re-imported with matching UIDs, so no data was moved.
## Rollout status (already partially applied in this session)
- `stacks/uptime-kuma` applied — SA + RBAC + CronJob changes live; FQDN fix verified
- `stacks/actualbudget` applied — budget-{viktor,anca,emo} all 200 OK externally
- `stacks/mailserver` + 21 other ingress_factory consumers applied — annotations live
- CronJob `external-monitor-sync` latest run: `source=k8s-api`, 26 monitors active
(was 13 on the central list)
## Deferred (separate work)
- 4 stacks show pre-existing DESTRUCTIVE drift in plan (metallb namespace, claude-memory,
rbac, redis) — NOT triggered by this commit but will be by CI's global-file cascade.
`[ci skip]` here so those don't auto-apply; they will be fixed manually before the
next CI push.
- Cleanup of `cloudflare_proxied_names` list once Helm-managed ingresses (authentik,
grafana, vault, forgejo) are annotated — separate PR.
## Test plan
### Automated
\`\`\`
\$ kubectl -n uptime-kuma logs \$(kubectl -n uptime-kuma get pods -l job-name -o name | tail -1)
Loaded 26 external monitor targets (source=k8s-api)
Sync complete: 7 created, 0 deleted, 17 unchanged
\$ curl -sk -o /dev/null -w "%{http_code}\n" -H "Accept: text/html" \\
https://dawarich.viktorbarzin.me/ https://nextcloud.viktorbarzin.me/ \\
https://budget-viktor.viktorbarzin.me/
200 302 200
\$ kubectl -n actualbudget get deploy,pvc -l app=budget-viktor
deployment.apps/budget-viktor 1/1 1 1 Ready
persistentvolumeclaim/budget-viktor-data-encrypted Bound 10Gi RWO proxmox-lvm-encrypted
\`\`\`
### Manual Verification
1. Confirm the annotation is present on an ingress_factory ingress:
\`\`\`
kubectl -n dawarich get ingress dawarich -o \\
jsonpath='{.metadata.annotations.uptime\.viktorbarzin\.me/external-monitor}'
# Expected: "true"
\`\`\`
2. Confirm the new `[External] <name>` monitor appears in Uptime Kuma within 10 min
(CronJob interval). For Immich specifically, it will appear after the immich stack
is re-applied.
3. Verify actualbudget plan is clean:
\`\`\`
cd stacks/actualbudget && scripts/tg plan --non-interactive
# Expected: no "Invalid count argument" errors
\`\`\`
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-04-17 10:34:32 +00:00
|
|
|
}, var.extra_annotations,
|
|
|
|
|
var.dns_type != "none" ? { "cloudflare.viktorbarzin.me/dns-type" = var.dns_type } : {},
|
|
|
|
|
local.external_monitor_annotations,
|
2026-04-16 13:45:04 +00:00
|
|
|
)
|
2026-03-17 21:42:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
spec {
|
|
|
|
|
ingress_class_name = "traefik"
|
|
|
|
|
tls {
|
|
|
|
|
hosts = ["${var.name}.viktorbarzin.me"]
|
|
|
|
|
secret_name = var.tls_secret_name
|
|
|
|
|
}
|
|
|
|
|
rule {
|
|
|
|
|
host = "${var.name}.viktorbarzin.me"
|
|
|
|
|
http {
|
|
|
|
|
dynamic "path" {
|
|
|
|
|
for_each = var.ingress_path
|
|
|
|
|
|
|
|
|
|
content {
|
|
|
|
|
path = path.value
|
|
|
|
|
backend {
|
|
|
|
|
service {
|
|
|
|
|
|
|
|
|
|
name = var.name
|
|
|
|
|
port {
|
|
|
|
|
number = var.port
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Custom CSP headers middleware - created per service when custom_content_security_policy is set
|
|
|
|
|
resource "kubernetes_manifest" "custom_csp" {
|
|
|
|
|
count = var.custom_content_security_policy != null ? 1 : 0
|
|
|
|
|
|
|
|
|
|
manifest = {
|
|
|
|
|
apiVersion = "traefik.io/v1alpha1"
|
|
|
|
|
kind = "Middleware"
|
|
|
|
|
metadata = {
|
|
|
|
|
name = "custom-csp-${var.name}"
|
|
|
|
|
namespace = var.namespace
|
|
|
|
|
}
|
|
|
|
|
spec = {
|
|
|
|
|
headers = {
|
|
|
|
|
contentSecurityPolicy = var.custom_content_security_policy
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
2026-04-16 13:45:04 +00:00
|
|
|
|
|
|
|
|
# Cloudflare DNS records — created automatically when dns_type is set.
|
|
|
|
|
resource "cloudflare_record" "proxied" {
|
|
|
|
|
count = var.dns_type == "proxied" ? 1 : 0
|
|
|
|
|
name = var.name
|
|
|
|
|
content = "${var.cloudflare_tunnel_id}.cfargotunnel.com"
|
|
|
|
|
proxied = true
|
|
|
|
|
ttl = 1
|
|
|
|
|
type = "CNAME"
|
|
|
|
|
zone_id = var.cloudflare_zone_id
|
|
|
|
|
allow_overwrite = true
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
resource "cloudflare_record" "non_proxied_a" {
|
|
|
|
|
count = var.dns_type == "non-proxied" ? 1 : 0
|
|
|
|
|
name = var.name
|
|
|
|
|
content = var.public_ip
|
|
|
|
|
proxied = false
|
|
|
|
|
ttl = 1
|
|
|
|
|
type = "A"
|
|
|
|
|
zone_id = var.cloudflare_zone_id
|
|
|
|
|
allow_overwrite = true
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
resource "cloudflare_record" "non_proxied_aaaa" {
|
|
|
|
|
count = var.dns_type == "non-proxied" ? 1 : 0
|
|
|
|
|
name = var.name
|
|
|
|
|
content = var.public_ipv6
|
|
|
|
|
proxied = false
|
|
|
|
|
ttl = 1
|
|
|
|
|
type = "AAAA"
|
|
|
|
|
zone_id = var.cloudflare_zone_id
|
|
|
|
|
allow_overwrite = true
|
|
|
|
|
}
|