Compare commits
2 Commits
2c39db104e
...
fix/argocd
| Author | SHA1 | Date | |
|---|---|---|---|
| 3d9d01c240 | |||
| b290d39d64 |
@@ -211,17 +211,28 @@ jobs:
|
||||
echo "ERROR: Found $ctype \"$cname\" image repo is \"$new_repo\" but expected \"$expected_repo\""
|
||||
exit 1
|
||||
fi
|
||||
if [ -n "${old_image:-}" ]; then
|
||||
old_tag="${old_image##*:}"
|
||||
else
|
||||
old_tag=""
|
||||
fi
|
||||
|
||||
registry="$(echo "$new_repo" | awk -F/ '{print $1}')"
|
||||
|
||||
{
|
||||
echo "changed=$([ "$old_tag" != "$new_tag" ] && echo true || echo false)"
|
||||
echo "new_image=$new_image"
|
||||
echo "new_repo=$new_repo"
|
||||
echo "new_tag=$new_tag"
|
||||
echo "registry=$registry"
|
||||
} >> "$GITHUB_OUTPUT"
|
||||
|
||||
- name: Skip if tag unchanged
|
||||
if: steps.img.outputs.changed != 'true'
|
||||
run: echo "${{ matrix.description }} image tag unchanged; skipping build."
|
||||
|
||||
- name: Check if image exists on registry
|
||||
if: steps.img.outputs.changed == 'true'
|
||||
id: check_image
|
||||
shell: bash
|
||||
run: |
|
||||
@@ -256,15 +267,15 @@ jobs:
|
||||
fi
|
||||
|
||||
- name: Skip if image already exists
|
||||
if: steps.check_image.outputs.exists == 'true'
|
||||
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'true'
|
||||
run: echo "${{ matrix.description }} image ${{ steps.img.outputs.new_image }} already exists on registry; skipping build."
|
||||
|
||||
- name: Set up Buildx
|
||||
if: steps.check_image.outputs.exists == 'false'
|
||||
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'false'
|
||||
uses: docker/setup-buildx-action@v3
|
||||
|
||||
- name: Log in to registry
|
||||
if: steps.check_image.outputs.exists == 'false'
|
||||
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'false'
|
||||
uses: docker/login-action@v3
|
||||
with:
|
||||
registry: ${{ steps.img.outputs.registry }}
|
||||
@@ -272,7 +283,7 @@ jobs:
|
||||
password: ${{ secrets.REGISTRY_PASSWORD }}
|
||||
|
||||
- name: Build and push ${{ matrix.description }} (exact tag from deployment)
|
||||
if: steps.check_image.outputs.exists == 'false'
|
||||
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'false'
|
||||
uses: docker/build-push-action@v6
|
||||
with:
|
||||
context: ${{ matrix.build_context }}
|
||||
|
||||
@@ -1,67 +0,0 @@
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- main
|
||||
- development
|
||||
pull_request:
|
||||
types: [opened, synchronize, reopened]
|
||||
|
||||
name: SonarQube Scan
|
||||
jobs:
|
||||
sonarqube:
|
||||
name: SonarQube Trigger
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checking out
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Set up Python
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: '3.11'
|
||||
|
||||
- name: Install dependencies
|
||||
run: |
|
||||
pip install -r requirements.txt
|
||||
|
||||
- name: Run tests with coverage
|
||||
run: |
|
||||
coverage run --source='.' manage.py test
|
||||
coverage xml
|
||||
|
||||
- name: Set up JDK 17
|
||||
uses: actions/setup-java@v3
|
||||
with:
|
||||
java-version: '17'
|
||||
distribution: 'temurin'
|
||||
|
||||
- name: Cache SonarQube packages
|
||||
uses: actions/cache@v3
|
||||
with:
|
||||
path: ~/.sonar/cache
|
||||
key: ${{ runner.os }}-sonar
|
||||
restore-keys: ${{ runner.os }}-sonar
|
||||
|
||||
- name: Download and setup SonarScanner
|
||||
run: |
|
||||
mkdir -p $HOME/.sonar
|
||||
wget -q https://binaries.sonarsource.com/Distribution/sonar-scanner-cli/sonar-scanner-cli-5.0.1.3006-linux.zip
|
||||
unzip -q sonar-scanner-cli-5.0.1.3006-linux.zip -d $HOME/.sonar/
|
||||
echo "$HOME/.sonar/sonar-scanner-5.0.1.3006-linux/bin" >> $GITHUB_PATH
|
||||
|
||||
- name: Verify Java version
|
||||
run: java -version
|
||||
|
||||
- name: SonarQube Scan
|
||||
env:
|
||||
SONAR_HOST_URL: ${{ secrets.SONARQUBE_HOST }}
|
||||
SONAR_TOKEN: ${{ secrets.SONARQUBE_TOKEN }}
|
||||
run: |
|
||||
sonar-scanner \
|
||||
-Dsonar.projectKey=${{ github.event.repository.name }} \
|
||||
-Dsonar.sources=. \
|
||||
-Dsonar.host.url=${SONAR_HOST_URL} \
|
||||
-Dsonar.token=${SONAR_TOKEN} \
|
||||
-Dsonar.python.coverage.reportPaths=coverage.xml
|
||||
13
Dockerfile
13
Dockerfile
@@ -1,4 +1,4 @@
|
||||
FROM python:3.14 AS baustelle
|
||||
FROM python:3.13-slim AS baustelle
|
||||
RUN mkdir /app
|
||||
WORKDIR /app
|
||||
ENV PYTHONDONTWRITEBYTECODE=1
|
||||
@@ -7,14 +7,15 @@ RUN pip install --upgrade pip
|
||||
COPY requirements.txt /app/
|
||||
RUN pip install --no-cache-dir -r requirements.txt
|
||||
|
||||
FROM python:3.14-slim
|
||||
FROM python:3.13-slim
|
||||
RUN useradd -m -r appuser && \
|
||||
mkdir /app && \
|
||||
chown -R appuser /app
|
||||
|
||||
COPY --from=baustelle /usr/local/lib/python3.14/site-packages/ /usr/local/lib/python3.14/site-packages/
|
||||
COPY --from=baustelle /usr/local/lib/python3.13/site-packages/ /usr/local/lib/python3.13/site-packages/
|
||||
COPY --from=baustelle /usr/local/bin/ /usr/local/bin/
|
||||
RUN rm /usr/bin/tar /usr/lib/x86_64-linux-gnu/libncur*
|
||||
RUN rm /usr/bin/tar
|
||||
RUN rm /usr/lib/x86_64-linux-gnu/libncur*
|
||||
WORKDIR /app
|
||||
COPY --chown=appuser:appuser . .
|
||||
ENV PYTHONDONTWRITEBYTECODE=1
|
||||
@@ -30,7 +31,7 @@ RUN rm -rf /app/Dockerfile* \
|
||||
/app/requirements.txt \
|
||||
/app/node_modules \
|
||||
/app/*.json \
|
||||
/app/test_*.py && \
|
||||
python3 manage.py collectstatic
|
||||
/app/test_*.py
|
||||
RUN python3 manage.py collectstatic
|
||||
CMD ["gunicorn","--bind","0.0.0.0:8000","--workers","3","VorgabenUI.wsgi:application"]
|
||||
|
||||
|
||||
95
Documentation/ARGOCD-INGRESS-FIX.md
Normal file
95
Documentation/ARGOCD-INGRESS-FIX.md
Normal file
@@ -0,0 +1,95 @@
|
||||
# ArgoCD Ingress "Progressing" State Fix
|
||||
|
||||
## Problem
|
||||
|
||||
The `django` Ingress resource in the `vorgabenui` namespace was stuck in "Progressing" state in ArgoCD and would not transition to "Healthy".
|
||||
|
||||
### Root Cause
|
||||
|
||||
ArgoCD determines Ingress health by checking if the `status.loadBalancer.ingress` field is populated with an IP address or hostname. Without this field, the Ingress is considered "Progressing" indefinitely.
|
||||
|
||||
The issue occurred because **Traefik was not configured to report its IP address** in the Ingress status field.
|
||||
|
||||
## Solution
|
||||
|
||||
Two changes were made to fix this issue:
|
||||
|
||||
### 1. Update Ingress Annotation (Applied)
|
||||
|
||||
**File**: `argocd/ingress.yaml`
|
||||
|
||||
**Change**:
|
||||
```yaml
|
||||
# Before
|
||||
annotations:
|
||||
argocd.argoproj.io/ignore-healthcheck: "true"
|
||||
|
||||
# After
|
||||
annotations:
|
||||
argocd.argoproj.io/sync-wave: "1"
|
||||
```
|
||||
|
||||
**Rationale**:
|
||||
- The `ignore-healthcheck` annotation was causing ArgoCD to not monitor the Ingress health at all
|
||||
- The `sync-wave: "1"` annotation ensures the Ingress syncs after the Deployment and Service are ready (which have default sync-wave of 0)
|
||||
- This allows ArgoCD to properly assess the Ingress health status
|
||||
|
||||
### 2. Configure Traefik to Report Ingress Status (Cluster Patch)
|
||||
|
||||
**Patch Command**:
|
||||
```bash
|
||||
kubectl patch deployment traefik -n traefik --type='json' \
|
||||
-p='[{"op": "add", "path": "/spec/template/spec/containers/0/args/-", "value": "--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik"}]'
|
||||
```
|
||||
|
||||
**Configuration Flag Added**:
|
||||
```
|
||||
--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik
|
||||
```
|
||||
|
||||
**Rationale**:
|
||||
This flag tells Traefik to:
|
||||
- Watch for changes to Ingress resources in the cluster
|
||||
- Monitor the Service `traefik/traefik` (the Traefik LoadBalancer service)
|
||||
- Automatically populate `status.loadBalancer.ingress[].ip` with the service's external IP address
|
||||
- Allow ArgoCD to detect when the Ingress has been assigned an IP and transition to "Healthy"
|
||||
|
||||
## Result
|
||||
|
||||
✅ **Status**: RESOLVED
|
||||
|
||||
**Current State**:
|
||||
- Ingress Address: `192.168.17.53` (Traefik LoadBalancer IP)
|
||||
- Ingress Health: Healthy
|
||||
- ArgoCD Application Health: Healthy
|
||||
- Accessible at: `http://vorgabenportal.knowyoursecurity.com/`
|
||||
|
||||
## Verification
|
||||
|
||||
To verify the fix is working:
|
||||
|
||||
```bash
|
||||
# Check Ingress status
|
||||
kubectl get ingress django -n vorgabenui -o jsonpath='{.status.loadBalancer.ingress[0].ip}'
|
||||
# Should output: 192.168.17.53
|
||||
|
||||
# Check ArgoCD application health
|
||||
kubectl get application vorgabenui -n argocd -o jsonpath='{.status.health.status}'
|
||||
# Should output: Healthy
|
||||
|
||||
# Check Traefik configuration
|
||||
kubectl get deploy traefik -n traefik -o jsonpath='{.spec.template.spec.containers[0].args}' | jq 'map(select(. | contains("publishedservice")))'
|
||||
# Should output the publishedservice flag
|
||||
```
|
||||
|
||||
## Documentation Location
|
||||
|
||||
The Traefik configuration patch is documented in:
|
||||
- `argocd/traefik-middleware.yaml` - ConfigMap with patch details and rationale
|
||||
|
||||
## Notes for Future Maintenance
|
||||
|
||||
- If Traefik is upgraded or redeployed via Helm, ensure the `--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik` flag is preserved
|
||||
- The flag must point to the correct LoadBalancer Service that has an external IP
|
||||
- In this case, it's `traefik/traefik` (namespace/service-name) with external IP `192.168.17.53`
|
||||
- If the Traefik service configuration changes, this flag may need adjustment
|
||||
@@ -24,7 +24,7 @@ BASE_DIR = Path(__file__).resolve().parent.parent
|
||||
SECRET_KEY = os.environ.get("SECRET_KEY")
|
||||
|
||||
# SECURITY WARNING: don't run with debug turned on in production!
|
||||
DEBUG = bool(os.environ.get("DEBUG", default=0))
|
||||
DEBUG = bool(os.environ.get("DEBUG", default=0)
|
||||
|
||||
ALLOWED_HOSTS = os.environ.get("DJANGO_ALLOWED_HOSTS","127.0.0.1").split(",")
|
||||
|
||||
|
||||
@@ -25,7 +25,7 @@ spec:
|
||||
mountPath: /data
|
||||
containers:
|
||||
- name: web
|
||||
image: git.baumann.gr/adebaumann/vui:0.963
|
||||
image: git.baumann.gr/adebaumann/vui:0.961
|
||||
imagePullPolicy: Always
|
||||
ports:
|
||||
- containerPort: 8000
|
||||
|
||||
@@ -4,7 +4,7 @@ metadata:
|
||||
name: django
|
||||
namespace: vorgabenui
|
||||
annotations:
|
||||
argocd.argoproj.io/ignore-healthcheck: "true"
|
||||
argocd.argoproj.io/sync-wave: "1"
|
||||
spec:
|
||||
ingressClassName: traefik
|
||||
rules:
|
||||
|
||||
24
argocd/traefik-middleware.yaml
Normal file
24
argocd/traefik-middleware.yaml
Normal file
@@ -0,0 +1,24 @@
|
||||
---
|
||||
# Traefik configuration to enable Ingress status updates
|
||||
# This patch configures Traefik to report its IP address in Ingress.status.loadBalancer
|
||||
# which is required for ArgoCD to properly assess Ingress health status
|
||||
apiVersion: v1
|
||||
kind: ConfigMap
|
||||
metadata:
|
||||
name: traefik-patch-note
|
||||
namespace: traefik
|
||||
annotations:
|
||||
description: "Manual patch applied to traefik deployment to enable ingress status reporting"
|
||||
data:
|
||||
patch-command: |
|
||||
kubectl patch deployment traefik -n traefik --type='json' \
|
||||
-p='[{"op": "add", "path": "/spec/template/spec/containers/0/args/-", "value": "--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik"}]'
|
||||
|
||||
rationale: |
|
||||
The Ingress resource needs its status.loadBalancer.ingress field populated for ArgoCD to assess health.
|
||||
Without this, Ingress resources remain in "Progressing" state indefinitely.
|
||||
|
||||
This flag tells Traefik to:
|
||||
- Monitor the specified Service (traefik/traefik - the LoadBalancer service)
|
||||
- Automatically update Ingress.status.loadBalancer with the service's external IP
|
||||
- Allow ArgoCD to transition the Ingress from "Progressing" to "Healthy"
|
||||
@@ -1620,25 +1620,19 @@ class GetVorgabeCommentsViewTest(TestCase):
|
||||
# Create users
|
||||
self.regular_user = User.objects.create_user(
|
||||
username='regularuser',
|
||||
password='testpass123',
|
||||
first_name='Regular',
|
||||
last_name='User'
|
||||
password='testpass123'
|
||||
)
|
||||
|
||||
self.staff_user = User.objects.create_user(
|
||||
username='staffuser',
|
||||
password='testpass123',
|
||||
first_name='Staff',
|
||||
last_name='User'
|
||||
password='testpass123'
|
||||
)
|
||||
self.staff_user.is_staff = True
|
||||
self.staff_user.save()
|
||||
|
||||
self.other_user = User.objects.create_user(
|
||||
username='otheruser',
|
||||
password='testpass123',
|
||||
first_name='Other',
|
||||
last_name='User'
|
||||
password='testpass123'
|
||||
)
|
||||
|
||||
# Create test data
|
||||
@@ -1703,7 +1697,7 @@ class GetVorgabeCommentsViewTest(TestCase):
|
||||
# Should only see their own comment
|
||||
self.assertEqual(len(data['comments']), 1)
|
||||
self.assertEqual(data['comments'][0]['text'], 'Kommentar von Regular User')
|
||||
self.assertEqual(data['comments'][0]['user'], 'Regular User')
|
||||
self.assertEqual(data['comments'][0]['user'], 'regularuser')
|
||||
self.assertTrue(data['comments'][0]['is_own'])
|
||||
|
||||
def test_staff_user_sees_all_comments(self):
|
||||
@@ -1721,8 +1715,8 @@ class GetVorgabeCommentsViewTest(TestCase):
|
||||
# Should see all comments
|
||||
self.assertEqual(len(data['comments']), 2)
|
||||
usernames = [c['user'] for c in data['comments']]
|
||||
self.assertIn('Regular User', usernames)
|
||||
self.assertIn('Other User', usernames)
|
||||
self.assertIn('regularuser', usernames)
|
||||
self.assertIn('otheruser', usernames)
|
||||
|
||||
def test_get_comments_returns_404_for_nonexistent_vorgabe(self):
|
||||
"""Test that requesting comments for non-existent Vorgabe returns 404"""
|
||||
@@ -2047,16 +2041,12 @@ class DeleteVorgabeCommentViewTest(TestCase):
|
||||
|
||||
self.other_user = User.objects.create_user(
|
||||
username='otheruser',
|
||||
password='testpass123',
|
||||
first_name='Other',
|
||||
last_name='User'
|
||||
password='testpass123'
|
||||
)
|
||||
|
||||
self.staff_user = User.objects.create_user(
|
||||
username='staffuser',
|
||||
password='testpass123',
|
||||
first_name='Staff',
|
||||
last_name='User'
|
||||
password='testpass123'
|
||||
)
|
||||
self.staff_user.is_staff = True
|
||||
self.staff_user.save()
|
||||
|
||||
@@ -215,7 +215,7 @@
|
||||
</p>
|
||||
</div>
|
||||
<div class="col-sm-6 text-right">
|
||||
<p class="text-muted">Version {{ version|default:"0.963" }}</p>
|
||||
<p class="text-muted">Version {{ version|default:"0.961" }}</p>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
|
||||
@@ -5,7 +5,7 @@ certifi==2025.8.3
|
||||
charset-normalizer==3.4.3
|
||||
curtsies==0.4.3
|
||||
cwcwidth==0.1.10
|
||||
Django==5.2.8
|
||||
Django==5.2.5
|
||||
django-admin-sortable2==2.2.8
|
||||
django-js-asset==3.1.2
|
||||
django-mptt==0.17.0
|
||||
@@ -33,4 +33,3 @@ sqlparse==0.5.3
|
||||
urllib3==2.5.0
|
||||
wcwidth==0.2.13
|
||||
bleach==6.1.0
|
||||
coverage==7.6.1
|
||||
|
||||
Reference in New Issue
Block a user