Compare commits
2 Commits
2c39db104e
...
fix/argocd
| Author | SHA1 | Date | |
|---|---|---|---|
| 3d9d01c240 | |||
| b290d39d64 |
@@ -211,17 +211,28 @@ jobs:
|
|||||||
echo "ERROR: Found $ctype \"$cname\" image repo is \"$new_repo\" but expected \"$expected_repo\""
|
echo "ERROR: Found $ctype \"$cname\" image repo is \"$new_repo\" but expected \"$expected_repo\""
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
if [ -n "${old_image:-}" ]; then
|
||||||
|
old_tag="${old_image##*:}"
|
||||||
|
else
|
||||||
|
old_tag=""
|
||||||
|
fi
|
||||||
|
|
||||||
registry="$(echo "$new_repo" | awk -F/ '{print $1}')"
|
registry="$(echo "$new_repo" | awk -F/ '{print $1}')"
|
||||||
|
|
||||||
{
|
{
|
||||||
|
echo "changed=$([ "$old_tag" != "$new_tag" ] && echo true || echo false)"
|
||||||
echo "new_image=$new_image"
|
echo "new_image=$new_image"
|
||||||
echo "new_repo=$new_repo"
|
echo "new_repo=$new_repo"
|
||||||
echo "new_tag=$new_tag"
|
echo "new_tag=$new_tag"
|
||||||
echo "registry=$registry"
|
echo "registry=$registry"
|
||||||
} >> "$GITHUB_OUTPUT"
|
} >> "$GITHUB_OUTPUT"
|
||||||
|
|
||||||
|
- name: Skip if tag unchanged
|
||||||
|
if: steps.img.outputs.changed != 'true'
|
||||||
|
run: echo "${{ matrix.description }} image tag unchanged; skipping build."
|
||||||
|
|
||||||
- name: Check if image exists on registry
|
- name: Check if image exists on registry
|
||||||
|
if: steps.img.outputs.changed == 'true'
|
||||||
id: check_image
|
id: check_image
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
@@ -256,15 +267,15 @@ jobs:
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Skip if image already exists
|
- name: Skip if image already exists
|
||||||
if: steps.check_image.outputs.exists == 'true'
|
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'true'
|
||||||
run: echo "${{ matrix.description }} image ${{ steps.img.outputs.new_image }} already exists on registry; skipping build."
|
run: echo "${{ matrix.description }} image ${{ steps.img.outputs.new_image }} already exists on registry; skipping build."
|
||||||
|
|
||||||
- name: Set up Buildx
|
- name: Set up Buildx
|
||||||
if: steps.check_image.outputs.exists == 'false'
|
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'false'
|
||||||
uses: docker/setup-buildx-action@v3
|
uses: docker/setup-buildx-action@v3
|
||||||
|
|
||||||
- name: Log in to registry
|
- name: Log in to registry
|
||||||
if: steps.check_image.outputs.exists == 'false'
|
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'false'
|
||||||
uses: docker/login-action@v3
|
uses: docker/login-action@v3
|
||||||
with:
|
with:
|
||||||
registry: ${{ steps.img.outputs.registry }}
|
registry: ${{ steps.img.outputs.registry }}
|
||||||
@@ -272,7 +283,7 @@ jobs:
|
|||||||
password: ${{ secrets.REGISTRY_PASSWORD }}
|
password: ${{ secrets.REGISTRY_PASSWORD }}
|
||||||
|
|
||||||
- name: Build and push ${{ matrix.description }} (exact tag from deployment)
|
- name: Build and push ${{ matrix.description }} (exact tag from deployment)
|
||||||
if: steps.check_image.outputs.exists == 'false'
|
if: steps.img.outputs.changed == 'true' && steps.check_image.outputs.exists == 'false'
|
||||||
uses: docker/build-push-action@v6
|
uses: docker/build-push-action@v6
|
||||||
with:
|
with:
|
||||||
context: ${{ matrix.build_context }}
|
context: ${{ matrix.build_context }}
|
||||||
|
|||||||
@@ -1,67 +0,0 @@
|
|||||||
on:
|
|
||||||
push:
|
|
||||||
branches:
|
|
||||||
- main
|
|
||||||
- development
|
|
||||||
pull_request:
|
|
||||||
types: [opened, synchronize, reopened]
|
|
||||||
|
|
||||||
name: SonarQube Scan
|
|
||||||
jobs:
|
|
||||||
sonarqube:
|
|
||||||
name: SonarQube Trigger
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- name: Checking out
|
|
||||||
uses: actions/checkout@v4
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
- name: Set up Python
|
|
||||||
uses: actions/setup-python@v4
|
|
||||||
with:
|
|
||||||
python-version: '3.11'
|
|
||||||
|
|
||||||
- name: Install dependencies
|
|
||||||
run: |
|
|
||||||
pip install -r requirements.txt
|
|
||||||
|
|
||||||
- name: Run tests with coverage
|
|
||||||
run: |
|
|
||||||
coverage run --source='.' manage.py test
|
|
||||||
coverage xml
|
|
||||||
|
|
||||||
- name: Set up JDK 17
|
|
||||||
uses: actions/setup-java@v3
|
|
||||||
with:
|
|
||||||
java-version: '17'
|
|
||||||
distribution: 'temurin'
|
|
||||||
|
|
||||||
- name: Cache SonarQube packages
|
|
||||||
uses: actions/cache@v3
|
|
||||||
with:
|
|
||||||
path: ~/.sonar/cache
|
|
||||||
key: ${{ runner.os }}-sonar
|
|
||||||
restore-keys: ${{ runner.os }}-sonar
|
|
||||||
|
|
||||||
- name: Download and setup SonarScanner
|
|
||||||
run: |
|
|
||||||
mkdir -p $HOME/.sonar
|
|
||||||
wget -q https://binaries.sonarsource.com/Distribution/sonar-scanner-cli/sonar-scanner-cli-5.0.1.3006-linux.zip
|
|
||||||
unzip -q sonar-scanner-cli-5.0.1.3006-linux.zip -d $HOME/.sonar/
|
|
||||||
echo "$HOME/.sonar/sonar-scanner-5.0.1.3006-linux/bin" >> $GITHUB_PATH
|
|
||||||
|
|
||||||
- name: Verify Java version
|
|
||||||
run: java -version
|
|
||||||
|
|
||||||
- name: SonarQube Scan
|
|
||||||
env:
|
|
||||||
SONAR_HOST_URL: ${{ secrets.SONARQUBE_HOST }}
|
|
||||||
SONAR_TOKEN: ${{ secrets.SONARQUBE_TOKEN }}
|
|
||||||
run: |
|
|
||||||
sonar-scanner \
|
|
||||||
-Dsonar.projectKey=${{ github.event.repository.name }} \
|
|
||||||
-Dsonar.sources=. \
|
|
||||||
-Dsonar.host.url=${SONAR_HOST_URL} \
|
|
||||||
-Dsonar.token=${SONAR_TOKEN} \
|
|
||||||
-Dsonar.python.coverage.reportPaths=coverage.xml
|
|
||||||
13
Dockerfile
13
Dockerfile
@@ -1,4 +1,4 @@
|
|||||||
FROM python:3.14 AS baustelle
|
FROM python:3.13-slim AS baustelle
|
||||||
RUN mkdir /app
|
RUN mkdir /app
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
ENV PYTHONDONTWRITEBYTECODE=1
|
ENV PYTHONDONTWRITEBYTECODE=1
|
||||||
@@ -7,14 +7,15 @@ RUN pip install --upgrade pip
|
|||||||
COPY requirements.txt /app/
|
COPY requirements.txt /app/
|
||||||
RUN pip install --no-cache-dir -r requirements.txt
|
RUN pip install --no-cache-dir -r requirements.txt
|
||||||
|
|
||||||
FROM python:3.14-slim
|
FROM python:3.13-slim
|
||||||
RUN useradd -m -r appuser && \
|
RUN useradd -m -r appuser && \
|
||||||
mkdir /app && \
|
mkdir /app && \
|
||||||
chown -R appuser /app
|
chown -R appuser /app
|
||||||
|
|
||||||
COPY --from=baustelle /usr/local/lib/python3.14/site-packages/ /usr/local/lib/python3.14/site-packages/
|
COPY --from=baustelle /usr/local/lib/python3.13/site-packages/ /usr/local/lib/python3.13/site-packages/
|
||||||
COPY --from=baustelle /usr/local/bin/ /usr/local/bin/
|
COPY --from=baustelle /usr/local/bin/ /usr/local/bin/
|
||||||
RUN rm /usr/bin/tar /usr/lib/x86_64-linux-gnu/libncur*
|
RUN rm /usr/bin/tar
|
||||||
|
RUN rm /usr/lib/x86_64-linux-gnu/libncur*
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
COPY --chown=appuser:appuser . .
|
COPY --chown=appuser:appuser . .
|
||||||
ENV PYTHONDONTWRITEBYTECODE=1
|
ENV PYTHONDONTWRITEBYTECODE=1
|
||||||
@@ -30,7 +31,7 @@ RUN rm -rf /app/Dockerfile* \
|
|||||||
/app/requirements.txt \
|
/app/requirements.txt \
|
||||||
/app/node_modules \
|
/app/node_modules \
|
||||||
/app/*.json \
|
/app/*.json \
|
||||||
/app/test_*.py && \
|
/app/test_*.py
|
||||||
python3 manage.py collectstatic
|
RUN python3 manage.py collectstatic
|
||||||
CMD ["gunicorn","--bind","0.0.0.0:8000","--workers","3","VorgabenUI.wsgi:application"]
|
CMD ["gunicorn","--bind","0.0.0.0:8000","--workers","3","VorgabenUI.wsgi:application"]
|
||||||
|
|
||||||
|
|||||||
95
Documentation/ARGOCD-INGRESS-FIX.md
Normal file
95
Documentation/ARGOCD-INGRESS-FIX.md
Normal file
@@ -0,0 +1,95 @@
|
|||||||
|
# ArgoCD Ingress "Progressing" State Fix
|
||||||
|
|
||||||
|
## Problem
|
||||||
|
|
||||||
|
The `django` Ingress resource in the `vorgabenui` namespace was stuck in "Progressing" state in ArgoCD and would not transition to "Healthy".
|
||||||
|
|
||||||
|
### Root Cause
|
||||||
|
|
||||||
|
ArgoCD determines Ingress health by checking if the `status.loadBalancer.ingress` field is populated with an IP address or hostname. Without this field, the Ingress is considered "Progressing" indefinitely.
|
||||||
|
|
||||||
|
The issue occurred because **Traefik was not configured to report its IP address** in the Ingress status field.
|
||||||
|
|
||||||
|
## Solution
|
||||||
|
|
||||||
|
Two changes were made to fix this issue:
|
||||||
|
|
||||||
|
### 1. Update Ingress Annotation (Applied)
|
||||||
|
|
||||||
|
**File**: `argocd/ingress.yaml`
|
||||||
|
|
||||||
|
**Change**:
|
||||||
|
```yaml
|
||||||
|
# Before
|
||||||
|
annotations:
|
||||||
|
argocd.argoproj.io/ignore-healthcheck: "true"
|
||||||
|
|
||||||
|
# After
|
||||||
|
annotations:
|
||||||
|
argocd.argoproj.io/sync-wave: "1"
|
||||||
|
```
|
||||||
|
|
||||||
|
**Rationale**:
|
||||||
|
- The `ignore-healthcheck` annotation was causing ArgoCD to not monitor the Ingress health at all
|
||||||
|
- The `sync-wave: "1"` annotation ensures the Ingress syncs after the Deployment and Service are ready (which have default sync-wave of 0)
|
||||||
|
- This allows ArgoCD to properly assess the Ingress health status
|
||||||
|
|
||||||
|
### 2. Configure Traefik to Report Ingress Status (Cluster Patch)
|
||||||
|
|
||||||
|
**Patch Command**:
|
||||||
|
```bash
|
||||||
|
kubectl patch deployment traefik -n traefik --type='json' \
|
||||||
|
-p='[{"op": "add", "path": "/spec/template/spec/containers/0/args/-", "value": "--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik"}]'
|
||||||
|
```
|
||||||
|
|
||||||
|
**Configuration Flag Added**:
|
||||||
|
```
|
||||||
|
--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik
|
||||||
|
```
|
||||||
|
|
||||||
|
**Rationale**:
|
||||||
|
This flag tells Traefik to:
|
||||||
|
- Watch for changes to Ingress resources in the cluster
|
||||||
|
- Monitor the Service `traefik/traefik` (the Traefik LoadBalancer service)
|
||||||
|
- Automatically populate `status.loadBalancer.ingress[].ip` with the service's external IP address
|
||||||
|
- Allow ArgoCD to detect when the Ingress has been assigned an IP and transition to "Healthy"
|
||||||
|
|
||||||
|
## Result
|
||||||
|
|
||||||
|
✅ **Status**: RESOLVED
|
||||||
|
|
||||||
|
**Current State**:
|
||||||
|
- Ingress Address: `192.168.17.53` (Traefik LoadBalancer IP)
|
||||||
|
- Ingress Health: Healthy
|
||||||
|
- ArgoCD Application Health: Healthy
|
||||||
|
- Accessible at: `http://vorgabenportal.knowyoursecurity.com/`
|
||||||
|
|
||||||
|
## Verification
|
||||||
|
|
||||||
|
To verify the fix is working:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Check Ingress status
|
||||||
|
kubectl get ingress django -n vorgabenui -o jsonpath='{.status.loadBalancer.ingress[0].ip}'
|
||||||
|
# Should output: 192.168.17.53
|
||||||
|
|
||||||
|
# Check ArgoCD application health
|
||||||
|
kubectl get application vorgabenui -n argocd -o jsonpath='{.status.health.status}'
|
||||||
|
# Should output: Healthy
|
||||||
|
|
||||||
|
# Check Traefik configuration
|
||||||
|
kubectl get deploy traefik -n traefik -o jsonpath='{.spec.template.spec.containers[0].args}' | jq 'map(select(. | contains("publishedservice")))'
|
||||||
|
# Should output the publishedservice flag
|
||||||
|
```
|
||||||
|
|
||||||
|
## Documentation Location
|
||||||
|
|
||||||
|
The Traefik configuration patch is documented in:
|
||||||
|
- `argocd/traefik-middleware.yaml` - ConfigMap with patch details and rationale
|
||||||
|
|
||||||
|
## Notes for Future Maintenance
|
||||||
|
|
||||||
|
- If Traefik is upgraded or redeployed via Helm, ensure the `--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik` flag is preserved
|
||||||
|
- The flag must point to the correct LoadBalancer Service that has an external IP
|
||||||
|
- In this case, it's `traefik/traefik` (namespace/service-name) with external IP `192.168.17.53`
|
||||||
|
- If the Traefik service configuration changes, this flag may need adjustment
|
||||||
@@ -24,7 +24,7 @@ BASE_DIR = Path(__file__).resolve().parent.parent
|
|||||||
SECRET_KEY = os.environ.get("SECRET_KEY")
|
SECRET_KEY = os.environ.get("SECRET_KEY")
|
||||||
|
|
||||||
# SECURITY WARNING: don't run with debug turned on in production!
|
# SECURITY WARNING: don't run with debug turned on in production!
|
||||||
DEBUG = bool(os.environ.get("DEBUG", default=0))
|
DEBUG = bool(os.environ.get("DEBUG", default=0)
|
||||||
|
|
||||||
ALLOWED_HOSTS = os.environ.get("DJANGO_ALLOWED_HOSTS","127.0.0.1").split(",")
|
ALLOWED_HOSTS = os.environ.get("DJANGO_ALLOWED_HOSTS","127.0.0.1").split(",")
|
||||||
|
|
||||||
|
|||||||
@@ -25,7 +25,7 @@ spec:
|
|||||||
mountPath: /data
|
mountPath: /data
|
||||||
containers:
|
containers:
|
||||||
- name: web
|
- name: web
|
||||||
image: git.baumann.gr/adebaumann/vui:0.963
|
image: git.baumann.gr/adebaumann/vui:0.961
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
ports:
|
ports:
|
||||||
- containerPort: 8000
|
- containerPort: 8000
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ metadata:
|
|||||||
name: django
|
name: django
|
||||||
namespace: vorgabenui
|
namespace: vorgabenui
|
||||||
annotations:
|
annotations:
|
||||||
argocd.argoproj.io/ignore-healthcheck: "true"
|
argocd.argoproj.io/sync-wave: "1"
|
||||||
spec:
|
spec:
|
||||||
ingressClassName: traefik
|
ingressClassName: traefik
|
||||||
rules:
|
rules:
|
||||||
|
|||||||
24
argocd/traefik-middleware.yaml
Normal file
24
argocd/traefik-middleware.yaml
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
---
|
||||||
|
# Traefik configuration to enable Ingress status updates
|
||||||
|
# This patch configures Traefik to report its IP address in Ingress.status.loadBalancer
|
||||||
|
# which is required for ArgoCD to properly assess Ingress health status
|
||||||
|
apiVersion: v1
|
||||||
|
kind: ConfigMap
|
||||||
|
metadata:
|
||||||
|
name: traefik-patch-note
|
||||||
|
namespace: traefik
|
||||||
|
annotations:
|
||||||
|
description: "Manual patch applied to traefik deployment to enable ingress status reporting"
|
||||||
|
data:
|
||||||
|
patch-command: |
|
||||||
|
kubectl patch deployment traefik -n traefik --type='json' \
|
||||||
|
-p='[{"op": "add", "path": "/spec/template/spec/containers/0/args/-", "value": "--providers.kubernetesingress.ingressendpoint.publishedservice=traefik/traefik"}]'
|
||||||
|
|
||||||
|
rationale: |
|
||||||
|
The Ingress resource needs its status.loadBalancer.ingress field populated for ArgoCD to assess health.
|
||||||
|
Without this, Ingress resources remain in "Progressing" state indefinitely.
|
||||||
|
|
||||||
|
This flag tells Traefik to:
|
||||||
|
- Monitor the specified Service (traefik/traefik - the LoadBalancer service)
|
||||||
|
- Automatically update Ingress.status.loadBalancer with the service's external IP
|
||||||
|
- Allow ArgoCD to transition the Ingress from "Progressing" to "Healthy"
|
||||||
@@ -1620,25 +1620,19 @@ class GetVorgabeCommentsViewTest(TestCase):
|
|||||||
# Create users
|
# Create users
|
||||||
self.regular_user = User.objects.create_user(
|
self.regular_user = User.objects.create_user(
|
||||||
username='regularuser',
|
username='regularuser',
|
||||||
password='testpass123',
|
password='testpass123'
|
||||||
first_name='Regular',
|
|
||||||
last_name='User'
|
|
||||||
)
|
)
|
||||||
|
|
||||||
self.staff_user = User.objects.create_user(
|
self.staff_user = User.objects.create_user(
|
||||||
username='staffuser',
|
username='staffuser',
|
||||||
password='testpass123',
|
password='testpass123'
|
||||||
first_name='Staff',
|
|
||||||
last_name='User'
|
|
||||||
)
|
)
|
||||||
self.staff_user.is_staff = True
|
self.staff_user.is_staff = True
|
||||||
self.staff_user.save()
|
self.staff_user.save()
|
||||||
|
|
||||||
self.other_user = User.objects.create_user(
|
self.other_user = User.objects.create_user(
|
||||||
username='otheruser',
|
username='otheruser',
|
||||||
password='testpass123',
|
password='testpass123'
|
||||||
first_name='Other',
|
|
||||||
last_name='User'
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# Create test data
|
# Create test data
|
||||||
@@ -1703,7 +1697,7 @@ class GetVorgabeCommentsViewTest(TestCase):
|
|||||||
# Should only see their own comment
|
# Should only see their own comment
|
||||||
self.assertEqual(len(data['comments']), 1)
|
self.assertEqual(len(data['comments']), 1)
|
||||||
self.assertEqual(data['comments'][0]['text'], 'Kommentar von Regular User')
|
self.assertEqual(data['comments'][0]['text'], 'Kommentar von Regular User')
|
||||||
self.assertEqual(data['comments'][0]['user'], 'Regular User')
|
self.assertEqual(data['comments'][0]['user'], 'regularuser')
|
||||||
self.assertTrue(data['comments'][0]['is_own'])
|
self.assertTrue(data['comments'][0]['is_own'])
|
||||||
|
|
||||||
def test_staff_user_sees_all_comments(self):
|
def test_staff_user_sees_all_comments(self):
|
||||||
@@ -1721,8 +1715,8 @@ class GetVorgabeCommentsViewTest(TestCase):
|
|||||||
# Should see all comments
|
# Should see all comments
|
||||||
self.assertEqual(len(data['comments']), 2)
|
self.assertEqual(len(data['comments']), 2)
|
||||||
usernames = [c['user'] for c in data['comments']]
|
usernames = [c['user'] for c in data['comments']]
|
||||||
self.assertIn('Regular User', usernames)
|
self.assertIn('regularuser', usernames)
|
||||||
self.assertIn('Other User', usernames)
|
self.assertIn('otheruser', usernames)
|
||||||
|
|
||||||
def test_get_comments_returns_404_for_nonexistent_vorgabe(self):
|
def test_get_comments_returns_404_for_nonexistent_vorgabe(self):
|
||||||
"""Test that requesting comments for non-existent Vorgabe returns 404"""
|
"""Test that requesting comments for non-existent Vorgabe returns 404"""
|
||||||
@@ -2047,16 +2041,12 @@ class DeleteVorgabeCommentViewTest(TestCase):
|
|||||||
|
|
||||||
self.other_user = User.objects.create_user(
|
self.other_user = User.objects.create_user(
|
||||||
username='otheruser',
|
username='otheruser',
|
||||||
password='testpass123',
|
password='testpass123'
|
||||||
first_name='Other',
|
|
||||||
last_name='User'
|
|
||||||
)
|
)
|
||||||
|
|
||||||
self.staff_user = User.objects.create_user(
|
self.staff_user = User.objects.create_user(
|
||||||
username='staffuser',
|
username='staffuser',
|
||||||
password='testpass123',
|
password='testpass123'
|
||||||
first_name='Staff',
|
|
||||||
last_name='User'
|
|
||||||
)
|
)
|
||||||
self.staff_user.is_staff = True
|
self.staff_user.is_staff = True
|
||||||
self.staff_user.save()
|
self.staff_user.save()
|
||||||
|
|||||||
@@ -215,7 +215,7 @@
|
|||||||
</p>
|
</p>
|
||||||
</div>
|
</div>
|
||||||
<div class="col-sm-6 text-right">
|
<div class="col-sm-6 text-right">
|
||||||
<p class="text-muted">Version {{ version|default:"0.963" }}</p>
|
<p class="text-muted">Version {{ version|default:"0.961" }}</p>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ certifi==2025.8.3
|
|||||||
charset-normalizer==3.4.3
|
charset-normalizer==3.4.3
|
||||||
curtsies==0.4.3
|
curtsies==0.4.3
|
||||||
cwcwidth==0.1.10
|
cwcwidth==0.1.10
|
||||||
Django==5.2.8
|
Django==5.2.5
|
||||||
django-admin-sortable2==2.2.8
|
django-admin-sortable2==2.2.8
|
||||||
django-js-asset==3.1.2
|
django-js-asset==3.1.2
|
||||||
django-mptt==0.17.0
|
django-mptt==0.17.0
|
||||||
@@ -33,4 +33,3 @@ sqlparse==0.5.3
|
|||||||
urllib3==2.5.0
|
urllib3==2.5.0
|
||||||
wcwidth==0.2.13
|
wcwidth==0.2.13
|
||||||
bleach==6.1.0
|
bleach==6.1.0
|
||||||
coverage==7.6.1
|
|
||||||
|
|||||||
Reference in New Issue
Block a user