Compare commits
49 commits
smusali/ev
...
trustchain
Author | SHA1 | Date | |
---|---|---|---|
1cd000dfe2 | |||
00ae97944a | |||
9f3ccfb7c7 | |||
9ed9c39ac8 | |||
30b6eeee9f | |||
afe2621783 | |||
8b12c6a01a | |||
f63adfed96 | |||
9c8fec21cf | |||
4776d2bcc5 | |||
a15a040362 | |||
fcd6dc1d60 | |||
acc3b59869 | |||
d9d5ac10e6 | |||
750669dcab | |||
88a3eed67e | |||
6c214fffc4 | |||
70100fc105 | |||
3c1163fabd | |||
539e8242ff | |||
2648333590 | |||
fe828ef993 | |||
29a6530742 | |||
a6b9274c4f | |||
a2a67161ac | |||
2e8263a99b | |||
6b9afed21f | |||
1eb1f4e0b8 | |||
7c3d60ec3a | |||
a494c6b6e8 | |||
6604d3577f | |||
f8bfa7e16a | |||
ea6cf6eabf | |||
769ce3ce7b | |||
3891fb3fa8 | |||
41eb965350 | |||
8d95612287 | |||
82b5274b15 | |||
af56ce3d78 | |||
f5c6e7aeb0 | |||
3809400e93 | |||
1def9865cf | |||
3716298639 | |||
c16317d7cf | |||
bbb8fa8269 | |||
e4c251a178 | |||
0fefd5f522 | |||
88057db0b0 | |||
91cb6c9beb |
|
@ -9,4 +9,3 @@ blueprints/local
|
||||||
.git
|
.git
|
||||||
!gen-ts-api/node_modules
|
!gen-ts-api/node_modules
|
||||||
!gen-ts-api/dist/**
|
!gen-ts-api/dist/**
|
||||||
!gen-go-api/
|
|
||||||
|
|
1
.github/codespell-words.txt
vendored
1
.github/codespell-words.txt
vendored
|
@ -2,4 +2,3 @@ keypair
|
||||||
keypairs
|
keypairs
|
||||||
hass
|
hass
|
||||||
warmup
|
warmup
|
||||||
ontext
|
|
||||||
|
|
47
.github/workflows/ci-main.yml
vendored
47
.github/workflows/ci-main.yml
vendored
|
@ -61,6 +61,10 @@ jobs:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
- name: Setup authentik env
|
||||||
|
uses: ./.github/actions/setup
|
||||||
|
with:
|
||||||
|
postgresql_version: ${{ matrix.psql }}
|
||||||
- name: checkout stable
|
- name: checkout stable
|
||||||
run: |
|
run: |
|
||||||
# Delete all poetry envs
|
# Delete all poetry envs
|
||||||
|
@ -72,7 +76,7 @@ jobs:
|
||||||
git checkout version/$(python -c "from authentik import __version__; print(__version__)")
|
git checkout version/$(python -c "from authentik import __version__; print(__version__)")
|
||||||
rm -rf .github/ scripts/
|
rm -rf .github/ scripts/
|
||||||
mv ../.github ../scripts .
|
mv ../.github ../scripts .
|
||||||
- name: Setup authentik env (stable)
|
- name: Setup authentik env (ensure stable deps are installed)
|
||||||
uses: ./.github/actions/setup
|
uses: ./.github/actions/setup
|
||||||
with:
|
with:
|
||||||
postgresql_version: ${{ matrix.psql }}
|
postgresql_version: ${{ matrix.psql }}
|
||||||
|
@ -86,20 +90,14 @@ jobs:
|
||||||
git clean -d -fx .
|
git clean -d -fx .
|
||||||
git checkout $GITHUB_SHA
|
git checkout $GITHUB_SHA
|
||||||
# Delete previous poetry env
|
# Delete previous poetry env
|
||||||
rm -rf /home/runner/.cache/pypoetry/virtualenvs/*
|
rm -rf $(poetry env info --path)
|
||||||
|
poetry install
|
||||||
- name: Setup authentik env (ensure latest deps are installed)
|
- name: Setup authentik env (ensure latest deps are installed)
|
||||||
uses: ./.github/actions/setup
|
uses: ./.github/actions/setup
|
||||||
with:
|
with:
|
||||||
postgresql_version: ${{ matrix.psql }}
|
postgresql_version: ${{ matrix.psql }}
|
||||||
- name: migrate to latest
|
- name: migrate to latest
|
||||||
run: |
|
run: poetry run python -m lifecycle.migrate
|
||||||
poetry run python -m lifecycle.migrate
|
|
||||||
- name: run tests
|
|
||||||
env:
|
|
||||||
# Test in the main database that we just migrated from the previous stable version
|
|
||||||
AUTHENTIK_POSTGRESQL__TEST__NAME: authentik
|
|
||||||
run: |
|
|
||||||
poetry run make test
|
|
||||||
test-unittest:
|
test-unittest:
|
||||||
name: test-unittest - PostgreSQL ${{ matrix.psql }}
|
name: test-unittest - PostgreSQL ${{ matrix.psql }}
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
@ -249,6 +247,12 @@ jobs:
|
||||||
VERSION_FAMILY=${{ steps.ev.outputs.versionFamily }}
|
VERSION_FAMILY=${{ steps.ev.outputs.versionFamily }}
|
||||||
cache-from: type=gha
|
cache-from: type=gha
|
||||||
cache-to: type=gha,mode=max
|
cache-to: type=gha,mode=max
|
||||||
|
- name: Comment on PR
|
||||||
|
if: github.event_name == 'pull_request'
|
||||||
|
continue-on-error: true
|
||||||
|
uses: ./.github/actions/comment-pr-instructions
|
||||||
|
with:
|
||||||
|
tag: gh-${{ steps.ev.outputs.branchNameContainer }}-${{ steps.ev.outputs.timestamp }}-${{ steps.ev.outputs.shortHash }}
|
||||||
build-arm64:
|
build-arm64:
|
||||||
needs: ci-core-mark
|
needs: ci-core-mark
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
@ -297,26 +301,3 @@ jobs:
|
||||||
platforms: linux/arm64
|
platforms: linux/arm64
|
||||||
cache-from: type=gha
|
cache-from: type=gha
|
||||||
cache-to: type=gha,mode=max
|
cache-to: type=gha,mode=max
|
||||||
pr-comment:
|
|
||||||
needs:
|
|
||||||
- build
|
|
||||||
- build-arm64
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
if: ${{ github.event_name == 'pull_request' }}
|
|
||||||
permissions:
|
|
||||||
# Needed to write comments on PRs
|
|
||||||
pull-requests: write
|
|
||||||
timeout-minutes: 120
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v4
|
|
||||||
with:
|
|
||||||
ref: ${{ github.event.pull_request.head.sha }}
|
|
||||||
- name: prepare variables
|
|
||||||
uses: ./.github/actions/docker-push-variables
|
|
||||||
id: ev
|
|
||||||
env:
|
|
||||||
DOCKER_USERNAME: ${{ secrets.DOCKER_USERNAME }}
|
|
||||||
- name: Comment on PR
|
|
||||||
uses: ./.github/actions/comment-pr-instructions
|
|
||||||
with:
|
|
||||||
tag: gh-${{ steps.ev.outputs.branchNameContainer }}-${{ steps.ev.outputs.timestamp }}-${{ steps.ev.outputs.shortHash }}
|
|
||||||
|
|
8
.github/workflows/ci-outpost.yml
vendored
8
.github/workflows/ci-outpost.yml
vendored
|
@ -16,7 +16,7 @@ jobs:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-go@v5
|
- uses: actions/setup-go@v4
|
||||||
with:
|
with:
|
||||||
go-version-file: "go.mod"
|
go-version-file: "go.mod"
|
||||||
- name: Prepare and generate API
|
- name: Prepare and generate API
|
||||||
|
@ -37,7 +37,7 @@ jobs:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-go@v5
|
- uses: actions/setup-go@v4
|
||||||
with:
|
with:
|
||||||
go-version-file: "go.mod"
|
go-version-file: "go.mod"
|
||||||
- name: Setup authentik env
|
- name: Setup authentik env
|
||||||
|
@ -65,7 +65,6 @@ jobs:
|
||||||
- proxy
|
- proxy
|
||||||
- ldap
|
- ldap
|
||||||
- radius
|
- radius
|
||||||
- rac
|
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
permissions:
|
permissions:
|
||||||
# Needed to upload contianer images to ghcr.io
|
# Needed to upload contianer images to ghcr.io
|
||||||
|
@ -120,14 +119,13 @@ jobs:
|
||||||
- proxy
|
- proxy
|
||||||
- ldap
|
- ldap
|
||||||
- radius
|
- radius
|
||||||
- rac
|
|
||||||
goos: [linux]
|
goos: [linux]
|
||||||
goarch: [amd64, arm64]
|
goarch: [amd64, arm64]
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
ref: ${{ github.event.pull_request.head.sha }}
|
ref: ${{ github.event.pull_request.head.sha }}
|
||||||
- uses: actions/setup-go@v5
|
- uses: actions/setup-go@v4
|
||||||
with:
|
with:
|
||||||
go-version-file: "go.mod"
|
go-version-file: "go.mod"
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v4
|
||||||
|
|
6
.github/workflows/codeql-analysis.yml
vendored
6
.github/workflows/codeql-analysis.yml
vendored
|
@ -27,10 +27,10 @@ jobs:
|
||||||
- name: Setup authentik env
|
- name: Setup authentik env
|
||||||
uses: ./.github/actions/setup
|
uses: ./.github/actions/setup
|
||||||
- name: Initialize CodeQL
|
- name: Initialize CodeQL
|
||||||
uses: github/codeql-action/init@v3
|
uses: github/codeql-action/init@v2
|
||||||
with:
|
with:
|
||||||
languages: ${{ matrix.language }}
|
languages: ${{ matrix.language }}
|
||||||
- name: Autobuild
|
- name: Autobuild
|
||||||
uses: github/codeql-action/autobuild@v3
|
uses: github/codeql-action/autobuild@v2
|
||||||
- name: Perform CodeQL Analysis
|
- name: Perform CodeQL Analysis
|
||||||
uses: github/codeql-action/analyze@v3
|
uses: github/codeql-action/analyze@v2
|
||||||
|
|
4
.github/workflows/gha-cache-cleanup.yml
vendored
4
.github/workflows/gha-cache-cleanup.yml
vendored
|
@ -6,10 +6,6 @@ on:
|
||||||
types:
|
types:
|
||||||
- closed
|
- closed
|
||||||
|
|
||||||
permissions:
|
|
||||||
# Permission to delete cache
|
|
||||||
actions: write
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
cleanup:
|
cleanup:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
5
.github/workflows/release-publish.yml
vendored
5
.github/workflows/release-publish.yml
vendored
|
@ -65,10 +65,9 @@ jobs:
|
||||||
- proxy
|
- proxy
|
||||||
- ldap
|
- ldap
|
||||||
- radius
|
- radius
|
||||||
- rac
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-go@v5
|
- uses: actions/setup-go@v4
|
||||||
with:
|
with:
|
||||||
go-version-file: "go.mod"
|
go-version-file: "go.mod"
|
||||||
- name: Set up QEMU
|
- name: Set up QEMU
|
||||||
|
@ -127,7 +126,7 @@ jobs:
|
||||||
goarch: [amd64, arm64]
|
goarch: [amd64, arm64]
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- uses: actions/setup-go@v5
|
- uses: actions/setup-go@v4
|
||||||
with:
|
with:
|
||||||
go-version-file: "go.mod"
|
go-version-file: "go.mod"
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v4
|
||||||
|
|
2
.github/workflows/release-tag.yml
vendored
2
.github/workflows/release-tag.yml
vendored
|
@ -30,7 +30,7 @@ jobs:
|
||||||
private_key: ${{ secrets.GH_APP_PRIVATE_KEY }}
|
private_key: ${{ secrets.GH_APP_PRIVATE_KEY }}
|
||||||
- name: Extract version number
|
- name: Extract version number
|
||||||
id: get_version
|
id: get_version
|
||||||
uses: actions/github-script@v7
|
uses: actions/github-script@v6
|
||||||
with:
|
with:
|
||||||
github-token: ${{ steps.generate_token.outputs.token }}
|
github-token: ${{ steps.generate_token.outputs.token }}
|
||||||
script: |
|
script: |
|
||||||
|
|
2
.github/workflows/repo-stale.yml
vendored
2
.github/workflows/repo-stale.yml
vendored
|
@ -18,7 +18,7 @@ jobs:
|
||||||
with:
|
with:
|
||||||
app_id: ${{ secrets.GH_APP_ID }}
|
app_id: ${{ secrets.GH_APP_ID }}
|
||||||
private_key: ${{ secrets.GH_APP_PRIVATE_KEY }}
|
private_key: ${{ secrets.GH_APP_PRIVATE_KEY }}
|
||||||
- uses: actions/stale@v9
|
- uses: actions/stale@v8
|
||||||
with:
|
with:
|
||||||
repo-token: ${{ steps.generate_token.outputs.token }}
|
repo-token: ${{ steps.generate_token.outputs.token }}
|
||||||
days-before-stale: 60
|
days-before-stale: 60
|
||||||
|
|
7
.github/workflows/translation-advice.yml
vendored
7
.github/workflows/translation-advice.yml
vendored
|
@ -7,12 +7,7 @@ on:
|
||||||
paths:
|
paths:
|
||||||
- "!**"
|
- "!**"
|
||||||
- "locale/**"
|
- "locale/**"
|
||||||
- "!locale/en/**"
|
- "web/src/locales/**"
|
||||||
- "web/xliff/**"
|
|
||||||
|
|
||||||
permissions:
|
|
||||||
# Permission to write comment
|
|
||||||
pull-requests: write
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
post-comment:
|
post-comment:
|
||||||
|
|
4
.github/workflows/translation-rename.yml
vendored
4
.github/workflows/translation-rename.yml
vendored
|
@ -6,10 +6,6 @@ on:
|
||||||
pull_request:
|
pull_request:
|
||||||
types: [opened, reopened]
|
types: [opened, reopened]
|
||||||
|
|
||||||
permissions:
|
|
||||||
# Permission to rename PR
|
|
||||||
pull-requests: write
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
rename_pr:
|
rename_pr:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
1
.vscode/extensions.json
vendored
1
.vscode/extensions.json
vendored
|
@ -14,7 +14,6 @@
|
||||||
"ms-python.pylint",
|
"ms-python.pylint",
|
||||||
"ms-python.python",
|
"ms-python.python",
|
||||||
"ms-python.vscode-pylance",
|
"ms-python.vscode-pylance",
|
||||||
"ms-python.black-formatter",
|
|
||||||
"redhat.vscode-yaml",
|
"redhat.vscode-yaml",
|
||||||
"Tobermory.es6-string-html",
|
"Tobermory.es6-string-html",
|
||||||
"unifiedjs.vscode-mdx",
|
"unifiedjs.vscode-mdx",
|
||||||
|
|
2
.vscode/settings.json
vendored
2
.vscode/settings.json
vendored
|
@ -19,8 +19,10 @@
|
||||||
"slo",
|
"slo",
|
||||||
"scim",
|
"scim",
|
||||||
],
|
],
|
||||||
|
"python.linting.pylintEnabled": true,
|
||||||
"todo-tree.tree.showCountsInTree": true,
|
"todo-tree.tree.showCountsInTree": true,
|
||||||
"todo-tree.tree.showBadges": true,
|
"todo-tree.tree.showBadges": true,
|
||||||
|
"python.formatting.provider": "black",
|
||||||
"yaml.customTags": [
|
"yaml.customTags": [
|
||||||
"!Find sequence",
|
"!Find sequence",
|
||||||
"!KeyOf scalar",
|
"!KeyOf scalar",
|
||||||
|
|
12
Dockerfile
12
Dockerfile
|
@ -37,7 +37,7 @@ COPY ./gen-ts-api /work/web/node_modules/@goauthentik/api
|
||||||
RUN npm run build
|
RUN npm run build
|
||||||
|
|
||||||
# Stage 3: Build go proxy
|
# Stage 3: Build go proxy
|
||||||
FROM --platform=${BUILDPLATFORM} docker.io/golang:1.21.6-bookworm AS go-builder
|
FROM --platform=${BUILDPLATFORM} docker.io/golang:1.21.4-bookworm AS go-builder
|
||||||
|
|
||||||
ARG TARGETOS
|
ARG TARGETOS
|
||||||
ARG TARGETARCH
|
ARG TARGETARCH
|
||||||
|
@ -69,9 +69,9 @@ RUN --mount=type=cache,sharing=locked,target=/go/pkg/mod \
|
||||||
GOARM="${TARGETVARIANT#v}" go build -o /go/authentik ./cmd/server
|
GOARM="${TARGETVARIANT#v}" go build -o /go/authentik ./cmd/server
|
||||||
|
|
||||||
# Stage 4: MaxMind GeoIP
|
# Stage 4: MaxMind GeoIP
|
||||||
FROM --platform=${BUILDPLATFORM} ghcr.io/maxmind/geoipupdate:v6.1 as geoip
|
FROM --platform=${BUILDPLATFORM} ghcr.io/maxmind/geoipupdate:v6.0 as geoip
|
||||||
|
|
||||||
ENV GEOIPUPDATE_EDITION_IDS="GeoLite2-City GeoLite2-ASN"
|
ENV GEOIPUPDATE_EDITION_IDS="GeoLite2-City"
|
||||||
ENV GEOIPUPDATE_VERBOSE="true"
|
ENV GEOIPUPDATE_VERBOSE="true"
|
||||||
ENV GEOIPUPDATE_ACCOUNT_ID_FILE="/run/secrets/GEOIPUPDATE_ACCOUNT_ID"
|
ENV GEOIPUPDATE_ACCOUNT_ID_FILE="/run/secrets/GEOIPUPDATE_ACCOUNT_ID"
|
||||||
ENV GEOIPUPDATE_LICENSE_KEY_FILE="/run/secrets/GEOIPUPDATE_LICENSE_KEY"
|
ENV GEOIPUPDATE_LICENSE_KEY_FILE="/run/secrets/GEOIPUPDATE_LICENSE_KEY"
|
||||||
|
@ -83,7 +83,7 @@ RUN --mount=type=secret,id=GEOIPUPDATE_ACCOUNT_ID \
|
||||||
/bin/sh -c "/usr/bin/entry.sh || echo 'Failed to get GeoIP database, disabling'; exit 0"
|
/bin/sh -c "/usr/bin/entry.sh || echo 'Failed to get GeoIP database, disabling'; exit 0"
|
||||||
|
|
||||||
# Stage 5: Python dependencies
|
# Stage 5: Python dependencies
|
||||||
FROM docker.io/python:3.12.1-slim-bookworm AS python-deps
|
FROM docker.io/python:3.11.5-bookworm AS python-deps
|
||||||
|
|
||||||
WORKDIR /ak-root/poetry
|
WORKDIR /ak-root/poetry
|
||||||
|
|
||||||
|
@ -108,7 +108,7 @@ RUN --mount=type=bind,target=./pyproject.toml,src=./pyproject.toml \
|
||||||
poetry install --only=main --no-ansi --no-interaction
|
poetry install --only=main --no-ansi --no-interaction
|
||||||
|
|
||||||
# Stage 6: Run
|
# Stage 6: Run
|
||||||
FROM docker.io/python:3.12.1-slim-bookworm AS final-image
|
FROM docker.io/python:3.11.5-slim-bookworm AS final-image
|
||||||
|
|
||||||
ARG GIT_BUILD_HASH
|
ARG GIT_BUILD_HASH
|
||||||
ARG VERSION
|
ARG VERSION
|
||||||
|
@ -125,7 +125,7 @@ WORKDIR /
|
||||||
# We cannot cache this layer otherwise we'll end up with a bigger image
|
# We cannot cache this layer otherwise we'll end up with a bigger image
|
||||||
RUN apt-get update && \
|
RUN apt-get update && \
|
||||||
# Required for runtime
|
# Required for runtime
|
||||||
apt-get install -y --no-install-recommends libpq5 openssl libxmlsec1-openssl libmaxminddb0 ca-certificates && \
|
apt-get install -y --no-install-recommends libpq5 openssl libxmlsec1-openssl libmaxminddb0 && \
|
||||||
# Required for bootstrap & healtcheck
|
# Required for bootstrap & healtcheck
|
||||||
apt-get install -y --no-install-recommends runit && \
|
apt-get install -y --no-install-recommends runit && \
|
||||||
apt-get clean && \
|
apt-get clean && \
|
||||||
|
|
9
Makefile
9
Makefile
|
@ -58,7 +58,7 @@ test: ## Run the server tests and produce a coverage report (locally)
|
||||||
lint-fix: ## Lint and automatically fix errors in the python source code. Reports spelling errors.
|
lint-fix: ## Lint and automatically fix errors in the python source code. Reports spelling errors.
|
||||||
isort $(PY_SOURCES)
|
isort $(PY_SOURCES)
|
||||||
black $(PY_SOURCES)
|
black $(PY_SOURCES)
|
||||||
ruff --fix $(PY_SOURCES)
|
ruff $(PY_SOURCES)
|
||||||
codespell -w $(CODESPELL_ARGS)
|
codespell -w $(CODESPELL_ARGS)
|
||||||
|
|
||||||
lint: ## Lint the python and golang sources
|
lint: ## Lint the python and golang sources
|
||||||
|
@ -110,14 +110,11 @@ gen-diff: ## (Release) generate the changelog diff between the current schema a
|
||||||
--markdown /local/diff.md \
|
--markdown /local/diff.md \
|
||||||
/local/old_schema.yml /local/schema.yml
|
/local/old_schema.yml /local/schema.yml
|
||||||
rm old_schema.yml
|
rm old_schema.yml
|
||||||
sed -i 's/{/{/g' diff.md
|
|
||||||
sed -i 's/}/}/g' diff.md
|
|
||||||
npx prettier --write diff.md
|
npx prettier --write diff.md
|
||||||
|
|
||||||
gen-clean:
|
gen-clean:
|
||||||
rm -rf gen-go-api/
|
rm -rf web/api/src/
|
||||||
rm -rf gen-ts-api/
|
rm -rf api/
|
||||||
rm -rf web/node_modules/@goauthentik/api/
|
|
||||||
|
|
||||||
gen-client-ts: ## Build and install the authentik API for Typescript into the authentik UI Application
|
gen-client-ts: ## Build and install the authentik API for Typescript into the authentik UI Application
|
||||||
docker run \
|
docker run \
|
||||||
|
|
|
@ -1,9 +1,5 @@
|
||||||
authentik takes security very seriously. We follow the rules of [responsible disclosure](https://en.wikipedia.org/wiki/Responsible_disclosure), and we urge our community to do so as well, instead of reporting vulnerabilities publicly. This allows us to patch the issue quickly, announce it's existence and release the fixed version.
|
authentik takes security very seriously. We follow the rules of [responsible disclosure](https://en.wikipedia.org/wiki/Responsible_disclosure), and we urge our community to do so as well, instead of reporting vulnerabilities publicly. This allows us to patch the issue quickly, announce it's existence and release the fixed version.
|
||||||
|
|
||||||
## Independent audits and pentests
|
|
||||||
|
|
||||||
In May/June of 2023 [Cure53](https://cure53.de) conducted an audit and pentest. The [results](https://cure53.de/pentest-report_authentik.pdf) are published on the [Cure53 website](https://cure53.de/#publications-2023). For more details about authentik's response to the findings of the audit refer to [2023-06 Cure53 Code audit](https://goauthentik.io/docs/security/2023-06-cure53).
|
|
||||||
|
|
||||||
## What authentik classifies as a CVE
|
## What authentik classifies as a CVE
|
||||||
|
|
||||||
CVE (Common Vulnerability and Exposure) is a system designed to aggregate all vulnerabilities. As such, a CVE will be issued when there is a either vulnerability or exposure. Per NIST, A vulnerability is:
|
CVE (Common Vulnerability and Exposure) is a system designed to aggregate all vulnerabilities. As such, a CVE will be issued when there is a either vulnerability or exposure. Per NIST, A vulnerability is:
|
||||||
|
|
|
@ -30,7 +30,7 @@ class RuntimeDict(TypedDict):
|
||||||
uname: str
|
uname: str
|
||||||
|
|
||||||
|
|
||||||
class SystemInfoSerializer(PassiveSerializer):
|
class SystemSerializer(PassiveSerializer):
|
||||||
"""Get system information."""
|
"""Get system information."""
|
||||||
|
|
||||||
http_headers = SerializerMethodField()
|
http_headers = SerializerMethodField()
|
||||||
|
@ -91,14 +91,14 @@ class SystemView(APIView):
|
||||||
permission_classes = [HasPermission("authentik_rbac.view_system_info")]
|
permission_classes = [HasPermission("authentik_rbac.view_system_info")]
|
||||||
pagination_class = None
|
pagination_class = None
|
||||||
filter_backends = []
|
filter_backends = []
|
||||||
serializer_class = SystemInfoSerializer
|
serializer_class = SystemSerializer
|
||||||
|
|
||||||
@extend_schema(responses={200: SystemInfoSerializer(many=False)})
|
@extend_schema(responses={200: SystemSerializer(many=False)})
|
||||||
def get(self, request: Request) -> Response:
|
def get(self, request: Request) -> Response:
|
||||||
"""Get system information."""
|
"""Get system information."""
|
||||||
return Response(SystemInfoSerializer(request).data)
|
return Response(SystemSerializer(request).data)
|
||||||
|
|
||||||
@extend_schema(responses={200: SystemInfoSerializer(many=False)})
|
@extend_schema(responses={200: SystemSerializer(many=False)})
|
||||||
def post(self, request: Request) -> Response:
|
def post(self, request: Request) -> Response:
|
||||||
"""Get system information."""
|
"""Get system information."""
|
||||||
return Response(SystemInfoSerializer(request).data)
|
return Response(SystemSerializer(request).data)
|
||||||
|
|
|
@ -12,8 +12,6 @@ from authentik.blueprints.tests import reconcile_app
|
||||||
from authentik.core.models import Token, TokenIntents, User, UserTypes
|
from authentik.core.models import Token, TokenIntents, User, UserTypes
|
||||||
from authentik.core.tests.utils import create_test_admin_user, create_test_flow
|
from authentik.core.tests.utils import create_test_admin_user, create_test_flow
|
||||||
from authentik.lib.generators import generate_id
|
from authentik.lib.generators import generate_id
|
||||||
from authentik.outposts.apps import MANAGED_OUTPOST
|
|
||||||
from authentik.outposts.models import Outpost
|
|
||||||
from authentik.providers.oauth2.constants import SCOPE_AUTHENTIK_API
|
from authentik.providers.oauth2.constants import SCOPE_AUTHENTIK_API
|
||||||
from authentik.providers.oauth2.models import AccessToken, OAuth2Provider
|
from authentik.providers.oauth2.models import AccessToken, OAuth2Provider
|
||||||
|
|
||||||
|
@ -51,12 +49,8 @@ class TestAPIAuth(TestCase):
|
||||||
with self.assertRaises(AuthenticationFailed):
|
with self.assertRaises(AuthenticationFailed):
|
||||||
bearer_auth(f"Bearer {token.key}".encode())
|
bearer_auth(f"Bearer {token.key}".encode())
|
||||||
|
|
||||||
@reconcile_app("authentik_outposts")
|
def test_managed_outpost(self):
|
||||||
def test_managed_outpost_fail(self):
|
|
||||||
"""Test managed outpost"""
|
"""Test managed outpost"""
|
||||||
outpost = Outpost.objects.filter(managed=MANAGED_OUTPOST).first()
|
|
||||||
outpost.user.delete()
|
|
||||||
outpost.delete()
|
|
||||||
with self.assertRaises(AuthenticationFailed):
|
with self.assertRaises(AuthenticationFailed):
|
||||||
bearer_auth(f"Bearer {settings.SECRET_KEY}".encode())
|
bearer_auth(f"Bearer {settings.SECRET_KEY}".encode())
|
||||||
|
|
||||||
|
|
|
@ -19,7 +19,7 @@ from rest_framework.response import Response
|
||||||
from rest_framework.views import APIView
|
from rest_framework.views import APIView
|
||||||
|
|
||||||
from authentik.core.api.utils import PassiveSerializer
|
from authentik.core.api.utils import PassiveSerializer
|
||||||
from authentik.events.context_processors.base import get_context_processors
|
from authentik.events.geo import GEOIP_READER
|
||||||
from authentik.lib.config import CONFIG
|
from authentik.lib.config import CONFIG
|
||||||
|
|
||||||
capabilities = Signal()
|
capabilities = Signal()
|
||||||
|
@ -30,7 +30,6 @@ class Capabilities(models.TextChoices):
|
||||||
|
|
||||||
CAN_SAVE_MEDIA = "can_save_media"
|
CAN_SAVE_MEDIA = "can_save_media"
|
||||||
CAN_GEO_IP = "can_geo_ip"
|
CAN_GEO_IP = "can_geo_ip"
|
||||||
CAN_ASN = "can_asn"
|
|
||||||
CAN_IMPERSONATE = "can_impersonate"
|
CAN_IMPERSONATE = "can_impersonate"
|
||||||
CAN_DEBUG = "can_debug"
|
CAN_DEBUG = "can_debug"
|
||||||
IS_ENTERPRISE = "is_enterprise"
|
IS_ENTERPRISE = "is_enterprise"
|
||||||
|
@ -69,9 +68,8 @@ class ConfigView(APIView):
|
||||||
deb_test = settings.DEBUG or settings.TEST
|
deb_test = settings.DEBUG or settings.TEST
|
||||||
if Path(settings.MEDIA_ROOT).is_mount() or deb_test:
|
if Path(settings.MEDIA_ROOT).is_mount() or deb_test:
|
||||||
caps.append(Capabilities.CAN_SAVE_MEDIA)
|
caps.append(Capabilities.CAN_SAVE_MEDIA)
|
||||||
for processor in get_context_processors():
|
if GEOIP_READER.enabled:
|
||||||
if cap := processor.capability():
|
caps.append(Capabilities.CAN_GEO_IP)
|
||||||
caps.append(cap)
|
|
||||||
if CONFIG.get_bool("impersonation"):
|
if CONFIG.get_bool("impersonation"):
|
||||||
caps.append(Capabilities.CAN_IMPERSONATE)
|
caps.append(Capabilities.CAN_IMPERSONATE)
|
||||||
if settings.DEBUG: # pragma: no cover
|
if settings.DEBUG: # pragma: no cover
|
||||||
|
@ -95,10 +93,10 @@ class ConfigView(APIView):
|
||||||
"traces_sample_rate": float(CONFIG.get("error_reporting.sample_rate", 0.4)),
|
"traces_sample_rate": float(CONFIG.get("error_reporting.sample_rate", 0.4)),
|
||||||
},
|
},
|
||||||
"capabilities": self.get_capabilities(),
|
"capabilities": self.get_capabilities(),
|
||||||
"cache_timeout": CONFIG.get_int("cache.timeout"),
|
"cache_timeout": CONFIG.get_int("redis.cache_timeout"),
|
||||||
"cache_timeout_flows": CONFIG.get_int("cache.timeout_flows"),
|
"cache_timeout_flows": CONFIG.get_int("redis.cache_timeout_flows"),
|
||||||
"cache_timeout_policies": CONFIG.get_int("cache.timeout_policies"),
|
"cache_timeout_policies": CONFIG.get_int("redis.cache_timeout_policies"),
|
||||||
"cache_timeout_reputation": CONFIG.get_int("cache.timeout_reputation"),
|
"cache_timeout_reputation": CONFIG.get_int("redis.cache_timeout_reputation"),
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
@ -3,7 +3,7 @@ from django.utils.translation import gettext_lazy as _
|
||||||
from drf_spectacular.utils import extend_schema, inline_serializer
|
from drf_spectacular.utils import extend_schema, inline_serializer
|
||||||
from rest_framework.decorators import action
|
from rest_framework.decorators import action
|
||||||
from rest_framework.exceptions import ValidationError
|
from rest_framework.exceptions import ValidationError
|
||||||
from rest_framework.fields import CharField, DateTimeField
|
from rest_framework.fields import CharField, DateTimeField, JSONField
|
||||||
from rest_framework.request import Request
|
from rest_framework.request import Request
|
||||||
from rest_framework.response import Response
|
from rest_framework.response import Response
|
||||||
from rest_framework.serializers import ListSerializer, ModelSerializer
|
from rest_framework.serializers import ListSerializer, ModelSerializer
|
||||||
|
@ -15,7 +15,7 @@ from authentik.blueprints.v1.importer import Importer
|
||||||
from authentik.blueprints.v1.oci import OCI_PREFIX
|
from authentik.blueprints.v1.oci import OCI_PREFIX
|
||||||
from authentik.blueprints.v1.tasks import apply_blueprint, blueprints_find_dict
|
from authentik.blueprints.v1.tasks import apply_blueprint, blueprints_find_dict
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.api.utils import JSONDictField, PassiveSerializer
|
from authentik.core.api.utils import PassiveSerializer
|
||||||
|
|
||||||
|
|
||||||
class ManagedSerializer:
|
class ManagedSerializer:
|
||||||
|
@ -28,7 +28,7 @@ class MetadataSerializer(PassiveSerializer):
|
||||||
"""Serializer for blueprint metadata"""
|
"""Serializer for blueprint metadata"""
|
||||||
|
|
||||||
name = CharField()
|
name = CharField()
|
||||||
labels = JSONDictField()
|
labels = JSONField()
|
||||||
|
|
||||||
|
|
||||||
class BlueprintInstanceSerializer(ModelSerializer):
|
class BlueprintInstanceSerializer(ModelSerializer):
|
||||||
|
|
|
@ -2,11 +2,11 @@
|
||||||
from typing import TYPE_CHECKING
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
from rest_framework.exceptions import ValidationError
|
from rest_framework.exceptions import ValidationError
|
||||||
from rest_framework.fields import BooleanField
|
from rest_framework.fields import BooleanField, JSONField
|
||||||
from structlog.stdlib import get_logger
|
from structlog.stdlib import get_logger
|
||||||
|
|
||||||
from authentik.blueprints.v1.meta.registry import BaseMetaModel, MetaResult, registry
|
from authentik.blueprints.v1.meta.registry import BaseMetaModel, MetaResult, registry
|
||||||
from authentik.core.api.utils import JSONDictField, PassiveSerializer
|
from authentik.core.api.utils import PassiveSerializer, is_dict
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
from authentik.blueprints.models import BlueprintInstance
|
from authentik.blueprints.models import BlueprintInstance
|
||||||
|
@ -17,7 +17,7 @@ LOGGER = get_logger()
|
||||||
class ApplyBlueprintMetaSerializer(PassiveSerializer):
|
class ApplyBlueprintMetaSerializer(PassiveSerializer):
|
||||||
"""Serializer for meta apply blueprint model"""
|
"""Serializer for meta apply blueprint model"""
|
||||||
|
|
||||||
identifiers = JSONDictField()
|
identifiers = JSONField(validators=[is_dict])
|
||||||
required = BooleanField(default=True)
|
required = BooleanField(default=True)
|
||||||
|
|
||||||
# We cannot override `instance` as that will confuse rest_framework
|
# We cannot override `instance` as that will confuse rest_framework
|
||||||
|
|
|
@ -14,8 +14,7 @@ from ua_parser import user_agent_parser
|
||||||
from authentik.api.authorization import OwnerSuperuserPermissions
|
from authentik.api.authorization import OwnerSuperuserPermissions
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.models import AuthenticatedSession
|
from authentik.core.models import AuthenticatedSession
|
||||||
from authentik.events.context_processors.asn import ASN_CONTEXT_PROCESSOR, ASNDict
|
from authentik.events.geo import GEOIP_READER, GeoIPDict
|
||||||
from authentik.events.context_processors.geoip import GEOIP_CONTEXT_PROCESSOR, GeoIPDict
|
|
||||||
|
|
||||||
|
|
||||||
class UserAgentDeviceDict(TypedDict):
|
class UserAgentDeviceDict(TypedDict):
|
||||||
|
@ -60,7 +59,6 @@ class AuthenticatedSessionSerializer(ModelSerializer):
|
||||||
current = SerializerMethodField()
|
current = SerializerMethodField()
|
||||||
user_agent = SerializerMethodField()
|
user_agent = SerializerMethodField()
|
||||||
geo_ip = SerializerMethodField()
|
geo_ip = SerializerMethodField()
|
||||||
asn = SerializerMethodField()
|
|
||||||
|
|
||||||
def get_current(self, instance: AuthenticatedSession) -> bool:
|
def get_current(self, instance: AuthenticatedSession) -> bool:
|
||||||
"""Check if session is currently active session"""
|
"""Check if session is currently active session"""
|
||||||
|
@ -72,12 +70,8 @@ class AuthenticatedSessionSerializer(ModelSerializer):
|
||||||
return user_agent_parser.Parse(instance.last_user_agent)
|
return user_agent_parser.Parse(instance.last_user_agent)
|
||||||
|
|
||||||
def get_geo_ip(self, instance: AuthenticatedSession) -> Optional[GeoIPDict]: # pragma: no cover
|
def get_geo_ip(self, instance: AuthenticatedSession) -> Optional[GeoIPDict]: # pragma: no cover
|
||||||
"""Get GeoIP Data"""
|
"""Get parsed user agent"""
|
||||||
return GEOIP_CONTEXT_PROCESSOR.city_dict(instance.last_ip)
|
return GEOIP_READER.city_dict(instance.last_ip)
|
||||||
|
|
||||||
def get_asn(self, instance: AuthenticatedSession) -> Optional[ASNDict]: # pragma: no cover
|
|
||||||
"""Get ASN Data"""
|
|
||||||
return ASN_CONTEXT_PROCESSOR.asn_dict(instance.last_ip)
|
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
model = AuthenticatedSession
|
model = AuthenticatedSession
|
||||||
|
@ -86,7 +80,6 @@ class AuthenticatedSessionSerializer(ModelSerializer):
|
||||||
"current",
|
"current",
|
||||||
"user_agent",
|
"user_agent",
|
||||||
"geo_ip",
|
"geo_ip",
|
||||||
"asn",
|
|
||||||
"user",
|
"user",
|
||||||
"last_ip",
|
"last_ip",
|
||||||
"last_user_agent",
|
"last_user_agent",
|
||||||
|
|
|
@ -8,7 +8,7 @@ from django_filters.filterset import FilterSet
|
||||||
from drf_spectacular.utils import OpenApiResponse, extend_schema
|
from drf_spectacular.utils import OpenApiResponse, extend_schema
|
||||||
from guardian.shortcuts import get_objects_for_user
|
from guardian.shortcuts import get_objects_for_user
|
||||||
from rest_framework.decorators import action
|
from rest_framework.decorators import action
|
||||||
from rest_framework.fields import CharField, IntegerField
|
from rest_framework.fields import CharField, IntegerField, JSONField
|
||||||
from rest_framework.request import Request
|
from rest_framework.request import Request
|
||||||
from rest_framework.response import Response
|
from rest_framework.response import Response
|
||||||
from rest_framework.serializers import ListSerializer, ModelSerializer, ValidationError
|
from rest_framework.serializers import ListSerializer, ModelSerializer, ValidationError
|
||||||
|
@ -16,7 +16,7 @@ from rest_framework.viewsets import ModelViewSet
|
||||||
|
|
||||||
from authentik.api.decorators import permission_required
|
from authentik.api.decorators import permission_required
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.api.utils import JSONDictField, PassiveSerializer
|
from authentik.core.api.utils import PassiveSerializer, is_dict
|
||||||
from authentik.core.models import Group, User
|
from authentik.core.models import Group, User
|
||||||
from authentik.rbac.api.roles import RoleSerializer
|
from authentik.rbac.api.roles import RoleSerializer
|
||||||
|
|
||||||
|
@ -24,7 +24,7 @@ from authentik.rbac.api.roles import RoleSerializer
|
||||||
class GroupMemberSerializer(ModelSerializer):
|
class GroupMemberSerializer(ModelSerializer):
|
||||||
"""Stripped down user serializer to show relevant users for groups"""
|
"""Stripped down user serializer to show relevant users for groups"""
|
||||||
|
|
||||||
attributes = JSONDictField(required=False)
|
attributes = JSONField(validators=[is_dict], required=False)
|
||||||
uid = CharField(read_only=True)
|
uid = CharField(read_only=True)
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
|
@ -44,7 +44,7 @@ class GroupMemberSerializer(ModelSerializer):
|
||||||
class GroupSerializer(ModelSerializer):
|
class GroupSerializer(ModelSerializer):
|
||||||
"""Group Serializer"""
|
"""Group Serializer"""
|
||||||
|
|
||||||
attributes = JSONDictField(required=False)
|
attributes = JSONField(validators=[is_dict], required=False)
|
||||||
users_obj = ListSerializer(
|
users_obj = ListSerializer(
|
||||||
child=GroupMemberSerializer(), read_only=True, source="users", required=False
|
child=GroupMemberSerializer(), read_only=True, source="users", required=False
|
||||||
)
|
)
|
||||||
|
|
|
@ -19,7 +19,6 @@ from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.api.utils import MetaNameSerializer, PassiveSerializer, TypeCreateSerializer
|
from authentik.core.api.utils import MetaNameSerializer, PassiveSerializer, TypeCreateSerializer
|
||||||
from authentik.core.expression.evaluator import PropertyMappingEvaluator
|
from authentik.core.expression.evaluator import PropertyMappingEvaluator
|
||||||
from authentik.core.models import PropertyMapping
|
from authentik.core.models import PropertyMapping
|
||||||
from authentik.enterprise.apps import EnterpriseConfig
|
|
||||||
from authentik.events.utils import sanitize_item
|
from authentik.events.utils import sanitize_item
|
||||||
from authentik.lib.utils.reflection import all_subclasses
|
from authentik.lib.utils.reflection import all_subclasses
|
||||||
from authentik.policies.api.exec import PolicyTestSerializer
|
from authentik.policies.api.exec import PolicyTestSerializer
|
||||||
|
@ -96,7 +95,6 @@ class PropertyMappingViewSet(
|
||||||
"description": subclass.__doc__,
|
"description": subclass.__doc__,
|
||||||
"component": subclass().component,
|
"component": subclass().component,
|
||||||
"model_name": subclass._meta.model_name,
|
"model_name": subclass._meta.model_name,
|
||||||
"requires_enterprise": isinstance(subclass._meta.app_config, EnterpriseConfig),
|
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
return Response(TypeCreateSerializer(data, many=True).data)
|
return Response(TypeCreateSerializer(data, many=True).data)
|
||||||
|
|
|
@ -16,7 +16,6 @@ from rest_framework.viewsets import GenericViewSet
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.api.utils import MetaNameSerializer, TypeCreateSerializer
|
from authentik.core.api.utils import MetaNameSerializer, TypeCreateSerializer
|
||||||
from authentik.core.models import Provider
|
from authentik.core.models import Provider
|
||||||
from authentik.enterprise.apps import EnterpriseConfig
|
|
||||||
from authentik.lib.utils.reflection import all_subclasses
|
from authentik.lib.utils.reflection import all_subclasses
|
||||||
|
|
||||||
|
|
||||||
|
@ -114,7 +113,6 @@ class ProviderViewSet(
|
||||||
"description": subclass.__doc__,
|
"description": subclass.__doc__,
|
||||||
"component": subclass().component,
|
"component": subclass().component,
|
||||||
"model_name": subclass._meta.model_name,
|
"model_name": subclass._meta.model_name,
|
||||||
"requires_enterprise": isinstance(subclass._meta.app_config, EnterpriseConfig),
|
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
data.append(
|
data.append(
|
||||||
|
|
|
@ -32,7 +32,13 @@ from drf_spectacular.utils import (
|
||||||
)
|
)
|
||||||
from guardian.shortcuts import get_anonymous_user, get_objects_for_user
|
from guardian.shortcuts import get_anonymous_user, get_objects_for_user
|
||||||
from rest_framework.decorators import action
|
from rest_framework.decorators import action
|
||||||
from rest_framework.fields import CharField, IntegerField, ListField, SerializerMethodField
|
from rest_framework.fields import (
|
||||||
|
CharField,
|
||||||
|
IntegerField,
|
||||||
|
JSONField,
|
||||||
|
ListField,
|
||||||
|
SerializerMethodField,
|
||||||
|
)
|
||||||
from rest_framework.request import Request
|
from rest_framework.request import Request
|
||||||
from rest_framework.response import Response
|
from rest_framework.response import Response
|
||||||
from rest_framework.serializers import (
|
from rest_framework.serializers import (
|
||||||
|
@ -51,7 +57,7 @@ from authentik.admin.api.metrics import CoordinateSerializer
|
||||||
from authentik.api.decorators import permission_required
|
from authentik.api.decorators import permission_required
|
||||||
from authentik.blueprints.v1.importer import SERIALIZER_CONTEXT_BLUEPRINT
|
from authentik.blueprints.v1.importer import SERIALIZER_CONTEXT_BLUEPRINT
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.api.utils import JSONDictField, LinkSerializer, PassiveSerializer
|
from authentik.core.api.utils import LinkSerializer, PassiveSerializer, is_dict
|
||||||
from authentik.core.middleware import (
|
from authentik.core.middleware import (
|
||||||
SESSION_KEY_IMPERSONATE_ORIGINAL_USER,
|
SESSION_KEY_IMPERSONATE_ORIGINAL_USER,
|
||||||
SESSION_KEY_IMPERSONATE_USER,
|
SESSION_KEY_IMPERSONATE_USER,
|
||||||
|
@ -83,7 +89,7 @@ LOGGER = get_logger()
|
||||||
class UserGroupSerializer(ModelSerializer):
|
class UserGroupSerializer(ModelSerializer):
|
||||||
"""Simplified Group Serializer for user's groups"""
|
"""Simplified Group Serializer for user's groups"""
|
||||||
|
|
||||||
attributes = JSONDictField(required=False)
|
attributes = JSONField(required=False)
|
||||||
parent_name = CharField(source="parent.name", read_only=True)
|
parent_name = CharField(source="parent.name", read_only=True)
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
|
@ -104,7 +110,7 @@ class UserSerializer(ModelSerializer):
|
||||||
|
|
||||||
is_superuser = BooleanField(read_only=True)
|
is_superuser = BooleanField(read_only=True)
|
||||||
avatar = CharField(read_only=True)
|
avatar = CharField(read_only=True)
|
||||||
attributes = JSONDictField(required=False)
|
attributes = JSONField(validators=[is_dict], required=False)
|
||||||
groups = PrimaryKeyRelatedField(
|
groups = PrimaryKeyRelatedField(
|
||||||
allow_empty=True, many=True, source="ak_groups", queryset=Group.objects.all(), default=list
|
allow_empty=True, many=True, source="ak_groups", queryset=Group.objects.all(), default=list
|
||||||
)
|
)
|
||||||
|
|
|
@ -2,10 +2,7 @@
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
from django.db.models import Model
|
from django.db.models import Model
|
||||||
from drf_spectacular.extensions import OpenApiSerializerFieldExtension
|
from rest_framework.fields import CharField, IntegerField, JSONField
|
||||||
from drf_spectacular.plumbing import build_basic_type
|
|
||||||
from drf_spectacular.types import OpenApiTypes
|
|
||||||
from rest_framework.fields import BooleanField, CharField, IntegerField, JSONField
|
|
||||||
from rest_framework.serializers import Serializer, SerializerMethodField, ValidationError
|
from rest_framework.serializers import Serializer, SerializerMethodField, ValidationError
|
||||||
|
|
||||||
|
|
||||||
|
@ -16,21 +13,6 @@ def is_dict(value: Any):
|
||||||
raise ValidationError("Value must be a dictionary, and not have any duplicate keys.")
|
raise ValidationError("Value must be a dictionary, and not have any duplicate keys.")
|
||||||
|
|
||||||
|
|
||||||
class JSONDictField(JSONField):
|
|
||||||
"""JSON Field which only allows dictionaries"""
|
|
||||||
|
|
||||||
default_validators = [is_dict]
|
|
||||||
|
|
||||||
|
|
||||||
class JSONExtension(OpenApiSerializerFieldExtension):
|
|
||||||
"""Generate API Schema for JSON fields as"""
|
|
||||||
|
|
||||||
target_class = "authentik.core.api.utils.JSONDictField"
|
|
||||||
|
|
||||||
def map_serializer_field(self, auto_schema, direction):
|
|
||||||
return build_basic_type(OpenApiTypes.OBJECT)
|
|
||||||
|
|
||||||
|
|
||||||
class PassiveSerializer(Serializer):
|
class PassiveSerializer(Serializer):
|
||||||
"""Base serializer class which doesn't implement create/update methods"""
|
"""Base serializer class which doesn't implement create/update methods"""
|
||||||
|
|
||||||
|
@ -44,7 +26,7 @@ class PassiveSerializer(Serializer):
|
||||||
class PropertyMappingPreviewSerializer(PassiveSerializer):
|
class PropertyMappingPreviewSerializer(PassiveSerializer):
|
||||||
"""Preview how the current user is mapped via the property mappings selected in a provider"""
|
"""Preview how the current user is mapped via the property mappings selected in a provider"""
|
||||||
|
|
||||||
preview = JSONDictField(read_only=True)
|
preview = JSONField(read_only=True)
|
||||||
|
|
||||||
|
|
||||||
class MetaNameSerializer(PassiveSerializer):
|
class MetaNameSerializer(PassiveSerializer):
|
||||||
|
@ -74,7 +56,6 @@ class TypeCreateSerializer(PassiveSerializer):
|
||||||
description = CharField(required=True)
|
description = CharField(required=True)
|
||||||
component = CharField(required=True)
|
component = CharField(required=True)
|
||||||
model_name = CharField(required=True)
|
model_name = CharField(required=True)
|
||||||
requires_enterprise = BooleanField(default=False)
|
|
||||||
|
|
||||||
|
|
||||||
class CacheSerializer(PassiveSerializer):
|
class CacheSerializer(PassiveSerializer):
|
||||||
|
|
|
@ -1,29 +1,22 @@
|
||||||
"""Channels base classes"""
|
"""Channels base classes"""
|
||||||
from channels.db import database_sync_to_async
|
|
||||||
from channels.exceptions import DenyConnection
|
from channels.exceptions import DenyConnection
|
||||||
|
from channels.generic.websocket import JsonWebsocketConsumer
|
||||||
from rest_framework.exceptions import AuthenticationFailed
|
from rest_framework.exceptions import AuthenticationFailed
|
||||||
from structlog.stdlib import get_logger
|
from structlog.stdlib import get_logger
|
||||||
|
|
||||||
from authentik.api.authentication import bearer_auth
|
from authentik.api.authentication import bearer_auth
|
||||||
|
from authentik.core.models import User
|
||||||
|
|
||||||
LOGGER = get_logger()
|
LOGGER = get_logger()
|
||||||
|
|
||||||
|
|
||||||
class TokenOutpostMiddleware:
|
class AuthJsonConsumer(JsonWebsocketConsumer):
|
||||||
"""Authorize a client with a token"""
|
"""Authorize a client with a token"""
|
||||||
|
|
||||||
def __init__(self, inner):
|
user: User
|
||||||
self.inner = inner
|
|
||||||
|
|
||||||
async def __call__(self, scope, receive, send):
|
def connect(self):
|
||||||
scope = dict(scope)
|
headers = dict(self.scope["headers"])
|
||||||
await self.auth(scope)
|
|
||||||
return await self.inner(scope, receive, send)
|
|
||||||
|
|
||||||
@database_sync_to_async
|
|
||||||
def auth(self, scope):
|
|
||||||
"""Authenticate request from header"""
|
|
||||||
headers = dict(scope["headers"])
|
|
||||||
if b"authorization" not in headers:
|
if b"authorization" not in headers:
|
||||||
LOGGER.warning("WS Request without authorization header")
|
LOGGER.warning("WS Request without authorization header")
|
||||||
raise DenyConnection()
|
raise DenyConnection()
|
||||||
|
@ -39,4 +32,4 @@ class TokenOutpostMiddleware:
|
||||||
LOGGER.warning("Failed to authenticate", exc=exc)
|
LOGGER.warning("Failed to authenticate", exc=exc)
|
||||||
raise DenyConnection()
|
raise DenyConnection()
|
||||||
|
|
||||||
scope["user"] = user
|
self.user = user
|
||||||
|
|
|
@ -30,6 +30,7 @@ from authentik.lib.models import (
|
||||||
DomainlessFormattedURLValidator,
|
DomainlessFormattedURLValidator,
|
||||||
SerializerModel,
|
SerializerModel,
|
||||||
)
|
)
|
||||||
|
from authentik.lib.utils.http import get_client_ip
|
||||||
from authentik.policies.models import PolicyBindingModel
|
from authentik.policies.models import PolicyBindingModel
|
||||||
from authentik.root.install_id import get_install_id
|
from authentik.root.install_id import get_install_id
|
||||||
|
|
||||||
|
@ -516,7 +517,7 @@ class Source(ManagedModel, SerializerModel, PolicyBindingModel):
|
||||||
objects = InheritanceManager()
|
objects = InheritanceManager()
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def icon_url(self) -> Optional[str]:
|
def get_icon(self) -> Optional[str]:
|
||||||
"""Get the URL to the Icon. If the name is /static or
|
"""Get the URL to the Icon. If the name is /static or
|
||||||
starts with http it is returned as-is"""
|
starts with http it is returned as-is"""
|
||||||
if not self.icon:
|
if not self.icon:
|
||||||
|
@ -747,14 +748,12 @@ class AuthenticatedSession(ExpiringModel):
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def from_request(request: HttpRequest, user: User) -> Optional["AuthenticatedSession"]:
|
def from_request(request: HttpRequest, user: User) -> Optional["AuthenticatedSession"]:
|
||||||
"""Create a new session from a http request"""
|
"""Create a new session from a http request"""
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
|
|
||||||
if not hasattr(request, "session") or not request.session.session_key:
|
if not hasattr(request, "session") or not request.session.session_key:
|
||||||
return None
|
return None
|
||||||
return AuthenticatedSession(
|
return AuthenticatedSession(
|
||||||
session_key=request.session.session_key,
|
session_key=request.session.session_key,
|
||||||
user=user,
|
user=user,
|
||||||
last_ip=ClientIPMiddleware.get_client_ip(request),
|
last_ip=get_client_ip(request),
|
||||||
last_user_agent=request.META.get("HTTP_USER_AGENT", ""),
|
last_user_agent=request.META.get("HTTP_USER_AGENT", ""),
|
||||||
expires=request.session.get_expiry_date(),
|
expires=request.session.get_expiry_date(),
|
||||||
)
|
)
|
||||||
|
|
|
@ -27,7 +27,7 @@ window.authentik.flow = {
|
||||||
|
|
||||||
{% block body %}
|
{% block body %}
|
||||||
<ak-message-container></ak-message-container>
|
<ak-message-container></ak-message-container>
|
||||||
<ak-flow-executor flowSlug="{{ flow.slug }}">
|
<ak-flow-executor>
|
||||||
<ak-loading></ak-loading>
|
<ak-loading></ak-loading>
|
||||||
</ak-flow-executor>
|
</ak-flow-executor>
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
|
|
|
@ -44,14 +44,28 @@
|
||||||
|
|
||||||
{% block body %}
|
{% block body %}
|
||||||
<div class="pf-c-background-image">
|
<div class="pf-c-background-image">
|
||||||
|
<svg xmlns="http://www.w3.org/2000/svg" class="pf-c-background-image__filter" width="0" height="0">
|
||||||
|
<filter id="image_overlay">
|
||||||
|
<feColorMatrix in="SourceGraphic" type="matrix" values="1.3 0 0 0 0 0 1.3 0 0 0 0 0 1.3 0 0 0 0 0 1 0" />
|
||||||
|
<feComponentTransfer color-interpolation-filters="sRGB" result="duotone">
|
||||||
|
<feFuncR type="table" tableValues="0.086274509803922 0.43921568627451"></feFuncR>
|
||||||
|
<feFuncG type="table" tableValues="0.086274509803922 0.43921568627451"></feFuncG>
|
||||||
|
<feFuncB type="table" tableValues="0.086274509803922 0.43921568627451"></feFuncB>
|
||||||
|
<feFuncA type="table" tableValues="0 1"></feFuncA>
|
||||||
|
</feComponentTransfer>
|
||||||
|
</filter>
|
||||||
|
</svg>
|
||||||
</div>
|
</div>
|
||||||
<ak-message-container></ak-message-container>
|
<ak-message-container></ak-message-container>
|
||||||
<div class="pf-c-login stacked">
|
<div class="pf-c-login">
|
||||||
<div class="ak-login-container">
|
<div class="ak-login-container">
|
||||||
<main class="pf-c-login__main">
|
<header class="pf-c-login__header">
|
||||||
<div class="pf-c-login__main-header pf-c-brand ak-brand">
|
<div class="pf-c-brand ak-brand">
|
||||||
<img src="{{ tenant.branding_logo }}" alt="authentik Logo" />
|
<img src="{{ tenant.branding_logo }}" alt="authentik Logo" />
|
||||||
</div>
|
</div>
|
||||||
|
</header>
|
||||||
|
{% block main_container %}
|
||||||
|
<main class="pf-c-login__main">
|
||||||
<header class="pf-c-login__main-header">
|
<header class="pf-c-login__main-header">
|
||||||
<h1 class="pf-c-title pf-m-3xl">
|
<h1 class="pf-c-title pf-m-3xl">
|
||||||
{% block card_title %}
|
{% block card_title %}
|
||||||
|
@ -63,6 +77,7 @@
|
||||||
{% endblock %}
|
{% endblock %}
|
||||||
</div>
|
</div>
|
||||||
</main>
|
</main>
|
||||||
|
{% endblock %}
|
||||||
<footer class="pf-c-login__footer">
|
<footer class="pf-c-login__footer">
|
||||||
<ul class="pf-c-list pf-m-inline">
|
<ul class="pf-c-list pf-m-inline">
|
||||||
{% for link in footer_links %}
|
{% for link in footer_links %}
|
||||||
|
|
|
@ -22,7 +22,6 @@ class InterfaceView(TemplateView):
|
||||||
kwargs["version_family"] = f"{LOCAL_VERSION.major}.{LOCAL_VERSION.minor}"
|
kwargs["version_family"] = f"{LOCAL_VERSION.major}.{LOCAL_VERSION.minor}"
|
||||||
kwargs["version_subdomain"] = f"version-{LOCAL_VERSION.major}-{LOCAL_VERSION.minor}"
|
kwargs["version_subdomain"] = f"version-{LOCAL_VERSION.major}-{LOCAL_VERSION.minor}"
|
||||||
kwargs["build"] = get_build_hash()
|
kwargs["build"] = get_build_hash()
|
||||||
kwargs["url_kwargs"] = self.kwargs
|
|
||||||
return super().get_context_data(**kwargs)
|
return super().get_context_data(**kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -2,11 +2,9 @@
|
||||||
from datetime import datetime, timedelta
|
from datetime import datetime, timedelta
|
||||||
|
|
||||||
from django.utils.timezone import now
|
from django.utils.timezone import now
|
||||||
from django.utils.translation import gettext as _
|
|
||||||
from drf_spectacular.types import OpenApiTypes
|
from drf_spectacular.types import OpenApiTypes
|
||||||
from drf_spectacular.utils import extend_schema, inline_serializer
|
from drf_spectacular.utils import extend_schema, inline_serializer
|
||||||
from rest_framework.decorators import action
|
from rest_framework.decorators import action
|
||||||
from rest_framework.exceptions import ValidationError
|
|
||||||
from rest_framework.fields import BooleanField, CharField, DateTimeField, IntegerField
|
from rest_framework.fields import BooleanField, CharField, DateTimeField, IntegerField
|
||||||
from rest_framework.permissions import IsAuthenticated
|
from rest_framework.permissions import IsAuthenticated
|
||||||
from rest_framework.request import Request
|
from rest_framework.request import Request
|
||||||
|
@ -22,18 +20,6 @@ from authentik.enterprise.models import License, LicenseKey
|
||||||
from authentik.root.install_id import get_install_id
|
from authentik.root.install_id import get_install_id
|
||||||
|
|
||||||
|
|
||||||
class EnterpriseRequiredMixin:
|
|
||||||
"""Mixin to validate that a valid enterprise license
|
|
||||||
exists before allowing to safe the object"""
|
|
||||||
|
|
||||||
def validate(self, attrs: dict) -> dict:
|
|
||||||
"""Check that a valid license exists"""
|
|
||||||
total = LicenseKey.get_total()
|
|
||||||
if not total.is_valid():
|
|
||||||
raise ValidationError(_("Enterprise is required to create/update this object."))
|
|
||||||
return super().validate(attrs)
|
|
||||||
|
|
||||||
|
|
||||||
class LicenseSerializer(ModelSerializer):
|
class LicenseSerializer(ModelSerializer):
|
||||||
"""License Serializer"""
|
"""License Serializer"""
|
||||||
|
|
||||||
|
|
|
@ -2,11 +2,7 @@
|
||||||
from authentik.blueprints.apps import ManagedAppConfig
|
from authentik.blueprints.apps import ManagedAppConfig
|
||||||
|
|
||||||
|
|
||||||
class EnterpriseConfig(ManagedAppConfig):
|
class AuthentikEnterpriseConfig(ManagedAppConfig):
|
||||||
"""Base app config for all enterprise apps"""
|
|
||||||
|
|
||||||
|
|
||||||
class AuthentikEnterpriseConfig(EnterpriseConfig):
|
|
||||||
"""Enterprise app config"""
|
"""Enterprise app config"""
|
||||||
|
|
||||||
name = "authentik.enterprise"
|
name = "authentik.enterprise"
|
||||||
|
|
|
@ -1,8 +1,6 @@
|
||||||
"""Enterprise license policies"""
|
"""Enterprise license policies"""
|
||||||
from typing import Optional
|
from typing import Optional
|
||||||
|
|
||||||
from django.utils.translation import gettext_lazy as _
|
|
||||||
|
|
||||||
from authentik.core.models import User, UserTypes
|
from authentik.core.models import User, UserTypes
|
||||||
from authentik.enterprise.models import LicenseKey
|
from authentik.enterprise.models import LicenseKey
|
||||||
from authentik.policies.types import PolicyRequest, PolicyResult
|
from authentik.policies.types import PolicyRequest, PolicyResult
|
||||||
|
@ -15,10 +13,10 @@ class EnterprisePolicyAccessView(PolicyAccessView):
|
||||||
def check_license(self):
|
def check_license(self):
|
||||||
"""Check license"""
|
"""Check license"""
|
||||||
if not LicenseKey.get_total().is_valid():
|
if not LicenseKey.get_total().is_valid():
|
||||||
return PolicyResult(False, _("Enterprise required to access this feature."))
|
return False
|
||||||
if self.request.user.type != UserTypes.INTERNAL:
|
if self.request.user.type != UserTypes.INTERNAL:
|
||||||
return PolicyResult(False, _("Feature only accessible for internal users."))
|
return False
|
||||||
return PolicyResult(True)
|
return True
|
||||||
|
|
||||||
def user_has_access(self, user: Optional[User] = None) -> PolicyResult:
|
def user_has_access(self, user: Optional[User] = None) -> PolicyResult:
|
||||||
user = user or self.request.user
|
user = user or self.request.user
|
||||||
|
@ -26,7 +24,7 @@ class EnterprisePolicyAccessView(PolicyAccessView):
|
||||||
request.http_request = self.request
|
request.http_request = self.request
|
||||||
result = super().user_has_access(user)
|
result = super().user_has_access(user)
|
||||||
enterprise_result = self.check_license()
|
enterprise_result = self.check_license()
|
||||||
if not enterprise_result.passing:
|
if not enterprise_result:
|
||||||
return enterprise_result
|
return enterprise_result
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
|
|
@ -1,135 +0,0 @@
|
||||||
"""RAC Provider API Views"""
|
|
||||||
from typing import Optional
|
|
||||||
|
|
||||||
from django.core.cache import cache
|
|
||||||
from django.db.models import QuerySet
|
|
||||||
from django.urls import reverse
|
|
||||||
from drf_spectacular.types import OpenApiTypes
|
|
||||||
from drf_spectacular.utils import OpenApiParameter, OpenApiResponse, extend_schema
|
|
||||||
from rest_framework.fields import SerializerMethodField
|
|
||||||
from rest_framework.request import Request
|
|
||||||
from rest_framework.response import Response
|
|
||||||
from rest_framework.serializers import ModelSerializer
|
|
||||||
from rest_framework.viewsets import ModelViewSet
|
|
||||||
from structlog.stdlib import get_logger
|
|
||||||
|
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
|
||||||
from authentik.core.models import Provider
|
|
||||||
from authentik.enterprise.api import EnterpriseRequiredMixin
|
|
||||||
from authentik.enterprise.providers.rac.api.providers import RACProviderSerializer
|
|
||||||
from authentik.enterprise.providers.rac.models import Endpoint
|
|
||||||
from authentik.policies.engine import PolicyEngine
|
|
||||||
from authentik.rbac.filters import ObjectFilter
|
|
||||||
|
|
||||||
LOGGER = get_logger()
|
|
||||||
|
|
||||||
|
|
||||||
def user_endpoint_cache_key(user_pk: str) -> str:
|
|
||||||
"""Cache key where endpoint list for user is saved"""
|
|
||||||
return f"goauthentik.io/providers/rac/endpoint_access/{user_pk}"
|
|
||||||
|
|
||||||
|
|
||||||
class EndpointSerializer(EnterpriseRequiredMixin, ModelSerializer):
|
|
||||||
"""Endpoint Serializer"""
|
|
||||||
|
|
||||||
provider_obj = RACProviderSerializer(source="provider", read_only=True)
|
|
||||||
launch_url = SerializerMethodField()
|
|
||||||
|
|
||||||
def get_launch_url(self, endpoint: Endpoint) -> Optional[str]:
|
|
||||||
"""Build actual launch URL (the provider itself does not have one, just
|
|
||||||
individual endpoints)"""
|
|
||||||
try:
|
|
||||||
# pylint: disable=no-member
|
|
||||||
return reverse(
|
|
||||||
"authentik_providers_rac:start",
|
|
||||||
kwargs={"app": endpoint.provider.application.slug, "endpoint": endpoint.pk},
|
|
||||||
)
|
|
||||||
except Provider.application.RelatedObjectDoesNotExist:
|
|
||||||
return None
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
model = Endpoint
|
|
||||||
fields = [
|
|
||||||
"pk",
|
|
||||||
"name",
|
|
||||||
"provider",
|
|
||||||
"provider_obj",
|
|
||||||
"protocol",
|
|
||||||
"host",
|
|
||||||
"settings",
|
|
||||||
"property_mappings",
|
|
||||||
"auth_mode",
|
|
||||||
"launch_url",
|
|
||||||
"maximum_connections",
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
class EndpointViewSet(UsedByMixin, ModelViewSet):
|
|
||||||
"""Endpoint Viewset"""
|
|
||||||
|
|
||||||
queryset = Endpoint.objects.all()
|
|
||||||
serializer_class = EndpointSerializer
|
|
||||||
filterset_fields = ["name", "provider"]
|
|
||||||
search_fields = ["name", "protocol"]
|
|
||||||
ordering = ["name", "protocol"]
|
|
||||||
|
|
||||||
def _filter_queryset_for_list(self, queryset: QuerySet) -> QuerySet:
|
|
||||||
"""Custom filter_queryset method which ignores guardian, but still supports sorting"""
|
|
||||||
for backend in list(self.filter_backends):
|
|
||||||
if backend == ObjectFilter:
|
|
||||||
continue
|
|
||||||
queryset = backend().filter_queryset(self.request, queryset, self)
|
|
||||||
return queryset
|
|
||||||
|
|
||||||
def _get_allowed_endpoints(self, queryset: QuerySet) -> list[Endpoint]:
|
|
||||||
endpoints = []
|
|
||||||
for endpoint in queryset:
|
|
||||||
engine = PolicyEngine(endpoint, self.request.user, self.request)
|
|
||||||
engine.build()
|
|
||||||
if engine.passing:
|
|
||||||
endpoints.append(endpoint)
|
|
||||||
return endpoints
|
|
||||||
|
|
||||||
@extend_schema(
|
|
||||||
parameters=[
|
|
||||||
OpenApiParameter(
|
|
||||||
"search",
|
|
||||||
OpenApiTypes.STR,
|
|
||||||
),
|
|
||||||
OpenApiParameter(
|
|
||||||
name="superuser_full_list",
|
|
||||||
location=OpenApiParameter.QUERY,
|
|
||||||
type=OpenApiTypes.BOOL,
|
|
||||||
),
|
|
||||||
],
|
|
||||||
responses={
|
|
||||||
200: EndpointSerializer(many=True),
|
|
||||||
400: OpenApiResponse(description="Bad request"),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
def list(self, request: Request, *args, **kwargs) -> Response:
|
|
||||||
"""List accessible endpoints"""
|
|
||||||
should_cache = request.GET.get("search", "") == ""
|
|
||||||
|
|
||||||
superuser_full_list = str(request.GET.get("superuser_full_list", "false")).lower() == "true"
|
|
||||||
if superuser_full_list and request.user.is_superuser:
|
|
||||||
return super().list(request)
|
|
||||||
|
|
||||||
queryset = self._filter_queryset_for_list(self.get_queryset())
|
|
||||||
self.paginate_queryset(queryset)
|
|
||||||
|
|
||||||
allowed_endpoints = []
|
|
||||||
if not should_cache:
|
|
||||||
allowed_endpoints = self._get_allowed_endpoints(queryset)
|
|
||||||
if should_cache:
|
|
||||||
allowed_endpoints = cache.get(user_endpoint_cache_key(self.request.user.pk))
|
|
||||||
if not allowed_endpoints:
|
|
||||||
LOGGER.debug("Caching allowed endpoint list")
|
|
||||||
allowed_endpoints = self._get_allowed_endpoints(queryset)
|
|
||||||
cache.set(
|
|
||||||
user_endpoint_cache_key(self.request.user.pk),
|
|
||||||
allowed_endpoints,
|
|
||||||
timeout=86400,
|
|
||||||
)
|
|
||||||
serializer = self.get_serializer(allowed_endpoints, many=True)
|
|
||||||
return self.get_paginated_response(serializer.data)
|
|
|
@ -1,36 +0,0 @@
|
||||||
"""RAC Provider API Views"""
|
|
||||||
from rest_framework.fields import CharField
|
|
||||||
from rest_framework.viewsets import ModelViewSet
|
|
||||||
|
|
||||||
from authentik.core.api.propertymappings import PropertyMappingSerializer
|
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
|
||||||
from authentik.core.api.utils import JSONDictField
|
|
||||||
from authentik.enterprise.api import EnterpriseRequiredMixin
|
|
||||||
from authentik.enterprise.providers.rac.models import RACPropertyMapping
|
|
||||||
|
|
||||||
|
|
||||||
class RACPropertyMappingSerializer(EnterpriseRequiredMixin, PropertyMappingSerializer):
|
|
||||||
"""RACPropertyMapping Serializer"""
|
|
||||||
|
|
||||||
static_settings = JSONDictField()
|
|
||||||
expression = CharField(allow_blank=True, required=False)
|
|
||||||
|
|
||||||
def validate_expression(self, expression: str) -> str:
|
|
||||||
"""Test Syntax"""
|
|
||||||
if expression == "":
|
|
||||||
return expression
|
|
||||||
return super().validate_expression(expression)
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
model = RACPropertyMapping
|
|
||||||
fields = PropertyMappingSerializer.Meta.fields + ["static_settings"]
|
|
||||||
|
|
||||||
|
|
||||||
class RACPropertyMappingViewSet(UsedByMixin, ModelViewSet):
|
|
||||||
"""RACPropertyMapping Viewset"""
|
|
||||||
|
|
||||||
queryset = RACPropertyMapping.objects.all()
|
|
||||||
serializer_class = RACPropertyMappingSerializer
|
|
||||||
search_fields = ["name"]
|
|
||||||
ordering = ["name"]
|
|
||||||
filterset_fields = ["name", "managed"]
|
|
|
@ -1,32 +0,0 @@
|
||||||
"""RAC Provider API Views"""
|
|
||||||
from rest_framework.fields import CharField, ListField
|
|
||||||
from rest_framework.viewsets import ModelViewSet
|
|
||||||
|
|
||||||
from authentik.core.api.providers import ProviderSerializer
|
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
|
||||||
from authentik.enterprise.api import EnterpriseRequiredMixin
|
|
||||||
from authentik.enterprise.providers.rac.models import RACProvider
|
|
||||||
|
|
||||||
|
|
||||||
class RACProviderSerializer(EnterpriseRequiredMixin, ProviderSerializer):
|
|
||||||
"""RACProvider Serializer"""
|
|
||||||
|
|
||||||
outpost_set = ListField(child=CharField(), read_only=True, source="outpost_set.all")
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
model = RACProvider
|
|
||||||
fields = ProviderSerializer.Meta.fields + ["settings", "outpost_set", "connection_expiry"]
|
|
||||||
extra_kwargs = ProviderSerializer.Meta.extra_kwargs
|
|
||||||
|
|
||||||
|
|
||||||
class RACProviderViewSet(UsedByMixin, ModelViewSet):
|
|
||||||
"""RACProvider Viewset"""
|
|
||||||
|
|
||||||
queryset = RACProvider.objects.all()
|
|
||||||
serializer_class = RACProviderSerializer
|
|
||||||
filterset_fields = {
|
|
||||||
"application": ["isnull"],
|
|
||||||
"name": ["iexact"],
|
|
||||||
}
|
|
||||||
search_fields = ["name"]
|
|
||||||
ordering = ["name"]
|
|
|
@ -1,17 +0,0 @@
|
||||||
"""RAC app config"""
|
|
||||||
from authentik.enterprise.apps import EnterpriseConfig
|
|
||||||
|
|
||||||
|
|
||||||
class AuthentikEnterpriseProviderRAC(EnterpriseConfig):
|
|
||||||
"""authentik enterprise rac app config"""
|
|
||||||
|
|
||||||
name = "authentik.enterprise.providers.rac"
|
|
||||||
label = "authentik_providers_rac"
|
|
||||||
verbose_name = "authentik Enterprise.Providers.RAC"
|
|
||||||
default = True
|
|
||||||
mountpoint = ""
|
|
||||||
ws_mountpoint = "authentik.enterprise.providers.rac.urls"
|
|
||||||
|
|
||||||
def reconcile_load_rac_signals(self):
|
|
||||||
"""Load rac signals"""
|
|
||||||
self.import_module("authentik.enterprise.providers.rac.signals")
|
|
|
@ -1,163 +0,0 @@
|
||||||
"""RAC Client consumer"""
|
|
||||||
from asgiref.sync import async_to_sync
|
|
||||||
from channels.db import database_sync_to_async
|
|
||||||
from channels.exceptions import ChannelFull, DenyConnection
|
|
||||||
from channels.generic.websocket import AsyncWebsocketConsumer
|
|
||||||
from django.http.request import QueryDict
|
|
||||||
from structlog.stdlib import BoundLogger, get_logger
|
|
||||||
|
|
||||||
from authentik.enterprise.providers.rac.models import ConnectionToken, RACProvider
|
|
||||||
from authentik.outposts.consumer import OUTPOST_GROUP_INSTANCE
|
|
||||||
from authentik.outposts.models import Outpost, OutpostState, OutpostType
|
|
||||||
|
|
||||||
# Global broadcast group, which messages are sent to when the outpost connects back
|
|
||||||
# to authentik for a specific connection
|
|
||||||
# The `RACClientConsumer` consumer adds itself to this group on connection,
|
|
||||||
# and removes itself once it has been assigned a specific outpost channel
|
|
||||||
RAC_CLIENT_GROUP = "group_enterprise_rac_client"
|
|
||||||
# A group for all connections in a given authentik session ID
|
|
||||||
# A disconnect message is sent to this group when the session expires/is deleted
|
|
||||||
RAC_CLIENT_GROUP_SESSION = "group_enterprise_rac_client_%(session)s"
|
|
||||||
# A group for all connections with a specific token, which in almost all cases
|
|
||||||
# is just one connection, however this is used to disconnect the connection
|
|
||||||
# when the token is deleted
|
|
||||||
RAC_CLIENT_GROUP_TOKEN = "group_enterprise_rac_token_%(token)s" # nosec
|
|
||||||
|
|
||||||
# Step 1: Client connects to this websocket endpoint
|
|
||||||
# Step 2: We prepare all the connection args for Guac
|
|
||||||
# Step 3: Send a websocket message to a single outpost that has this provider assigned
|
|
||||||
# (Currently sending to all of them)
|
|
||||||
# (Should probably do different load balancing algorithms)
|
|
||||||
# Step 4: Outpost creates a websocket connection back to authentik
|
|
||||||
# with /ws/outpost_rac/<our_channel_id>/
|
|
||||||
# Step 5: This consumer transfers data between the two channels
|
|
||||||
|
|
||||||
|
|
||||||
class RACClientConsumer(AsyncWebsocketConsumer):
|
|
||||||
"""RAC client consumer the browser connects to"""
|
|
||||||
|
|
||||||
dest_channel_id: str = ""
|
|
||||||
provider: RACProvider
|
|
||||||
token: ConnectionToken
|
|
||||||
logger: BoundLogger
|
|
||||||
|
|
||||||
async def connect(self):
|
|
||||||
await self.accept("guacamole")
|
|
||||||
await self.channel_layer.group_add(RAC_CLIENT_GROUP, self.channel_name)
|
|
||||||
await self.channel_layer.group_add(
|
|
||||||
RAC_CLIENT_GROUP_SESSION % {"session": self.scope["session"].session_key},
|
|
||||||
self.channel_name,
|
|
||||||
)
|
|
||||||
await self.init_outpost_connection()
|
|
||||||
|
|
||||||
async def disconnect(self, code):
|
|
||||||
self.logger.debug("Disconnecting")
|
|
||||||
# Tell the outpost we're disconnecting
|
|
||||||
await self.channel_layer.send(
|
|
||||||
self.dest_channel_id,
|
|
||||||
{
|
|
||||||
"type": "event.disconnect",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
@database_sync_to_async
|
|
||||||
def init_outpost_connection(self):
|
|
||||||
"""Initialize guac connection settings"""
|
|
||||||
self.token = ConnectionToken.filter_not_expired(
|
|
||||||
token=self.scope["url_route"]["kwargs"]["token"]
|
|
||||||
).first()
|
|
||||||
if not self.token:
|
|
||||||
raise DenyConnection()
|
|
||||||
self.provider = self.token.provider
|
|
||||||
params = self.token.get_settings()
|
|
||||||
self.logger = get_logger().bind(
|
|
||||||
endpoint=self.token.endpoint.name, user=self.scope["user"].username
|
|
||||||
)
|
|
||||||
msg = {
|
|
||||||
"type": "event.provider.specific",
|
|
||||||
"sub_type": "init_connection",
|
|
||||||
"dest_channel_id": self.channel_name,
|
|
||||||
"params": params,
|
|
||||||
"protocol": self.token.endpoint.protocol,
|
|
||||||
}
|
|
||||||
query = QueryDict(self.scope["query_string"].decode())
|
|
||||||
for key in ["screen_width", "screen_height", "screen_dpi", "audio"]:
|
|
||||||
value = query.get(key, None)
|
|
||||||
if not value:
|
|
||||||
continue
|
|
||||||
msg[key] = str(value)
|
|
||||||
outposts = Outpost.objects.filter(
|
|
||||||
type=OutpostType.RAC,
|
|
||||||
providers__in=[self.provider],
|
|
||||||
)
|
|
||||||
if not outposts.exists():
|
|
||||||
self.logger.warning("Provider has no outpost")
|
|
||||||
raise DenyConnection()
|
|
||||||
for outpost in outposts:
|
|
||||||
# Sort all states for the outpost by connection count
|
|
||||||
states = sorted(
|
|
||||||
OutpostState.for_outpost(outpost),
|
|
||||||
key=lambda state: int(state.args.get("active_connections", 0)),
|
|
||||||
)
|
|
||||||
if len(states) < 1:
|
|
||||||
continue
|
|
||||||
self.logger.debug("Sending out connection broadcast")
|
|
||||||
async_to_sync(self.channel_layer.group_send)(
|
|
||||||
OUTPOST_GROUP_INSTANCE % {"outpost_pk": str(outpost.pk), "instance": states[0].uid},
|
|
||||||
msg,
|
|
||||||
)
|
|
||||||
|
|
||||||
async def receive(self, text_data=None, bytes_data=None):
|
|
||||||
"""Mirror data received from client to the dest_channel_id
|
|
||||||
which is the channel talking to guacd"""
|
|
||||||
if self.dest_channel_id == "":
|
|
||||||
return
|
|
||||||
if self.token.is_expired:
|
|
||||||
await self.event_disconnect({"reason": "token_expiry"})
|
|
||||||
return
|
|
||||||
try:
|
|
||||||
await self.channel_layer.send(
|
|
||||||
self.dest_channel_id,
|
|
||||||
{
|
|
||||||
"type": "event.send",
|
|
||||||
"text_data": text_data,
|
|
||||||
"bytes_data": bytes_data,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
except ChannelFull:
|
|
||||||
pass
|
|
||||||
|
|
||||||
async def event_outpost_connected(self, event: dict):
|
|
||||||
"""Handle event broadcasted from outpost consumer, and check if they
|
|
||||||
created a connection for us"""
|
|
||||||
outpost_channel = event.get("outpost_channel")
|
|
||||||
if event.get("client_channel") != self.channel_name:
|
|
||||||
return
|
|
||||||
if self.dest_channel_id != "":
|
|
||||||
# We've already selected an outpost channel, so tell the other channel to disconnect
|
|
||||||
# This should never happen since we remove ourselves from the broadcast group
|
|
||||||
await self.channel_layer.send(
|
|
||||||
outpost_channel,
|
|
||||||
{
|
|
||||||
"type": "event.disconnect",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
return
|
|
||||||
self.logger.debug("Connected to a single outpost instance")
|
|
||||||
self.dest_channel_id = outpost_channel
|
|
||||||
# Since we have a specific outpost channel now, we can remove
|
|
||||||
# ourselves from the global broadcast group
|
|
||||||
await self.channel_layer.group_discard(RAC_CLIENT_GROUP, self.channel_name)
|
|
||||||
|
|
||||||
async def event_send(self, event: dict):
|
|
||||||
"""Handler called by outpost websocket that sends data to this specific
|
|
||||||
client connection"""
|
|
||||||
if self.token.is_expired:
|
|
||||||
await self.event_disconnect({"reason": "token_expiry"})
|
|
||||||
return
|
|
||||||
await self.send(text_data=event.get("text_data"), bytes_data=event.get("bytes_data"))
|
|
||||||
|
|
||||||
async def event_disconnect(self, event: dict):
|
|
||||||
"""Disconnect when the session ends"""
|
|
||||||
self.logger.info("Disconnecting RAC connection", reason=event.get("reason"))
|
|
||||||
await self.close()
|
|
|
@ -1,48 +0,0 @@
|
||||||
"""RAC consumer"""
|
|
||||||
from channels.exceptions import ChannelFull
|
|
||||||
from channels.generic.websocket import AsyncWebsocketConsumer
|
|
||||||
|
|
||||||
from authentik.enterprise.providers.rac.consumer_client import RAC_CLIENT_GROUP
|
|
||||||
|
|
||||||
|
|
||||||
class RACOutpostConsumer(AsyncWebsocketConsumer):
|
|
||||||
"""Consumer the outpost connects to, to send specific data back to a client connection"""
|
|
||||||
|
|
||||||
dest_channel_id: str
|
|
||||||
|
|
||||||
async def connect(self):
|
|
||||||
self.dest_channel_id = self.scope["url_route"]["kwargs"]["channel"]
|
|
||||||
await self.accept()
|
|
||||||
await self.channel_layer.group_send(
|
|
||||||
RAC_CLIENT_GROUP,
|
|
||||||
{
|
|
||||||
"type": "event.outpost.connected",
|
|
||||||
"outpost_channel": self.channel_name,
|
|
||||||
"client_channel": self.dest_channel_id,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
async def receive(self, text_data=None, bytes_data=None):
|
|
||||||
"""Mirror data received from guacd running in the outpost
|
|
||||||
to the dest_channel_id which is the channel talking to the browser"""
|
|
||||||
try:
|
|
||||||
await self.channel_layer.send(
|
|
||||||
self.dest_channel_id,
|
|
||||||
{
|
|
||||||
"type": "event.send",
|
|
||||||
"text_data": text_data,
|
|
||||||
"bytes_data": bytes_data,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
except ChannelFull:
|
|
||||||
pass
|
|
||||||
|
|
||||||
async def event_send(self, event: dict):
|
|
||||||
"""Handler called by client websocket that sends data to this specific
|
|
||||||
outpost connection"""
|
|
||||||
await self.send(text_data=event.get("text_data"), bytes_data=event.get("bytes_data"))
|
|
||||||
|
|
||||||
async def event_disconnect(self, event: dict):
|
|
||||||
"""Tell outpost we're about to disconnect"""
|
|
||||||
await self.send(text_data="0.authentik.disconnect")
|
|
||||||
await self.close()
|
|
|
@ -1,11 +0,0 @@
|
||||||
"""RAC Provider Docker Controller"""
|
|
||||||
from authentik.outposts.controllers.docker import DockerController
|
|
||||||
from authentik.outposts.models import DockerServiceConnection, Outpost
|
|
||||||
|
|
||||||
|
|
||||||
class RACDockerController(DockerController):
|
|
||||||
"""RAC Provider Docker Controller"""
|
|
||||||
|
|
||||||
def __init__(self, outpost: Outpost, connection: DockerServiceConnection):
|
|
||||||
super().__init__(outpost, connection)
|
|
||||||
self.deployment_ports = []
|
|
|
@ -1,13 +0,0 @@
|
||||||
"""RAC Provider Kubernetes Controller"""
|
|
||||||
from authentik.outposts.controllers.k8s.service import ServiceReconciler
|
|
||||||
from authentik.outposts.controllers.kubernetes import KubernetesController
|
|
||||||
from authentik.outposts.models import KubernetesServiceConnection, Outpost
|
|
||||||
|
|
||||||
|
|
||||||
class RACKubernetesController(KubernetesController):
|
|
||||||
"""RAC Provider Kubernetes Controller"""
|
|
||||||
|
|
||||||
def __init__(self, outpost: Outpost, connection: KubernetesServiceConnection):
|
|
||||||
super().__init__(outpost, connection)
|
|
||||||
self.deployment_ports = []
|
|
||||||
del self.reconcilers[ServiceReconciler.reconciler_name()]
|
|
|
@ -1,164 +0,0 @@
|
||||||
# Generated by Django 4.2.8 on 2023-12-29 15:58
|
|
||||||
|
|
||||||
import uuid
|
|
||||||
|
|
||||||
import django.db.models.deletion
|
|
||||||
from django.db import migrations, models
|
|
||||||
|
|
||||||
import authentik.core.models
|
|
||||||
import authentik.lib.utils.time
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
initial = True
|
|
||||||
|
|
||||||
dependencies = [
|
|
||||||
("authentik_policies", "0011_policybinding_failure_result_and_more"),
|
|
||||||
("authentik_core", "0032_group_roles"),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.CreateModel(
|
|
||||||
name="RACPropertyMapping",
|
|
||||||
fields=[
|
|
||||||
(
|
|
||||||
"propertymapping_ptr",
|
|
||||||
models.OneToOneField(
|
|
||||||
auto_created=True,
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
parent_link=True,
|
|
||||||
primary_key=True,
|
|
||||||
serialize=False,
|
|
||||||
to="authentik_core.propertymapping",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
("static_settings", models.JSONField(default=dict)),
|
|
||||||
],
|
|
||||||
options={
|
|
||||||
"verbose_name": "RAC Property Mapping",
|
|
||||||
"verbose_name_plural": "RAC Property Mappings",
|
|
||||||
},
|
|
||||||
bases=("authentik_core.propertymapping",),
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name="RACProvider",
|
|
||||||
fields=[
|
|
||||||
(
|
|
||||||
"provider_ptr",
|
|
||||||
models.OneToOneField(
|
|
||||||
auto_created=True,
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
parent_link=True,
|
|
||||||
primary_key=True,
|
|
||||||
serialize=False,
|
|
||||||
to="authentik_core.provider",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
("settings", models.JSONField(default=dict)),
|
|
||||||
(
|
|
||||||
"auth_mode",
|
|
||||||
models.TextField(
|
|
||||||
choices=[("static", "Static"), ("prompt", "Prompt")], default="prompt"
|
|
||||||
),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"connection_expiry",
|
|
||||||
models.TextField(
|
|
||||||
default="hours=8",
|
|
||||||
help_text="Determines how long a session lasts. Default of 0 means that the sessions lasts until the browser is closed. (Format: hours=-1;minutes=-2;seconds=-3)",
|
|
||||||
validators=[authentik.lib.utils.time.timedelta_string_validator],
|
|
||||||
),
|
|
||||||
),
|
|
||||||
],
|
|
||||||
options={
|
|
||||||
"verbose_name": "RAC Provider",
|
|
||||||
"verbose_name_plural": "RAC Providers",
|
|
||||||
},
|
|
||||||
bases=("authentik_core.provider",),
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name="Endpoint",
|
|
||||||
fields=[
|
|
||||||
(
|
|
||||||
"policybindingmodel_ptr",
|
|
||||||
models.OneToOneField(
|
|
||||||
auto_created=True,
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
parent_link=True,
|
|
||||||
primary_key=True,
|
|
||||||
serialize=False,
|
|
||||||
to="authentik_policies.policybindingmodel",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
("name", models.TextField()),
|
|
||||||
("host", models.TextField()),
|
|
||||||
(
|
|
||||||
"protocol",
|
|
||||||
models.TextField(choices=[("rdp", "Rdp"), ("vnc", "Vnc"), ("ssh", "Ssh")]),
|
|
||||||
),
|
|
||||||
("settings", models.JSONField(default=dict)),
|
|
||||||
(
|
|
||||||
"auth_mode",
|
|
||||||
models.TextField(choices=[("static", "Static"), ("prompt", "Prompt")]),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"property_mappings",
|
|
||||||
models.ManyToManyField(
|
|
||||||
blank=True, default=None, to="authentik_core.propertymapping"
|
|
||||||
),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"provider",
|
|
||||||
models.ForeignKey(
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
to="authentik_providers_rac.racprovider",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
],
|
|
||||||
options={
|
|
||||||
"verbose_name": "RAC Endpoint",
|
|
||||||
"verbose_name_plural": "RAC Endpoints",
|
|
||||||
},
|
|
||||||
bases=("authentik_policies.policybindingmodel", models.Model),
|
|
||||||
),
|
|
||||||
migrations.CreateModel(
|
|
||||||
name="ConnectionToken",
|
|
||||||
fields=[
|
|
||||||
(
|
|
||||||
"expires",
|
|
||||||
models.DateTimeField(default=authentik.core.models.default_token_duration),
|
|
||||||
),
|
|
||||||
("expiring", models.BooleanField(default=True)),
|
|
||||||
(
|
|
||||||
"connection_token_uuid",
|
|
||||||
models.UUIDField(default=uuid.uuid4, primary_key=True, serialize=False),
|
|
||||||
),
|
|
||||||
("token", models.TextField(default=authentik.core.models.default_token_key)),
|
|
||||||
("settings", models.JSONField(default=dict)),
|
|
||||||
(
|
|
||||||
"endpoint",
|
|
||||||
models.ForeignKey(
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
to="authentik_providers_rac.endpoint",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"provider",
|
|
||||||
models.ForeignKey(
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
to="authentik_providers_rac.racprovider",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"session",
|
|
||||||
models.ForeignKey(
|
|
||||||
on_delete=django.db.models.deletion.CASCADE,
|
|
||||||
to="authentik_core.authenticatedsession",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
],
|
|
||||||
options={
|
|
||||||
"abstract": False,
|
|
||||||
},
|
|
||||||
),
|
|
||||||
]
|
|
|
@ -1,17 +0,0 @@
|
||||||
# Generated by Django 5.0 on 2024-01-03 23:44
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
dependencies = [
|
|
||||||
("authentik_providers_rac", "0001_initial"),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AddField(
|
|
||||||
model_name="endpoint",
|
|
||||||
name="maximum_connections",
|
|
||||||
field=models.IntegerField(default=1),
|
|
||||||
),
|
|
||||||
]
|
|
|
@ -1,192 +0,0 @@
|
||||||
"""RAC Models"""
|
|
||||||
from typing import Optional
|
|
||||||
from uuid import uuid4
|
|
||||||
|
|
||||||
from deepmerge import always_merger
|
|
||||||
from django.db import models
|
|
||||||
from django.db.models import QuerySet
|
|
||||||
from django.utils.translation import gettext as _
|
|
||||||
from rest_framework.serializers import Serializer
|
|
||||||
from structlog.stdlib import get_logger
|
|
||||||
|
|
||||||
from authentik.core.exceptions import PropertyMappingExpressionException
|
|
||||||
from authentik.core.models import ExpiringModel, PropertyMapping, Provider, default_token_key
|
|
||||||
from authentik.events.models import Event, EventAction
|
|
||||||
from authentik.lib.models import SerializerModel
|
|
||||||
from authentik.lib.utils.time import timedelta_string_validator
|
|
||||||
from authentik.policies.models import PolicyBindingModel
|
|
||||||
|
|
||||||
LOGGER = get_logger()
|
|
||||||
|
|
||||||
|
|
||||||
class Protocols(models.TextChoices):
|
|
||||||
"""Supported protocols"""
|
|
||||||
|
|
||||||
RDP = "rdp"
|
|
||||||
VNC = "vnc"
|
|
||||||
SSH = "ssh"
|
|
||||||
|
|
||||||
|
|
||||||
class AuthenticationMode(models.TextChoices):
|
|
||||||
"""Authentication modes"""
|
|
||||||
|
|
||||||
STATIC = "static"
|
|
||||||
PROMPT = "prompt"
|
|
||||||
|
|
||||||
|
|
||||||
class RACProvider(Provider):
|
|
||||||
"""Remotely access computers/servers via RDP/SSH/VNC."""
|
|
||||||
|
|
||||||
settings = models.JSONField(default=dict)
|
|
||||||
auth_mode = models.TextField(
|
|
||||||
choices=AuthenticationMode.choices, default=AuthenticationMode.PROMPT
|
|
||||||
)
|
|
||||||
connection_expiry = models.TextField(
|
|
||||||
default="hours=8",
|
|
||||||
validators=[timedelta_string_validator],
|
|
||||||
help_text=_(
|
|
||||||
"Determines how long a session lasts. Default of 0 means "
|
|
||||||
"that the sessions lasts until the browser is closed. "
|
|
||||||
"(Format: hours=-1;minutes=-2;seconds=-3)"
|
|
||||||
),
|
|
||||||
)
|
|
||||||
|
|
||||||
@property
|
|
||||||
def launch_url(self) -> Optional[str]:
|
|
||||||
"""URL to this provider and initiate authorization for the user.
|
|
||||||
Can return None for providers that are not URL-based"""
|
|
||||||
return "goauthentik.io://providers/rac/launch"
|
|
||||||
|
|
||||||
@property
|
|
||||||
def component(self) -> str:
|
|
||||||
return "ak-provider-rac-form"
|
|
||||||
|
|
||||||
@property
|
|
||||||
def serializer(self) -> type[Serializer]:
|
|
||||||
from authentik.enterprise.providers.rac.api.providers import RACProviderSerializer
|
|
||||||
|
|
||||||
return RACProviderSerializer
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
verbose_name = _("RAC Provider")
|
|
||||||
verbose_name_plural = _("RAC Providers")
|
|
||||||
|
|
||||||
|
|
||||||
class Endpoint(SerializerModel, PolicyBindingModel):
|
|
||||||
"""Remote-accessible endpoint"""
|
|
||||||
|
|
||||||
name = models.TextField()
|
|
||||||
host = models.TextField()
|
|
||||||
protocol = models.TextField(choices=Protocols.choices)
|
|
||||||
settings = models.JSONField(default=dict)
|
|
||||||
auth_mode = models.TextField(choices=AuthenticationMode.choices)
|
|
||||||
provider = models.ForeignKey("RACProvider", on_delete=models.CASCADE)
|
|
||||||
maximum_connections = models.IntegerField(default=1)
|
|
||||||
|
|
||||||
property_mappings = models.ManyToManyField(
|
|
||||||
"authentik_core.PropertyMapping", default=None, blank=True
|
|
||||||
)
|
|
||||||
|
|
||||||
@property
|
|
||||||
def serializer(self) -> type[Serializer]:
|
|
||||||
from authentik.enterprise.providers.rac.api.endpoints import EndpointSerializer
|
|
||||||
|
|
||||||
return EndpointSerializer
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
return f"RAC Endpoint {self.name}"
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
verbose_name = _("RAC Endpoint")
|
|
||||||
verbose_name_plural = _("RAC Endpoints")
|
|
||||||
|
|
||||||
|
|
||||||
class RACPropertyMapping(PropertyMapping):
|
|
||||||
"""Configure settings for remote access endpoints."""
|
|
||||||
|
|
||||||
static_settings = models.JSONField(default=dict)
|
|
||||||
|
|
||||||
@property
|
|
||||||
def component(self) -> str:
|
|
||||||
return "ak-property-mapping-rac-form"
|
|
||||||
|
|
||||||
@property
|
|
||||||
def serializer(self) -> type[Serializer]:
|
|
||||||
from authentik.enterprise.providers.rac.api.property_mappings import (
|
|
||||||
RACPropertyMappingSerializer,
|
|
||||||
)
|
|
||||||
|
|
||||||
return RACPropertyMappingSerializer
|
|
||||||
|
|
||||||
class Meta:
|
|
||||||
verbose_name = _("RAC Property Mapping")
|
|
||||||
verbose_name_plural = _("RAC Property Mappings")
|
|
||||||
|
|
||||||
|
|
||||||
class ConnectionToken(ExpiringModel):
|
|
||||||
"""Token for a single connection to a specified endpoint"""
|
|
||||||
|
|
||||||
connection_token_uuid = models.UUIDField(default=uuid4, primary_key=True)
|
|
||||||
provider = models.ForeignKey(RACProvider, on_delete=models.CASCADE)
|
|
||||||
endpoint = models.ForeignKey(Endpoint, on_delete=models.CASCADE)
|
|
||||||
token = models.TextField(default=default_token_key)
|
|
||||||
settings = models.JSONField(default=dict)
|
|
||||||
session = models.ForeignKey("authentik_core.AuthenticatedSession", on_delete=models.CASCADE)
|
|
||||||
|
|
||||||
def get_settings(self) -> dict:
|
|
||||||
"""Get settings"""
|
|
||||||
default_settings = {}
|
|
||||||
if ":" in self.endpoint.host:
|
|
||||||
host, _, port = self.endpoint.host.partition(":")
|
|
||||||
default_settings["hostname"] = host
|
|
||||||
default_settings["port"] = str(port)
|
|
||||||
else:
|
|
||||||
default_settings["hostname"] = self.endpoint.host
|
|
||||||
default_settings["client-name"] = "authentik"
|
|
||||||
# default_settings["enable-drive"] = "true"
|
|
||||||
# default_settings["drive-name"] = "authentik"
|
|
||||||
settings = {}
|
|
||||||
always_merger.merge(settings, default_settings)
|
|
||||||
always_merger.merge(settings, self.endpoint.provider.settings)
|
|
||||||
always_merger.merge(settings, self.endpoint.settings)
|
|
||||||
always_merger.merge(settings, self.settings)
|
|
||||||
|
|
||||||
def mapping_evaluator(mappings: QuerySet):
|
|
||||||
for mapping in mappings:
|
|
||||||
mapping: RACPropertyMapping
|
|
||||||
if len(mapping.static_settings) > 0:
|
|
||||||
always_merger.merge(settings, mapping.static_settings)
|
|
||||||
continue
|
|
||||||
try:
|
|
||||||
mapping_settings = mapping.evaluate(
|
|
||||||
self.session.user, None, endpoint=self.endpoint, provider=self.provider
|
|
||||||
)
|
|
||||||
always_merger.merge(settings, mapping_settings)
|
|
||||||
except PropertyMappingExpressionException as exc:
|
|
||||||
Event.new(
|
|
||||||
EventAction.CONFIGURATION_ERROR,
|
|
||||||
message=f"Failed to evaluate property-mapping: '{mapping.name}'",
|
|
||||||
provider=self.provider,
|
|
||||||
mapping=mapping,
|
|
||||||
).set_user(self.session.user).save()
|
|
||||||
LOGGER.warning("Failed to evaluate property mapping", exc=exc)
|
|
||||||
|
|
||||||
mapping_evaluator(
|
|
||||||
RACPropertyMapping.objects.filter(provider__in=[self.provider]).order_by("name")
|
|
||||||
)
|
|
||||||
mapping_evaluator(
|
|
||||||
RACPropertyMapping.objects.filter(endpoint__in=[self.endpoint]).order_by("name")
|
|
||||||
)
|
|
||||||
|
|
||||||
settings["drive-path"] = f"/tmp/connection/{self.token}" # nosec
|
|
||||||
settings["create-drive-path"] = "true"
|
|
||||||
# Ensure all values of the settings dict are strings
|
|
||||||
for key, value in settings.items():
|
|
||||||
if isinstance(value, str):
|
|
||||||
continue
|
|
||||||
# Special case for bools
|
|
||||||
if isinstance(value, bool):
|
|
||||||
settings[key] = str(value).lower()
|
|
||||||
continue
|
|
||||||
settings[key] = str(value)
|
|
||||||
return settings
|
|
|
@ -1,54 +0,0 @@
|
||||||
"""RAC Signals"""
|
|
||||||
from asgiref.sync import async_to_sync
|
|
||||||
from channels.layers import get_channel_layer
|
|
||||||
from django.contrib.auth.signals import user_logged_out
|
|
||||||
from django.core.cache import cache
|
|
||||||
from django.db.models import Model
|
|
||||||
from django.db.models.signals import post_save, pre_delete
|
|
||||||
from django.dispatch import receiver
|
|
||||||
from django.http import HttpRequest
|
|
||||||
|
|
||||||
from authentik.core.models import User
|
|
||||||
from authentik.enterprise.providers.rac.api.endpoints import user_endpoint_cache_key
|
|
||||||
from authentik.enterprise.providers.rac.consumer_client import (
|
|
||||||
RAC_CLIENT_GROUP_SESSION,
|
|
||||||
RAC_CLIENT_GROUP_TOKEN,
|
|
||||||
)
|
|
||||||
from authentik.enterprise.providers.rac.models import ConnectionToken, Endpoint
|
|
||||||
|
|
||||||
|
|
||||||
@receiver(user_logged_out)
|
|
||||||
def user_logged_out_session(sender, request: HttpRequest, user: User, **_):
|
|
||||||
"""Disconnect any open RAC connections"""
|
|
||||||
layer = get_channel_layer()
|
|
||||||
async_to_sync(layer.group_send)(
|
|
||||||
RAC_CLIENT_GROUP_SESSION
|
|
||||||
% {
|
|
||||||
"session": request.session.session_key,
|
|
||||||
},
|
|
||||||
{"type": "event.disconnect", "reason": "session_logout"},
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@receiver(pre_delete, sender=ConnectionToken)
|
|
||||||
def pre_delete_connection_token_disconnect(sender, instance: ConnectionToken, **_):
|
|
||||||
"""Disconnect session when connection token is deleted"""
|
|
||||||
layer = get_channel_layer()
|
|
||||||
async_to_sync(layer.group_send)(
|
|
||||||
RAC_CLIENT_GROUP_TOKEN
|
|
||||||
% {
|
|
||||||
"token": instance.token,
|
|
||||||
},
|
|
||||||
{"type": "event.disconnect", "reason": "token_delete"},
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@receiver(post_save, sender=Endpoint)
|
|
||||||
def post_save_application(sender: type[Model], instance, created: bool, **_):
|
|
||||||
"""Clear user's application cache upon application creation"""
|
|
||||||
if not created: # pragma: no cover
|
|
||||||
return
|
|
||||||
|
|
||||||
# Delete user endpoint cache
|
|
||||||
keys = cache.keys(user_endpoint_cache_key("*"))
|
|
||||||
cache.delete_many(keys)
|
|
|
@ -1,18 +0,0 @@
|
||||||
{% extends "base/skeleton.html" %}
|
|
||||||
|
|
||||||
{% load static %}
|
|
||||||
|
|
||||||
{% block head %}
|
|
||||||
<script src="{% static 'dist/enterprise/rac/index.js' %}?version={{ version }}" type="module"></script>
|
|
||||||
<meta name="theme-color" content="#18191a" media="(prefers-color-scheme: dark)">
|
|
||||||
<meta name="theme-color" content="#ffffff" media="(prefers-color-scheme: light)">
|
|
||||||
<link rel="icon" href="{{ tenant.branding_favicon }}">
|
|
||||||
<link rel="shortcut icon" href="{{ tenant.branding_favicon }}">
|
|
||||||
{% include "base/header_js.html" %}
|
|
||||||
{% endblock %}
|
|
||||||
|
|
||||||
{% block body %}
|
|
||||||
<ak-rac token="{{ url_kwargs.token }}" endpointName="{{ token.endpoint.name }}">
|
|
||||||
<ak-loading></ak-loading>
|
|
||||||
</ak-rac>
|
|
||||||
{% endblock %}
|
|
|
@ -1,171 +0,0 @@
|
||||||
"""Test Endpoints API"""
|
|
||||||
|
|
||||||
from django.urls import reverse
|
|
||||||
from rest_framework.test import APITestCase
|
|
||||||
|
|
||||||
from authentik.core.models import Application
|
|
||||||
from authentik.core.tests.utils import create_test_admin_user
|
|
||||||
from authentik.enterprise.providers.rac.models import Endpoint, Protocols, RACProvider
|
|
||||||
from authentik.lib.generators import generate_id
|
|
||||||
from authentik.policies.dummy.models import DummyPolicy
|
|
||||||
from authentik.policies.models import PolicyBinding
|
|
||||||
|
|
||||||
|
|
||||||
class TestEndpointsAPI(APITestCase):
|
|
||||||
"""Test endpoints API"""
|
|
||||||
|
|
||||||
def setUp(self) -> None:
|
|
||||||
self.user = create_test_admin_user()
|
|
||||||
self.provider = RACProvider.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
)
|
|
||||||
self.app = Application.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
slug=generate_id(),
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
self.allowed = Endpoint.objects.create(
|
|
||||||
name=f"a-{generate_id()}",
|
|
||||||
host=generate_id(),
|
|
||||||
protocol=Protocols.RDP,
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
self.denied = Endpoint.objects.create(
|
|
||||||
name=f"b-{generate_id()}",
|
|
||||||
host=generate_id(),
|
|
||||||
protocol=Protocols.RDP,
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
PolicyBinding.objects.create(
|
|
||||||
target=self.denied,
|
|
||||||
policy=DummyPolicy.objects.create(name="deny", result=False, wait_min=1, wait_max=2),
|
|
||||||
order=0,
|
|
||||||
)
|
|
||||||
|
|
||||||
def test_list(self):
|
|
||||||
"""Test list operation without superuser_full_list"""
|
|
||||||
self.client.force_login(self.user)
|
|
||||||
response = self.client.get(reverse("authentik_api:endpoint-list"))
|
|
||||||
self.assertJSONEqual(
|
|
||||||
response.content.decode(),
|
|
||||||
{
|
|
||||||
"pagination": {
|
|
||||||
"next": 0,
|
|
||||||
"previous": 0,
|
|
||||||
"count": 2,
|
|
||||||
"current": 1,
|
|
||||||
"total_pages": 1,
|
|
||||||
"start_index": 1,
|
|
||||||
"end_index": 2,
|
|
||||||
},
|
|
||||||
"results": [
|
|
||||||
{
|
|
||||||
"pk": str(self.allowed.pk),
|
|
||||||
"name": self.allowed.name,
|
|
||||||
"provider": self.provider.pk,
|
|
||||||
"provider_obj": {
|
|
||||||
"pk": self.provider.pk,
|
|
||||||
"name": self.provider.name,
|
|
||||||
"authentication_flow": None,
|
|
||||||
"authorization_flow": None,
|
|
||||||
"property_mappings": [],
|
|
||||||
"connection_expiry": "hours=8",
|
|
||||||
"component": "ak-provider-rac-form",
|
|
||||||
"assigned_application_slug": self.app.slug,
|
|
||||||
"assigned_application_name": self.app.name,
|
|
||||||
"verbose_name": "RAC Provider",
|
|
||||||
"verbose_name_plural": "RAC Providers",
|
|
||||||
"meta_model_name": "authentik_providers_rac.racprovider",
|
|
||||||
"settings": {},
|
|
||||||
"outpost_set": [],
|
|
||||||
},
|
|
||||||
"protocol": "rdp",
|
|
||||||
"host": self.allowed.host,
|
|
||||||
"maximum_connections": 1,
|
|
||||||
"settings": {},
|
|
||||||
"property_mappings": [],
|
|
||||||
"auth_mode": "",
|
|
||||||
"launch_url": f"/application/rac/{self.app.slug}/{str(self.allowed.pk)}/",
|
|
||||||
},
|
|
||||||
],
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
def test_list_superuser_full_list(self):
|
|
||||||
"""Test list operation with superuser_full_list"""
|
|
||||||
self.client.force_login(self.user)
|
|
||||||
response = self.client.get(
|
|
||||||
reverse("authentik_api:endpoint-list") + "?superuser_full_list=true"
|
|
||||||
)
|
|
||||||
self.assertJSONEqual(
|
|
||||||
response.content.decode(),
|
|
||||||
{
|
|
||||||
"pagination": {
|
|
||||||
"next": 0,
|
|
||||||
"previous": 0,
|
|
||||||
"count": 2,
|
|
||||||
"current": 1,
|
|
||||||
"total_pages": 1,
|
|
||||||
"start_index": 1,
|
|
||||||
"end_index": 2,
|
|
||||||
},
|
|
||||||
"results": [
|
|
||||||
{
|
|
||||||
"pk": str(self.allowed.pk),
|
|
||||||
"name": self.allowed.name,
|
|
||||||
"provider": self.provider.pk,
|
|
||||||
"provider_obj": {
|
|
||||||
"pk": self.provider.pk,
|
|
||||||
"name": self.provider.name,
|
|
||||||
"authentication_flow": None,
|
|
||||||
"authorization_flow": None,
|
|
||||||
"property_mappings": [],
|
|
||||||
"component": "ak-provider-rac-form",
|
|
||||||
"assigned_application_slug": self.app.slug,
|
|
||||||
"assigned_application_name": self.app.name,
|
|
||||||
"connection_expiry": "hours=8",
|
|
||||||
"verbose_name": "RAC Provider",
|
|
||||||
"verbose_name_plural": "RAC Providers",
|
|
||||||
"meta_model_name": "authentik_providers_rac.racprovider",
|
|
||||||
"settings": {},
|
|
||||||
"outpost_set": [],
|
|
||||||
},
|
|
||||||
"protocol": "rdp",
|
|
||||||
"host": self.allowed.host,
|
|
||||||
"maximum_connections": 1,
|
|
||||||
"settings": {},
|
|
||||||
"property_mappings": [],
|
|
||||||
"auth_mode": "",
|
|
||||||
"launch_url": f"/application/rac/{self.app.slug}/{str(self.allowed.pk)}/",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"pk": str(self.denied.pk),
|
|
||||||
"name": self.denied.name,
|
|
||||||
"provider": self.provider.pk,
|
|
||||||
"provider_obj": {
|
|
||||||
"pk": self.provider.pk,
|
|
||||||
"name": self.provider.name,
|
|
||||||
"authentication_flow": None,
|
|
||||||
"authorization_flow": None,
|
|
||||||
"property_mappings": [],
|
|
||||||
"component": "ak-provider-rac-form",
|
|
||||||
"assigned_application_slug": self.app.slug,
|
|
||||||
"assigned_application_name": self.app.name,
|
|
||||||
"connection_expiry": "hours=8",
|
|
||||||
"verbose_name": "RAC Provider",
|
|
||||||
"verbose_name_plural": "RAC Providers",
|
|
||||||
"meta_model_name": "authentik_providers_rac.racprovider",
|
|
||||||
"settings": {},
|
|
||||||
"outpost_set": [],
|
|
||||||
},
|
|
||||||
"protocol": "rdp",
|
|
||||||
"host": self.denied.host,
|
|
||||||
"maximum_connections": 1,
|
|
||||||
"settings": {},
|
|
||||||
"property_mappings": [],
|
|
||||||
"auth_mode": "",
|
|
||||||
"launch_url": f"/application/rac/{self.app.slug}/{str(self.denied.pk)}/",
|
|
||||||
},
|
|
||||||
],
|
|
||||||
},
|
|
||||||
)
|
|
|
@ -1,144 +0,0 @@
|
||||||
"""Test RAC Models"""
|
|
||||||
from django.test import TransactionTestCase
|
|
||||||
|
|
||||||
from authentik.core.models import Application, AuthenticatedSession
|
|
||||||
from authentik.core.tests.utils import create_test_admin_user
|
|
||||||
from authentik.enterprise.providers.rac.models import (
|
|
||||||
ConnectionToken,
|
|
||||||
Endpoint,
|
|
||||||
Protocols,
|
|
||||||
RACPropertyMapping,
|
|
||||||
RACProvider,
|
|
||||||
)
|
|
||||||
from authentik.lib.generators import generate_id
|
|
||||||
|
|
||||||
|
|
||||||
class TestModels(TransactionTestCase):
|
|
||||||
"""Test RAC Models"""
|
|
||||||
|
|
||||||
def setUp(self):
|
|
||||||
self.user = create_test_admin_user()
|
|
||||||
self.provider = RACProvider.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
)
|
|
||||||
self.app = Application.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
slug=generate_id(),
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
self.endpoint = Endpoint.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
host=f"{generate_id()}:1324",
|
|
||||||
protocol=Protocols.RDP,
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
|
|
||||||
def test_settings_merge(self):
|
|
||||||
"""Test settings merge"""
|
|
||||||
token = ConnectionToken.objects.create(
|
|
||||||
provider=self.provider,
|
|
||||||
endpoint=self.endpoint,
|
|
||||||
session=AuthenticatedSession.objects.create(
|
|
||||||
user=self.user,
|
|
||||||
session_key=generate_id(),
|
|
||||||
),
|
|
||||||
)
|
|
||||||
path = f"/tmp/connection/{token.token}" # nosec
|
|
||||||
self.assertEqual(
|
|
||||||
token.get_settings(),
|
|
||||||
{
|
|
||||||
"hostname": self.endpoint.host.split(":")[0],
|
|
||||||
"port": "1324",
|
|
||||||
"client-name": "authentik",
|
|
||||||
"drive-path": path,
|
|
||||||
"create-drive-path": "true",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
# Set settings in provider
|
|
||||||
self.provider.settings = {"level": "provider"}
|
|
||||||
self.provider.save()
|
|
||||||
self.assertEqual(
|
|
||||||
token.get_settings(),
|
|
||||||
{
|
|
||||||
"hostname": self.endpoint.host.split(":")[0],
|
|
||||||
"port": "1324",
|
|
||||||
"client-name": "authentik",
|
|
||||||
"drive-path": path,
|
|
||||||
"create-drive-path": "true",
|
|
||||||
"level": "provider",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
# Set settings in endpoint
|
|
||||||
self.endpoint.settings = {
|
|
||||||
"level": "endpoint",
|
|
||||||
}
|
|
||||||
self.endpoint.save()
|
|
||||||
self.assertEqual(
|
|
||||||
token.get_settings(),
|
|
||||||
{
|
|
||||||
"hostname": self.endpoint.host.split(":")[0],
|
|
||||||
"port": "1324",
|
|
||||||
"client-name": "authentik",
|
|
||||||
"drive-path": path,
|
|
||||||
"create-drive-path": "true",
|
|
||||||
"level": "endpoint",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
# Set settings in token
|
|
||||||
token.settings = {
|
|
||||||
"level": "token",
|
|
||||||
}
|
|
||||||
token.save()
|
|
||||||
self.assertEqual(
|
|
||||||
token.get_settings(),
|
|
||||||
{
|
|
||||||
"hostname": self.endpoint.host.split(":")[0],
|
|
||||||
"port": "1324",
|
|
||||||
"client-name": "authentik",
|
|
||||||
"drive-path": path,
|
|
||||||
"create-drive-path": "true",
|
|
||||||
"level": "token",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
# Set settings in property mapping (provider)
|
|
||||||
mapping = RACPropertyMapping.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
expression="""return {
|
|
||||||
"level": "property_mapping_provider"
|
|
||||||
}""",
|
|
||||||
)
|
|
||||||
self.provider.property_mappings.add(mapping)
|
|
||||||
self.assertEqual(
|
|
||||||
token.get_settings(),
|
|
||||||
{
|
|
||||||
"hostname": self.endpoint.host.split(":")[0],
|
|
||||||
"port": "1324",
|
|
||||||
"client-name": "authentik",
|
|
||||||
"drive-path": path,
|
|
||||||
"create-drive-path": "true",
|
|
||||||
"level": "property_mapping_provider",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
# Set settings in property mapping (endpoint)
|
|
||||||
mapping = RACPropertyMapping.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
static_settings={
|
|
||||||
"level": "property_mapping_endpoint",
|
|
||||||
"foo": True,
|
|
||||||
"bar": 6,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
self.endpoint.property_mappings.add(mapping)
|
|
||||||
self.assertEqual(
|
|
||||||
token.get_settings(),
|
|
||||||
{
|
|
||||||
"hostname": self.endpoint.host.split(":")[0],
|
|
||||||
"port": "1324",
|
|
||||||
"client-name": "authentik",
|
|
||||||
"drive-path": path,
|
|
||||||
"create-drive-path": "true",
|
|
||||||
"level": "property_mapping_endpoint",
|
|
||||||
"foo": "true",
|
|
||||||
"bar": "6",
|
|
||||||
},
|
|
||||||
)
|
|
|
@ -1,132 +0,0 @@
|
||||||
"""RAC Views tests"""
|
|
||||||
from datetime import timedelta
|
|
||||||
from json import loads
|
|
||||||
from time import mktime
|
|
||||||
from unittest.mock import MagicMock, patch
|
|
||||||
|
|
||||||
from django.urls import reverse
|
|
||||||
from django.utils.timezone import now
|
|
||||||
from rest_framework.test import APITestCase
|
|
||||||
|
|
||||||
from authentik.core.models import Application
|
|
||||||
from authentik.core.tests.utils import create_test_admin_user, create_test_flow
|
|
||||||
from authentik.enterprise.models import License, LicenseKey
|
|
||||||
from authentik.enterprise.providers.rac.models import Endpoint, Protocols, RACProvider
|
|
||||||
from authentik.lib.generators import generate_id
|
|
||||||
from authentik.policies.denied import AccessDeniedResponse
|
|
||||||
from authentik.policies.dummy.models import DummyPolicy
|
|
||||||
from authentik.policies.models import PolicyBinding
|
|
||||||
|
|
||||||
|
|
||||||
class TestRACViews(APITestCase):
|
|
||||||
"""RAC Views tests"""
|
|
||||||
|
|
||||||
def setUp(self):
|
|
||||||
self.user = create_test_admin_user()
|
|
||||||
self.flow = create_test_flow()
|
|
||||||
self.provider = RACProvider.objects.create(name=generate_id(), authorization_flow=self.flow)
|
|
||||||
self.app = Application.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
slug=generate_id(),
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
self.endpoint = Endpoint.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
host=f"{generate_id()}:1324",
|
|
||||||
protocol=Protocols.RDP,
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
|
|
||||||
@patch(
|
|
||||||
"authentik.enterprise.models.LicenseKey.validate",
|
|
||||||
MagicMock(
|
|
||||||
return_value=LicenseKey(
|
|
||||||
aud="",
|
|
||||||
exp=int(mktime((now() + timedelta(days=3000)).timetuple())),
|
|
||||||
name=generate_id(),
|
|
||||||
internal_users=100,
|
|
||||||
external_users=100,
|
|
||||||
)
|
|
||||||
),
|
|
||||||
)
|
|
||||||
def test_no_policy(self):
|
|
||||||
"""Test request"""
|
|
||||||
License.objects.create(key=generate_id())
|
|
||||||
self.client.force_login(self.user)
|
|
||||||
response = self.client.get(
|
|
||||||
reverse(
|
|
||||||
"authentik_providers_rac:start",
|
|
||||||
kwargs={"app": self.app.slug, "endpoint": str(self.endpoint.pk)},
|
|
||||||
)
|
|
||||||
)
|
|
||||||
self.assertEqual(response.status_code, 302)
|
|
||||||
flow_response = self.client.get(
|
|
||||||
reverse("authentik_api:flow-executor", kwargs={"flow_slug": self.flow.slug})
|
|
||||||
)
|
|
||||||
body = loads(flow_response.content)
|
|
||||||
next_url = body["to"]
|
|
||||||
final_response = self.client.get(next_url)
|
|
||||||
self.assertEqual(final_response.status_code, 200)
|
|
||||||
|
|
||||||
@patch(
|
|
||||||
"authentik.enterprise.models.LicenseKey.validate",
|
|
||||||
MagicMock(
|
|
||||||
return_value=LicenseKey(
|
|
||||||
aud="",
|
|
||||||
exp=int(mktime((now() + timedelta(days=3000)).timetuple())),
|
|
||||||
name=generate_id(),
|
|
||||||
internal_users=100,
|
|
||||||
external_users=100,
|
|
||||||
)
|
|
||||||
),
|
|
||||||
)
|
|
||||||
def test_app_deny(self):
|
|
||||||
"""Test request (deny on app level)"""
|
|
||||||
PolicyBinding.objects.create(
|
|
||||||
target=self.app,
|
|
||||||
policy=DummyPolicy.objects.create(name="deny", result=False, wait_min=1, wait_max=2),
|
|
||||||
order=0,
|
|
||||||
)
|
|
||||||
License.objects.create(key=generate_id())
|
|
||||||
self.client.force_login(self.user)
|
|
||||||
response = self.client.get(
|
|
||||||
reverse(
|
|
||||||
"authentik_providers_rac:start",
|
|
||||||
kwargs={"app": self.app.slug, "endpoint": str(self.endpoint.pk)},
|
|
||||||
)
|
|
||||||
)
|
|
||||||
self.assertIsInstance(response, AccessDeniedResponse)
|
|
||||||
|
|
||||||
@patch(
|
|
||||||
"authentik.enterprise.models.LicenseKey.validate",
|
|
||||||
MagicMock(
|
|
||||||
return_value=LicenseKey(
|
|
||||||
aud="",
|
|
||||||
exp=int(mktime((now() + timedelta(days=3000)).timetuple())),
|
|
||||||
name=generate_id(),
|
|
||||||
internal_users=100,
|
|
||||||
external_users=100,
|
|
||||||
)
|
|
||||||
),
|
|
||||||
)
|
|
||||||
def test_endpoint_deny(self):
|
|
||||||
"""Test request (deny on endpoint level)"""
|
|
||||||
PolicyBinding.objects.create(
|
|
||||||
target=self.endpoint,
|
|
||||||
policy=DummyPolicy.objects.create(name="deny", result=False, wait_min=1, wait_max=2),
|
|
||||||
order=0,
|
|
||||||
)
|
|
||||||
License.objects.create(key=generate_id())
|
|
||||||
self.client.force_login(self.user)
|
|
||||||
response = self.client.get(
|
|
||||||
reverse(
|
|
||||||
"authentik_providers_rac:start",
|
|
||||||
kwargs={"app": self.app.slug, "endpoint": str(self.endpoint.pk)},
|
|
||||||
)
|
|
||||||
)
|
|
||||||
self.assertEqual(response.status_code, 302)
|
|
||||||
flow_response = self.client.get(
|
|
||||||
reverse("authentik_api:flow-executor", kwargs={"flow_slug": self.flow.slug})
|
|
||||||
)
|
|
||||||
body = loads(flow_response.content)
|
|
||||||
self.assertEqual(body["component"], "ak-stage-access-denied")
|
|
|
@ -1,47 +0,0 @@
|
||||||
"""rac urls"""
|
|
||||||
from channels.auth import AuthMiddleware
|
|
||||||
from channels.sessions import CookieMiddleware
|
|
||||||
from django.urls import path
|
|
||||||
from django.views.decorators.csrf import ensure_csrf_cookie
|
|
||||||
|
|
||||||
from authentik.core.channels import TokenOutpostMiddleware
|
|
||||||
from authentik.enterprise.providers.rac.api.endpoints import EndpointViewSet
|
|
||||||
from authentik.enterprise.providers.rac.api.property_mappings import RACPropertyMappingViewSet
|
|
||||||
from authentik.enterprise.providers.rac.api.providers import RACProviderViewSet
|
|
||||||
from authentik.enterprise.providers.rac.consumer_client import RACClientConsumer
|
|
||||||
from authentik.enterprise.providers.rac.consumer_outpost import RACOutpostConsumer
|
|
||||||
from authentik.enterprise.providers.rac.views import RACInterface, RACStartView
|
|
||||||
from authentik.root.asgi_middleware import SessionMiddleware
|
|
||||||
from authentik.root.middleware import ChannelsLoggingMiddleware
|
|
||||||
|
|
||||||
urlpatterns = [
|
|
||||||
path(
|
|
||||||
"application/rac/<slug:app>/<uuid:endpoint>/",
|
|
||||||
ensure_csrf_cookie(RACStartView.as_view()),
|
|
||||||
name="start",
|
|
||||||
),
|
|
||||||
path(
|
|
||||||
"if/rac/<str:token>/",
|
|
||||||
ensure_csrf_cookie(RACInterface.as_view()),
|
|
||||||
name="if-rac",
|
|
||||||
),
|
|
||||||
]
|
|
||||||
|
|
||||||
websocket_urlpatterns = [
|
|
||||||
path(
|
|
||||||
"ws/rac/<str:token>/",
|
|
||||||
ChannelsLoggingMiddleware(
|
|
||||||
CookieMiddleware(SessionMiddleware(AuthMiddleware(RACClientConsumer.as_asgi())))
|
|
||||||
),
|
|
||||||
),
|
|
||||||
path(
|
|
||||||
"ws/outpost_rac/<str:channel>/",
|
|
||||||
ChannelsLoggingMiddleware(TokenOutpostMiddleware(RACOutpostConsumer.as_asgi())),
|
|
||||||
),
|
|
||||||
]
|
|
||||||
|
|
||||||
api_urlpatterns = [
|
|
||||||
("providers/rac", RACProviderViewSet),
|
|
||||||
("propertymappings/rac", RACPropertyMappingViewSet),
|
|
||||||
("rac/endpoints", EndpointViewSet),
|
|
||||||
]
|
|
|
@ -1,140 +0,0 @@
|
||||||
"""RAC Views"""
|
|
||||||
from typing import Any
|
|
||||||
|
|
||||||
from django.http import Http404, HttpRequest, HttpResponse
|
|
||||||
from django.shortcuts import get_object_or_404, redirect
|
|
||||||
from django.urls import reverse
|
|
||||||
from django.utils.timezone import now
|
|
||||||
from django.utils.translation import gettext as _
|
|
||||||
|
|
||||||
from authentik.core.models import Application, AuthenticatedSession
|
|
||||||
from authentik.core.views.interface import InterfaceView
|
|
||||||
from authentik.enterprise.policy import EnterprisePolicyAccessView
|
|
||||||
from authentik.enterprise.providers.rac.models import ConnectionToken, Endpoint, RACProvider
|
|
||||||
from authentik.events.models import Event, EventAction
|
|
||||||
from authentik.flows.challenge import RedirectChallenge
|
|
||||||
from authentik.flows.exceptions import FlowNonApplicableException
|
|
||||||
from authentik.flows.models import in_memory_stage
|
|
||||||
from authentik.flows.planner import FlowPlanner
|
|
||||||
from authentik.flows.stage import RedirectStage
|
|
||||||
from authentik.flows.views.executor import SESSION_KEY_PLAN
|
|
||||||
from authentik.lib.utils.time import timedelta_from_string
|
|
||||||
from authentik.lib.utils.urls import redirect_with_qs
|
|
||||||
from authentik.policies.engine import PolicyEngine
|
|
||||||
|
|
||||||
|
|
||||||
class RACStartView(EnterprisePolicyAccessView):
|
|
||||||
"""Start a RAC connection by checking access and creating a connection token"""
|
|
||||||
|
|
||||||
endpoint: Endpoint
|
|
||||||
|
|
||||||
def resolve_provider_application(self):
|
|
||||||
self.application = get_object_or_404(Application, slug=self.kwargs["app"])
|
|
||||||
# Endpoint permissions are validated in the RACFinalStage below
|
|
||||||
self.endpoint = get_object_or_404(Endpoint, pk=self.kwargs["endpoint"])
|
|
||||||
self.provider = RACProvider.objects.get(application=self.application)
|
|
||||||
|
|
||||||
def get(self, request: HttpRequest, *args, **kwargs) -> HttpResponse:
|
|
||||||
"""Start flow planner for RAC provider"""
|
|
||||||
planner = FlowPlanner(self.provider.authorization_flow)
|
|
||||||
planner.allow_empty_flows = True
|
|
||||||
try:
|
|
||||||
plan = planner.plan(self.request)
|
|
||||||
except FlowNonApplicableException:
|
|
||||||
raise Http404
|
|
||||||
plan.insert_stage(
|
|
||||||
in_memory_stage(
|
|
||||||
RACFinalStage,
|
|
||||||
application=self.application,
|
|
||||||
endpoint=self.endpoint,
|
|
||||||
provider=self.provider,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
request.session[SESSION_KEY_PLAN] = plan
|
|
||||||
return redirect_with_qs(
|
|
||||||
"authentik_core:if-flow",
|
|
||||||
request.GET,
|
|
||||||
flow_slug=self.provider.authorization_flow.slug,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class RACInterface(InterfaceView):
|
|
||||||
"""Start RAC connection"""
|
|
||||||
|
|
||||||
template_name = "if/rac.html"
|
|
||||||
token: ConnectionToken
|
|
||||||
|
|
||||||
def dispatch(self, request: HttpRequest, *args: Any, **kwargs: Any) -> HttpResponse:
|
|
||||||
# Early sanity check to ensure token still exists
|
|
||||||
token = ConnectionToken.filter_not_expired(token=self.kwargs["token"]).first()
|
|
||||||
if not token:
|
|
||||||
return redirect("authentik_core:if-user")
|
|
||||||
self.token = token
|
|
||||||
return super().dispatch(request, *args, **kwargs)
|
|
||||||
|
|
||||||
def get_context_data(self, **kwargs: Any) -> dict[str, Any]:
|
|
||||||
kwargs["token"] = self.token
|
|
||||||
return super().get_context_data(**kwargs)
|
|
||||||
|
|
||||||
|
|
||||||
class RACFinalStage(RedirectStage):
|
|
||||||
"""RAC Connection final stage, set the connection token in the stage"""
|
|
||||||
|
|
||||||
endpoint: Endpoint
|
|
||||||
provider: RACProvider
|
|
||||||
application: Application
|
|
||||||
|
|
||||||
def dispatch(self, request: HttpRequest, *args: Any, **kwargs: Any) -> HttpResponse:
|
|
||||||
self.endpoint = self.executor.current_stage.endpoint
|
|
||||||
self.provider = self.executor.current_stage.provider
|
|
||||||
self.application = self.executor.current_stage.application
|
|
||||||
# Check policies bound to endpoint directly
|
|
||||||
engine = PolicyEngine(self.endpoint, self.request.user, self.request)
|
|
||||||
engine.use_cache = False
|
|
||||||
engine.build()
|
|
||||||
passing = engine.result
|
|
||||||
if not passing.passing:
|
|
||||||
return self.executor.stage_invalid(", ".join(passing.messages))
|
|
||||||
# Check if we're already at the maximum connection limit
|
|
||||||
all_tokens = ConnectionToken.filter_not_expired(
|
|
||||||
endpoint=self.endpoint,
|
|
||||||
).exclude(endpoint__maximum_connections__lte=-1)
|
|
||||||
if all_tokens.count() >= self.endpoint.maximum_connections:
|
|
||||||
msg = [_("Maximum connection limit reached.")]
|
|
||||||
# Check if any other tokens exist for the current user, and inform them
|
|
||||||
# they are already connected
|
|
||||||
if all_tokens.filter(session__user=self.request.user).exists():
|
|
||||||
msg.append(_("(You are already connected in another tab/window)"))
|
|
||||||
return self.executor.stage_invalid(" ".join(msg))
|
|
||||||
return super().dispatch(request, *args, **kwargs)
|
|
||||||
|
|
||||||
def get_challenge(self, *args, **kwargs) -> RedirectChallenge:
|
|
||||||
token = ConnectionToken.objects.create(
|
|
||||||
provider=self.provider,
|
|
||||||
endpoint=self.endpoint,
|
|
||||||
settings=self.executor.plan.context.get("connection_settings", {}),
|
|
||||||
session=AuthenticatedSession.objects.filter(
|
|
||||||
session_key=self.request.session.session_key
|
|
||||||
).first(),
|
|
||||||
expires=now() + timedelta_from_string(self.provider.connection_expiry),
|
|
||||||
expiring=True,
|
|
||||||
)
|
|
||||||
Event.new(
|
|
||||||
EventAction.AUTHORIZE_APPLICATION,
|
|
||||||
authorized_application=self.application,
|
|
||||||
flow=self.executor.plan.flow_pk,
|
|
||||||
endpoint=self.endpoint.name,
|
|
||||||
).from_http(self.request)
|
|
||||||
setattr(
|
|
||||||
self.executor.current_stage,
|
|
||||||
"destination",
|
|
||||||
self.request.build_absolute_uri(
|
|
||||||
reverse(
|
|
||||||
"authentik_providers_rac:if-rac",
|
|
||||||
kwargs={
|
|
||||||
"token": str(token.token),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
),
|
|
||||||
)
|
|
||||||
return super().get_challenge(*args, **kwargs)
|
|
|
@ -10,7 +10,3 @@ CELERY_BEAT_SCHEDULE = {
|
||||||
"options": {"queue": "authentik_scheduled"},
|
"options": {"queue": "authentik_scheduled"},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
INSTALLED_APPS = [
|
|
||||||
"authentik.enterprise.providers.rac",
|
|
||||||
]
|
|
||||||
|
|
|
@ -5,8 +5,7 @@ from json import loads
|
||||||
import django_filters
|
import django_filters
|
||||||
from django.db.models.aggregates import Count
|
from django.db.models.aggregates import Count
|
||||||
from django.db.models.fields.json import KeyTextTransform, KeyTransform
|
from django.db.models.fields.json import KeyTextTransform, KeyTransform
|
||||||
from django.db.models.functions import ExtractDay, ExtractHour
|
from django.db.models.functions import ExtractDay
|
||||||
from django.db.models.query_utils import Q
|
|
||||||
from drf_spectacular.types import OpenApiTypes
|
from drf_spectacular.types import OpenApiTypes
|
||||||
from drf_spectacular.utils import OpenApiParameter, extend_schema
|
from drf_spectacular.utils import OpenApiParameter, extend_schema
|
||||||
from guardian.shortcuts import get_objects_for_user
|
from guardian.shortcuts import get_objects_for_user
|
||||||
|
@ -37,7 +36,6 @@ class EventSerializer(ModelSerializer):
|
||||||
"created",
|
"created",
|
||||||
"expires",
|
"expires",
|
||||||
"tenant",
|
"tenant",
|
||||||
"batch_id",
|
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
@ -89,12 +87,7 @@ class EventsFilter(django_filters.FilterSet):
|
||||||
we need to remove the dashes that a client may send. We can't use a
|
we need to remove the dashes that a client may send. We can't use a
|
||||||
UUIDField for this, as some models might not have a UUID PK"""
|
UUIDField for this, as some models might not have a UUID PK"""
|
||||||
value = str(value).replace("-", "")
|
value = str(value).replace("-", "")
|
||||||
query = Q(context__model__pk=value)
|
return queryset.filter(context__model__pk=value)
|
||||||
try:
|
|
||||||
query |= Q(context__model__pk=int(value))
|
|
||||||
except ValueError:
|
|
||||||
pass
|
|
||||||
return queryset.filter(query)
|
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
model = Event
|
model = Event
|
||||||
|
@ -156,15 +149,7 @@ class EventViewSet(ModelViewSet):
|
||||||
return Response(EventTopPerUserSerializer(instance=events, many=True).data)
|
return Response(EventTopPerUserSerializer(instance=events, many=True).data)
|
||||||
|
|
||||||
@extend_schema(
|
@extend_schema(
|
||||||
responses={200: CoordinateSerializer(many=True)},
|
methods=["GET"],
|
||||||
)
|
|
||||||
@action(detail=False, methods=["GET"], pagination_class=None)
|
|
||||||
def volume(self, request: Request) -> Response:
|
|
||||||
"""Get event volume for specified filters and timeframe"""
|
|
||||||
queryset = self.filter_queryset(self.get_queryset())
|
|
||||||
return Response(queryset.get_events_per(timedelta(days=7), ExtractHour, 7 * 3))
|
|
||||||
|
|
||||||
@extend_schema(
|
|
||||||
responses={200: CoordinateSerializer(many=True)},
|
responses={200: CoordinateSerializer(many=True)},
|
||||||
filters=[],
|
filters=[],
|
||||||
parameters=[
|
parameters=[
|
||||||
|
|
|
@ -52,9 +52,6 @@ class NotificationTransportSerializer(ModelSerializer):
|
||||||
"webhook_url",
|
"webhook_url",
|
||||||
"webhook_mapping",
|
"webhook_mapping",
|
||||||
"send_once",
|
"send_once",
|
||||||
"enable_batching",
|
|
||||||
"batch_timeout",
|
|
||||||
"max_batch_size",
|
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
from prometheus_client import Gauge
|
from prometheus_client import Gauge
|
||||||
|
|
||||||
from authentik.blueprints.apps import ManagedAppConfig
|
from authentik.blueprints.apps import ManagedAppConfig
|
||||||
from authentik.lib.config import CONFIG, ENV_PREFIX
|
|
||||||
|
|
||||||
GAUGE_TASKS = Gauge(
|
GAUGE_TASKS = Gauge(
|
||||||
"authentik_system_tasks",
|
"authentik_system_tasks",
|
||||||
|
@ -22,24 +21,3 @@ class AuthentikEventsConfig(ManagedAppConfig):
|
||||||
def reconcile_load_events_signals(self):
|
def reconcile_load_events_signals(self):
|
||||||
"""Load events signals"""
|
"""Load events signals"""
|
||||||
self.import_module("authentik.events.signals")
|
self.import_module("authentik.events.signals")
|
||||||
|
|
||||||
def reconcile_check_deprecations(self):
|
|
||||||
"""Check for config deprecations"""
|
|
||||||
from authentik.events.models import Event, EventAction
|
|
||||||
|
|
||||||
for key_replace, msg in CONFIG.deprecations.items():
|
|
||||||
key, replace = key_replace
|
|
||||||
key_env = f"{ENV_PREFIX}_{key.replace('.', '__')}".upper()
|
|
||||||
replace_env = f"{ENV_PREFIX}_{replace.replace('.', '__')}".upper()
|
|
||||||
if Event.objects.filter(
|
|
||||||
action=EventAction.CONFIGURATION_ERROR, context__deprecated_option=key
|
|
||||||
).exists():
|
|
||||||
continue
|
|
||||||
Event.new(
|
|
||||||
EventAction.CONFIGURATION_ERROR,
|
|
||||||
deprecated_option=key,
|
|
||||||
deprecated_env=key_env,
|
|
||||||
replacement_option=replace,
|
|
||||||
replacement_env=replace_env,
|
|
||||||
message=msg,
|
|
||||||
).save()
|
|
||||||
|
|
|
@ -1,81 +0,0 @@
|
||||||
"""ASN Enricher"""
|
|
||||||
from typing import TYPE_CHECKING, Optional, TypedDict
|
|
||||||
|
|
||||||
from django.http import HttpRequest
|
|
||||||
from geoip2.errors import GeoIP2Error
|
|
||||||
from geoip2.models import ASN
|
|
||||||
from sentry_sdk import Hub
|
|
||||||
|
|
||||||
from authentik.events.context_processors.mmdb import MMDBContextProcessor
|
|
||||||
from authentik.lib.config import CONFIG
|
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from authentik.api.v3.config import Capabilities
|
|
||||||
from authentik.events.models import Event
|
|
||||||
|
|
||||||
|
|
||||||
class ASNDict(TypedDict):
|
|
||||||
"""ASN Details"""
|
|
||||||
|
|
||||||
asn: int
|
|
||||||
as_org: str | None
|
|
||||||
network: str | None
|
|
||||||
|
|
||||||
|
|
||||||
class ASNContextProcessor(MMDBContextProcessor):
|
|
||||||
"""ASN Database reader wrapper"""
|
|
||||||
|
|
||||||
def capability(self) -> Optional["Capabilities"]:
|
|
||||||
from authentik.api.v3.config import Capabilities
|
|
||||||
|
|
||||||
return Capabilities.CAN_ASN
|
|
||||||
|
|
||||||
def path(self) -> str | None:
|
|
||||||
return CONFIG.get("events.context_processors.asn")
|
|
||||||
|
|
||||||
def enrich_event(self, event: "Event"):
|
|
||||||
asn = self.asn_dict(event.client_ip)
|
|
||||||
if not asn:
|
|
||||||
return
|
|
||||||
event.context["asn"] = asn
|
|
||||||
|
|
||||||
def enrich_context(self, request: HttpRequest) -> dict:
|
|
||||||
return {
|
|
||||||
"asn": self.asn_dict(ClientIPMiddleware.get_client_ip(request)),
|
|
||||||
}
|
|
||||||
|
|
||||||
def asn(self, ip_address: str) -> Optional[ASN]:
|
|
||||||
"""Wrapper for Reader.asn"""
|
|
||||||
with Hub.current.start_span(
|
|
||||||
op="authentik.events.asn.asn",
|
|
||||||
description=ip_address,
|
|
||||||
):
|
|
||||||
if not self.configured():
|
|
||||||
return None
|
|
||||||
self.check_expired()
|
|
||||||
try:
|
|
||||||
return self.reader.asn(ip_address)
|
|
||||||
except (GeoIP2Error, ValueError):
|
|
||||||
return None
|
|
||||||
|
|
||||||
def asn_to_dict(self, asn: ASN | None) -> ASNDict:
|
|
||||||
"""Convert ASN to dict"""
|
|
||||||
if not asn:
|
|
||||||
return {}
|
|
||||||
asn_dict: ASNDict = {
|
|
||||||
"asn": asn.autonomous_system_number,
|
|
||||||
"as_org": asn.autonomous_system_organization,
|
|
||||||
"network": str(asn.network) if asn.network else None,
|
|
||||||
}
|
|
||||||
return asn_dict
|
|
||||||
|
|
||||||
def asn_dict(self, ip_address: str) -> Optional[ASNDict]:
|
|
||||||
"""Wrapper for self.asn that returns a dict"""
|
|
||||||
asn = self.asn(ip_address)
|
|
||||||
if not asn:
|
|
||||||
return None
|
|
||||||
return self.asn_to_dict(asn)
|
|
||||||
|
|
||||||
|
|
||||||
ASN_CONTEXT_PROCESSOR = ASNContextProcessor()
|
|
|
@ -1,43 +0,0 @@
|
||||||
"""Base event enricher"""
|
|
||||||
from functools import cache
|
|
||||||
from typing import TYPE_CHECKING, Optional
|
|
||||||
|
|
||||||
from django.http import HttpRequest
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from authentik.api.v3.config import Capabilities
|
|
||||||
from authentik.events.models import Event
|
|
||||||
|
|
||||||
|
|
||||||
class EventContextProcessor:
|
|
||||||
"""Base event enricher"""
|
|
||||||
|
|
||||||
def capability(self) -> Optional["Capabilities"]:
|
|
||||||
"""Return the capability this context processor provides"""
|
|
||||||
return None
|
|
||||||
|
|
||||||
def configured(self) -> bool:
|
|
||||||
"""Return true if this context processor is configured"""
|
|
||||||
return False
|
|
||||||
|
|
||||||
def enrich_event(self, event: "Event"):
|
|
||||||
"""Modify event"""
|
|
||||||
raise NotImplementedError
|
|
||||||
|
|
||||||
def enrich_context(self, request: HttpRequest) -> dict:
|
|
||||||
"""Modify context"""
|
|
||||||
raise NotImplementedError
|
|
||||||
|
|
||||||
|
|
||||||
@cache
|
|
||||||
def get_context_processors() -> list[EventContextProcessor]:
|
|
||||||
"""Get a list of all configured context processors"""
|
|
||||||
from authentik.events.context_processors.asn import ASN_CONTEXT_PROCESSOR
|
|
||||||
from authentik.events.context_processors.geoip import GEOIP_CONTEXT_PROCESSOR
|
|
||||||
|
|
||||||
processors_types = [ASN_CONTEXT_PROCESSOR, GEOIP_CONTEXT_PROCESSOR]
|
|
||||||
processors = []
|
|
||||||
for _type in processors_types:
|
|
||||||
if _type.configured():
|
|
||||||
processors.append(_type)
|
|
||||||
return processors
|
|
|
@ -1,86 +0,0 @@
|
||||||
"""events GeoIP Reader"""
|
|
||||||
from typing import TYPE_CHECKING, Optional, TypedDict
|
|
||||||
|
|
||||||
from django.http import HttpRequest
|
|
||||||
from geoip2.errors import GeoIP2Error
|
|
||||||
from geoip2.models import City
|
|
||||||
from sentry_sdk.hub import Hub
|
|
||||||
|
|
||||||
from authentik.events.context_processors.mmdb import MMDBContextProcessor
|
|
||||||
from authentik.lib.config import CONFIG
|
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from authentik.api.v3.config import Capabilities
|
|
||||||
from authentik.events.models import Event
|
|
||||||
|
|
||||||
|
|
||||||
class GeoIPDict(TypedDict):
|
|
||||||
"""GeoIP Details"""
|
|
||||||
|
|
||||||
continent: str
|
|
||||||
country: str
|
|
||||||
lat: float
|
|
||||||
long: float
|
|
||||||
city: str
|
|
||||||
|
|
||||||
|
|
||||||
class GeoIPContextProcessor(MMDBContextProcessor):
|
|
||||||
"""Slim wrapper around GeoIP API"""
|
|
||||||
|
|
||||||
def capability(self) -> Optional["Capabilities"]:
|
|
||||||
from authentik.api.v3.config import Capabilities
|
|
||||||
|
|
||||||
return Capabilities.CAN_GEO_IP
|
|
||||||
|
|
||||||
def path(self) -> str | None:
|
|
||||||
return CONFIG.get("events.context_processors.geoip")
|
|
||||||
|
|
||||||
def enrich_event(self, event: "Event"):
|
|
||||||
city = self.city_dict(event.client_ip)
|
|
||||||
if not city:
|
|
||||||
return
|
|
||||||
event.context["geo"] = city
|
|
||||||
|
|
||||||
def enrich_context(self, request: HttpRequest) -> dict:
|
|
||||||
# Different key `geoip` vs `geo` for legacy reasons
|
|
||||||
return {"geoip": self.city(ClientIPMiddleware.get_client_ip(request))}
|
|
||||||
|
|
||||||
def city(self, ip_address: str) -> Optional[City]:
|
|
||||||
"""Wrapper for Reader.city"""
|
|
||||||
with Hub.current.start_span(
|
|
||||||
op="authentik.events.geo.city",
|
|
||||||
description=ip_address,
|
|
||||||
):
|
|
||||||
if not self.configured():
|
|
||||||
return None
|
|
||||||
self.check_expired()
|
|
||||||
try:
|
|
||||||
return self.reader.city(ip_address)
|
|
||||||
except (GeoIP2Error, ValueError):
|
|
||||||
return None
|
|
||||||
|
|
||||||
def city_to_dict(self, city: City | None) -> GeoIPDict:
|
|
||||||
"""Convert City to dict"""
|
|
||||||
if not city:
|
|
||||||
return {}
|
|
||||||
city_dict: GeoIPDict = {
|
|
||||||
"continent": city.continent.code,
|
|
||||||
"country": city.country.iso_code,
|
|
||||||
"lat": city.location.latitude,
|
|
||||||
"long": city.location.longitude,
|
|
||||||
"city": "",
|
|
||||||
}
|
|
||||||
if city.city.name:
|
|
||||||
city_dict["city"] = city.city.name
|
|
||||||
return city_dict
|
|
||||||
|
|
||||||
def city_dict(self, ip_address: str) -> Optional[GeoIPDict]:
|
|
||||||
"""Wrapper for self.city that returns a dict"""
|
|
||||||
city = self.city(ip_address)
|
|
||||||
if not city:
|
|
||||||
return None
|
|
||||||
return self.city_to_dict(city)
|
|
||||||
|
|
||||||
|
|
||||||
GEOIP_CONTEXT_PROCESSOR = GeoIPContextProcessor()
|
|
|
@ -1,53 +0,0 @@
|
||||||
"""Common logic for reading MMDB files"""
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import Optional
|
|
||||||
|
|
||||||
from geoip2.database import Reader
|
|
||||||
from structlog.stdlib import get_logger
|
|
||||||
|
|
||||||
from authentik.events.context_processors.base import EventContextProcessor
|
|
||||||
|
|
||||||
|
|
||||||
class MMDBContextProcessor(EventContextProcessor):
|
|
||||||
"""Common logic for reading MaxMind DB files, including re-loading if the file has changed"""
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
self.reader: Optional[Reader] = None
|
|
||||||
self._last_mtime: float = 0.0
|
|
||||||
self.logger = get_logger()
|
|
||||||
self.open()
|
|
||||||
|
|
||||||
def path(self) -> str | None:
|
|
||||||
"""Get the path to the MMDB file to load"""
|
|
||||||
raise NotImplementedError
|
|
||||||
|
|
||||||
def open(self):
|
|
||||||
"""Get GeoIP Reader, if configured, otherwise none"""
|
|
||||||
path = self.path()
|
|
||||||
if path == "" or not path:
|
|
||||||
return
|
|
||||||
try:
|
|
||||||
self.reader = Reader(path)
|
|
||||||
self._last_mtime = Path(path).stat().st_mtime
|
|
||||||
self.logger.info("Loaded MMDB database", last_write=self._last_mtime, file=path)
|
|
||||||
except OSError as exc:
|
|
||||||
self.logger.warning("Failed to load MMDB database", path=path, exc=exc)
|
|
||||||
|
|
||||||
def check_expired(self):
|
|
||||||
"""Check if the modification date of the MMDB database has
|
|
||||||
changed, and reload it if so"""
|
|
||||||
path = self.path()
|
|
||||||
if path == "" or not path:
|
|
||||||
return
|
|
||||||
try:
|
|
||||||
mtime = Path(path).stat().st_mtime
|
|
||||||
diff = self._last_mtime < mtime
|
|
||||||
if diff > 0:
|
|
||||||
self.logger.info("Found new MMDB Database, reopening", diff=diff, path=path)
|
|
||||||
self.open()
|
|
||||||
except OSError as exc:
|
|
||||||
self.logger.warning("Failed to check MMDB age", exc=exc)
|
|
||||||
|
|
||||||
def configured(self) -> bool:
|
|
||||||
"""Return true if this context processor is configured"""
|
|
||||||
return bool(self.reader)
|
|
100
authentik/events/geo.py
Normal file
100
authentik/events/geo.py
Normal file
|
@ -0,0 +1,100 @@
|
||||||
|
"""events GeoIP Reader"""
|
||||||
|
from os import stat
|
||||||
|
from typing import Optional, TypedDict
|
||||||
|
|
||||||
|
from geoip2.database import Reader
|
||||||
|
from geoip2.errors import GeoIP2Error
|
||||||
|
from geoip2.models import City
|
||||||
|
from sentry_sdk.hub import Hub
|
||||||
|
from structlog.stdlib import get_logger
|
||||||
|
|
||||||
|
from authentik.lib.config import CONFIG
|
||||||
|
|
||||||
|
LOGGER = get_logger()
|
||||||
|
|
||||||
|
|
||||||
|
class GeoIPDict(TypedDict):
|
||||||
|
"""GeoIP Details"""
|
||||||
|
|
||||||
|
continent: str
|
||||||
|
country: str
|
||||||
|
lat: float
|
||||||
|
long: float
|
||||||
|
city: str
|
||||||
|
|
||||||
|
|
||||||
|
class GeoIPReader:
|
||||||
|
"""Slim wrapper around GeoIP API"""
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
self.__reader: Optional[Reader] = None
|
||||||
|
self.__last_mtime: float = 0.0
|
||||||
|
self.__open()
|
||||||
|
|
||||||
|
def __open(self):
|
||||||
|
"""Get GeoIP Reader, if configured, otherwise none"""
|
||||||
|
path = CONFIG.get("geoip")
|
||||||
|
if path == "" or not path:
|
||||||
|
return
|
||||||
|
try:
|
||||||
|
self.__reader = Reader(path)
|
||||||
|
self.__last_mtime = stat(path).st_mtime
|
||||||
|
LOGGER.info("Loaded GeoIP database", last_write=self.__last_mtime)
|
||||||
|
except OSError as exc:
|
||||||
|
LOGGER.warning("Failed to load GeoIP database", exc=exc)
|
||||||
|
|
||||||
|
def __check_expired(self):
|
||||||
|
"""Check if the modification date of the GeoIP database has
|
||||||
|
changed, and reload it if so"""
|
||||||
|
path = CONFIG.get("geoip")
|
||||||
|
try:
|
||||||
|
mtime = stat(path).st_mtime
|
||||||
|
diff = self.__last_mtime < mtime
|
||||||
|
if diff > 0:
|
||||||
|
LOGGER.info("Found new GeoIP Database, reopening", diff=diff)
|
||||||
|
self.__open()
|
||||||
|
except OSError as exc:
|
||||||
|
LOGGER.warning("Failed to check GeoIP age", exc=exc)
|
||||||
|
return
|
||||||
|
|
||||||
|
@property
|
||||||
|
def enabled(self) -> bool:
|
||||||
|
"""Check if GeoIP is enabled"""
|
||||||
|
return bool(self.__reader)
|
||||||
|
|
||||||
|
def city(self, ip_address: str) -> Optional[City]:
|
||||||
|
"""Wrapper for Reader.city"""
|
||||||
|
with Hub.current.start_span(
|
||||||
|
op="authentik.events.geo.city",
|
||||||
|
description=ip_address,
|
||||||
|
):
|
||||||
|
if not self.enabled:
|
||||||
|
return None
|
||||||
|
self.__check_expired()
|
||||||
|
try:
|
||||||
|
return self.__reader.city(ip_address)
|
||||||
|
except (GeoIP2Error, ValueError):
|
||||||
|
return None
|
||||||
|
|
||||||
|
def city_to_dict(self, city: City) -> GeoIPDict:
|
||||||
|
"""Convert City to dict"""
|
||||||
|
city_dict: GeoIPDict = {
|
||||||
|
"continent": city.continent.code,
|
||||||
|
"country": city.country.iso_code,
|
||||||
|
"lat": city.location.latitude,
|
||||||
|
"long": city.location.longitude,
|
||||||
|
"city": "",
|
||||||
|
}
|
||||||
|
if city.city.name:
|
||||||
|
city_dict["city"] = city.city.name
|
||||||
|
return city_dict
|
||||||
|
|
||||||
|
def city_dict(self, ip_address: str) -> Optional[GeoIPDict]:
|
||||||
|
"""Wrapper for self.city that returns a dict"""
|
||||||
|
city = self.city(ip_address)
|
||||||
|
if not city:
|
||||||
|
return None
|
||||||
|
return self.city_to_dict(city)
|
||||||
|
|
||||||
|
|
||||||
|
GEOIP_READER = GeoIPReader()
|
|
@ -20,7 +20,6 @@ from authentik.core.models import (
|
||||||
User,
|
User,
|
||||||
UserSourceConnection,
|
UserSourceConnection,
|
||||||
)
|
)
|
||||||
from authentik.enterprise.providers.rac.models import ConnectionToken
|
|
||||||
from authentik.events.models import Event, EventAction, Notification
|
from authentik.events.models import Event, EventAction, Notification
|
||||||
from authentik.events.utils import model_to_dict
|
from authentik.events.utils import model_to_dict
|
||||||
from authentik.flows.models import FlowToken, Stage
|
from authentik.flows.models import FlowToken, Stage
|
||||||
|
@ -55,7 +54,6 @@ IGNORED_MODELS = (
|
||||||
SCIMUser,
|
SCIMUser,
|
||||||
SCIMGroup,
|
SCIMGroup,
|
||||||
Reputation,
|
Reputation,
|
||||||
ConnectionToken,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -26,7 +26,7 @@ from authentik.core.middleware import (
|
||||||
SESSION_KEY_IMPERSONATE_USER,
|
SESSION_KEY_IMPERSONATE_USER,
|
||||||
)
|
)
|
||||||
from authentik.core.models import ExpiringModel, Group, PropertyMapping, User
|
from authentik.core.models import ExpiringModel, Group, PropertyMapping, User
|
||||||
from authentik.events.context_processors.base import get_context_processors
|
from authentik.events.geo import GEOIP_READER
|
||||||
from authentik.events.utils import (
|
from authentik.events.utils import (
|
||||||
cleanse_dict,
|
cleanse_dict,
|
||||||
get_user,
|
get_user,
|
||||||
|
@ -36,10 +36,9 @@ from authentik.events.utils import (
|
||||||
)
|
)
|
||||||
from authentik.lib.models import DomainlessURLValidator, SerializerModel
|
from authentik.lib.models import DomainlessURLValidator, SerializerModel
|
||||||
from authentik.lib.sentry import SentryIgnoredException
|
from authentik.lib.sentry import SentryIgnoredException
|
||||||
from authentik.lib.utils.http import get_http_session
|
from authentik.lib.utils.http import get_client_ip, get_http_session
|
||||||
from authentik.lib.utils.time import timedelta_from_string
|
from authentik.lib.utils.time import timedelta_from_string
|
||||||
from authentik.policies.models import PolicyBindingModel
|
from authentik.policies.models import PolicyBindingModel
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
from authentik.stages.email.utils import TemplateEmailMessage
|
from authentik.stages.email.utils import TemplateEmailMessage
|
||||||
from authentik.tenants.models import Tenant
|
from authentik.tenants.models import Tenant
|
||||||
from authentik.tenants.utils import DEFAULT_TENANT
|
from authentik.tenants.utils import DEFAULT_TENANT
|
||||||
|
@ -161,63 +160,6 @@ class EventManager(Manager):
|
||||||
return self.get_queryset().get_events_per(time_since, extract, data_points)
|
return self.get_queryset().get_events_per(time_since, extract, data_points)
|
||||||
|
|
||||||
|
|
||||||
class EventBatch(ExpiringModel):
|
|
||||||
"""Model to store information about batches of events."""
|
|
||||||
|
|
||||||
batch_id = models.UUIDField(primary_key=True, default=uuid4, editable=False)
|
|
||||||
event_type = models.CharField(max_length=255)
|
|
||||||
event_app = models.CharField(max_length=255)
|
|
||||||
event_user = models.CharField(max_length=255)
|
|
||||||
start_time = models.DateTimeField(auto_now_add=True)
|
|
||||||
end_time = models.DateTimeField(null=True, blank=True)
|
|
||||||
event_count = models.IntegerField(default=0)
|
|
||||||
last_updated = models.DateTimeField(auto_now=True)
|
|
||||||
max_batch_size = models.IntegerField(default=10)
|
|
||||||
batch_timeout = models.IntegerField(default=60) # Timeout in seconds
|
|
||||||
sent = models.BooleanField(default=False)
|
|
||||||
|
|
||||||
def add_event_to_batch(self, event):
|
|
||||||
"""Add an event to the batch and check if it's ready to send."""
|
|
||||||
self.add_event(event)
|
|
||||||
if self.check_batch_limits():
|
|
||||||
self.process_batch()
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def get_or_create_batch(action, app, user):
|
|
||||||
"""Get or create a batch for a given action."""
|
|
||||||
return EventBatch.objects.filter(
|
|
||||||
event_type=action, event_app=app, event_user=user, end_time__isnull=True
|
|
||||||
).first() or EventBatch.objects.create(event_type=action, event_app=app, event_user=user)
|
|
||||||
|
|
||||||
def check_batch_limits(self):
|
|
||||||
"""Check if the batch has reached its size or timeout limits."""
|
|
||||||
time_elapsed = now() - self.start_time
|
|
||||||
return self.event_count >= self.max_batch_size or time_elapsed >= timedelta(
|
|
||||||
seconds=self.batch_timeout
|
|
||||||
)
|
|
||||||
|
|
||||||
def add_event(self, event):
|
|
||||||
"""Add an event to the batch."""
|
|
||||||
self.event_count += 1
|
|
||||||
self.save()
|
|
||||||
|
|
||||||
def create_batch_summary(self):
|
|
||||||
"""Create a summary message for the batch."""
|
|
||||||
return f"Batched Event Summary: {self.event_type} action \
|
|
||||||
on {self.event_app} app by {self.event_user} user \
|
|
||||||
occurred {self.event_count} times between {self.start_time} and {now()}"
|
|
||||||
|
|
||||||
def process_batch(self):
|
|
||||||
"""Process the batch and check if it's ready to send."""
|
|
||||||
summary_message = self.create_batch_summary()
|
|
||||||
return summary_message
|
|
||||||
|
|
||||||
def send_notification(self):
|
|
||||||
"""Send notification for this batch."""
|
|
||||||
# Implement the logic to send notification
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
class Event(SerializerModel, ExpiringModel):
|
class Event(SerializerModel, ExpiringModel):
|
||||||
"""An individual Audit/Metrics/Notification/Error Event"""
|
"""An individual Audit/Metrics/Notification/Error Event"""
|
||||||
|
|
||||||
|
@ -233,8 +175,6 @@ class Event(SerializerModel, ExpiringModel):
|
||||||
# Shadow the expires attribute from ExpiringModel to override the default duration
|
# Shadow the expires attribute from ExpiringModel to override the default duration
|
||||||
expires = models.DateTimeField(default=default_event_duration)
|
expires = models.DateTimeField(default=default_event_duration)
|
||||||
|
|
||||||
batch_id = models.UUIDField(null=True, blank=True)
|
|
||||||
|
|
||||||
objects = EventManager()
|
objects = EventManager()
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
@ -256,7 +196,6 @@ class Event(SerializerModel, ExpiringModel):
|
||||||
current = currentframe()
|
current = currentframe()
|
||||||
parent = current.f_back
|
parent = current.f_back
|
||||||
app = parent.f_globals["__name__"]
|
app = parent.f_globals["__name__"]
|
||||||
|
|
||||||
cleaned_kwargs = cleanse_dict(sanitize_dict(kwargs))
|
cleaned_kwargs = cleanse_dict(sanitize_dict(kwargs))
|
||||||
event = Event(action=action, app=app, context=cleaned_kwargs)
|
event = Event(action=action, app=app, context=cleaned_kwargs)
|
||||||
return event
|
return event
|
||||||
|
@ -305,20 +244,23 @@ class Event(SerializerModel, ExpiringModel):
|
||||||
self.user = get_user(request.session[SESSION_KEY_IMPERSONATE_ORIGINAL_USER])
|
self.user = get_user(request.session[SESSION_KEY_IMPERSONATE_ORIGINAL_USER])
|
||||||
self.user["on_behalf_of"] = get_user(request.session[SESSION_KEY_IMPERSONATE_USER])
|
self.user["on_behalf_of"] = get_user(request.session[SESSION_KEY_IMPERSONATE_USER])
|
||||||
# User 255.255.255.255 as fallback if IP cannot be determined
|
# User 255.255.255.255 as fallback if IP cannot be determined
|
||||||
self.client_ip = ClientIPMiddleware.get_client_ip(request)
|
self.client_ip = get_client_ip(request)
|
||||||
# Enrich event data
|
# Apply GeoIP Data, when enabled
|
||||||
for processor in get_context_processors():
|
self.with_geoip()
|
||||||
processor.enrich_event(self)
|
|
||||||
# If there's no app set, we get it from the requests too
|
# If there's no app set, we get it from the requests too
|
||||||
if not self.app:
|
if not self.app:
|
||||||
self.app = Event._get_app_from_request(request)
|
self.app = Event._get_app_from_request(request)
|
||||||
self.save()
|
self.save()
|
||||||
return self
|
return self
|
||||||
|
|
||||||
|
def with_geoip(self): # pragma: no cover
|
||||||
|
"""Apply GeoIP Data, when enabled"""
|
||||||
|
city = GEOIP_READER.city_dict(self.client_ip)
|
||||||
|
if not city:
|
||||||
|
return
|
||||||
|
self.context["geo"] = city
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, **kwargs):
|
||||||
# Creating a batch for this event in the save method
|
|
||||||
batch = EventBatch.get_or_create_batch(self.action, self.user, self.app)
|
|
||||||
self.batch_id = batch.batch_id
|
|
||||||
if self._state.adding:
|
if self._state.adding:
|
||||||
LOGGER.info(
|
LOGGER.info(
|
||||||
"Created Event",
|
"Created Event",
|
||||||
|
@ -378,17 +320,7 @@ class NotificationTransport(SerializerModel):
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
enable_batching = models.BooleanField(default=False)
|
|
||||||
batch_timeout = models.IntegerField(default=60) # Timeout in seconds
|
|
||||||
max_batch_size = models.IntegerField(default=10)
|
|
||||||
|
|
||||||
def send(self, notification: "Notification") -> list[str]:
|
def send(self, notification: "Notification") -> list[str]:
|
||||||
"""Send a batched notification or a single notification"""
|
|
||||||
if self.enable_batching:
|
|
||||||
return self.process_batch(notification)
|
|
||||||
return self.send_notification(notification)
|
|
||||||
|
|
||||||
def send_notification(self, notification: "Notification") -> list[str]:
|
|
||||||
"""Send notification to user, called from async task"""
|
"""Send notification to user, called from async task"""
|
||||||
if self.mode == TransportMode.LOCAL:
|
if self.mode == TransportMode.LOCAL:
|
||||||
return self.send_local(notification)
|
return self.send_local(notification)
|
||||||
|
@ -534,7 +466,7 @@ class NotificationTransport(SerializerModel):
|
||||||
}
|
}
|
||||||
mail = TemplateEmailMessage(
|
mail = TemplateEmailMessage(
|
||||||
subject=subject_prefix + context["title"],
|
subject=subject_prefix + context["title"],
|
||||||
to=[f"{notification.user.name} <{notification.user.email}>"],
|
to=[notification.user.email],
|
||||||
language=notification.user.locale(),
|
language=notification.user.locale(),
|
||||||
template_name="email/event_notification.html",
|
template_name="email/event_notification.html",
|
||||||
template_context=context,
|
template_context=context,
|
||||||
|
|
|
@ -45,14 +45,9 @@ def get_login_event(request: HttpRequest) -> Optional[Event]:
|
||||||
|
|
||||||
|
|
||||||
@receiver(user_logged_out)
|
@receiver(user_logged_out)
|
||||||
def on_user_logged_out(sender, request: HttpRequest, user: User, **kwargs):
|
def on_user_logged_out(sender, request: HttpRequest, user: User, **_):
|
||||||
"""Log successfully logout"""
|
"""Log successfully logout"""
|
||||||
# Check if this even comes from the user_login stage's middleware, which will set an extra
|
Event.new(EventAction.LOGOUT).from_http(request, user=user)
|
||||||
# argument
|
|
||||||
event = Event.new(EventAction.LOGOUT)
|
|
||||||
if "event_extra" in kwargs:
|
|
||||||
event.context.update(kwargs["event_extra"])
|
|
||||||
event.from_http(request, user=user)
|
|
||||||
|
|
||||||
|
|
||||||
@receiver(user_write)
|
@receiver(user_write)
|
||||||
|
|
|
@ -2,13 +2,11 @@
|
||||||
from typing import Optional
|
from typing import Optional
|
||||||
|
|
||||||
from django.db.models.query_utils import Q
|
from django.db.models.query_utils import Q
|
||||||
from django.utils import timezone
|
|
||||||
from guardian.shortcuts import get_anonymous_user
|
from guardian.shortcuts import get_anonymous_user
|
||||||
from structlog.stdlib import get_logger
|
from structlog.stdlib import get_logger
|
||||||
|
|
||||||
from authentik.core.exceptions import PropertyMappingExpressionException
|
from authentik.core.exceptions import PropertyMappingExpressionException
|
||||||
from authentik.core.models import User
|
from authentik.core.models import User
|
||||||
from authentik.events.models import EventBatch # Importing the EventBatch model
|
|
||||||
from authentik.events.models import (
|
from authentik.events.models import (
|
||||||
Event,
|
Event,
|
||||||
Notification,
|
Notification,
|
||||||
|
@ -21,7 +19,6 @@ from authentik.events.monitored_tasks import (
|
||||||
TaskResult,
|
TaskResult,
|
||||||
TaskResultStatus,
|
TaskResultStatus,
|
||||||
prefill_task,
|
prefill_task,
|
||||||
shared_task,
|
|
||||||
)
|
)
|
||||||
from authentik.policies.engine import PolicyEngine
|
from authentik.policies.engine import PolicyEngine
|
||||||
from authentik.policies.models import PolicyBinding, PolicyEngineMode
|
from authentik.policies.models import PolicyBinding, PolicyEngineMode
|
||||||
|
@ -113,36 +110,19 @@ def notification_transport(
|
||||||
event = Event.objects.filter(pk=event_pk).first()
|
event = Event.objects.filter(pk=event_pk).first()
|
||||||
if not event:
|
if not event:
|
||||||
return
|
return
|
||||||
|
|
||||||
user = User.objects.filter(pk=user_pk).first()
|
user = User.objects.filter(pk=user_pk).first()
|
||||||
if not user:
|
if not user:
|
||||||
return
|
return
|
||||||
trigger = NotificationRule.objects.filter(pk=trigger_pk).first()
|
trigger = NotificationRule.objects.filter(pk=trigger_pk).first()
|
||||||
if not trigger:
|
if not trigger:
|
||||||
return
|
return
|
||||||
|
notification = Notification(
|
||||||
# Check if batching is enabled and process accordingly
|
severity=trigger.severity, body=event.summary, event=event, user=user
|
||||||
|
)
|
||||||
transport = NotificationTransport.objects.filter(pk=transport_pk).first()
|
transport = NotificationTransport.objects.filter(pk=transport_pk).first()
|
||||||
if transport and transport.enable_batching:
|
if not transport:
|
||||||
# Process the event for batching
|
return
|
||||||
batch = EventBatch.get_or_create_batch(event.action, event.app, event.user)
|
transport.send(notification)
|
||||||
batch.add_event_to_batch(event)
|
|
||||||
# Check if the batch has reached its limits
|
|
||||||
if not batch.check_batch_limits():
|
|
||||||
return
|
|
||||||
|
|
||||||
batch_summary = batch.process_batch()
|
|
||||||
batch.delete()
|
|
||||||
notification = Notification(
|
|
||||||
severity=trigger.severity, body=batch_summary, event=event, user=user
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
notification = Notification(
|
|
||||||
severity=trigger.severity, body=event.summary, event=event, user=user
|
|
||||||
)
|
|
||||||
|
|
||||||
transport.send_notification(notification)
|
|
||||||
|
|
||||||
self.set_status(TaskResult(TaskResultStatus.SUCCESSFUL))
|
self.set_status(TaskResult(TaskResultStatus.SUCCESSFUL))
|
||||||
except (NotificationTransportError, PropertyMappingExpressionException) as exc:
|
except (NotificationTransportError, PropertyMappingExpressionException) as exc:
|
||||||
self.set_status(TaskResult(TaskResultStatus.ERROR).with_error(exc))
|
self.set_status(TaskResult(TaskResultStatus.ERROR).with_error(exc))
|
||||||
|
@ -167,16 +147,3 @@ def notification_cleanup(self: MonitoredTask):
|
||||||
notification.delete()
|
notification.delete()
|
||||||
LOGGER.debug("Expired notifications", amount=amount)
|
LOGGER.debug("Expired notifications", amount=amount)
|
||||||
self.set_status(TaskResult(TaskResultStatus.SUCCESSFUL, [f"Expired {amount} Notifications"]))
|
self.set_status(TaskResult(TaskResultStatus.SUCCESSFUL, [f"Expired {amount} Notifications"]))
|
||||||
|
|
||||||
# Scheduled task to check and send pending batches
|
|
||||||
@CELERY_APP.task(base=MonitoredTask)
|
|
||||||
@shared_task
|
|
||||||
def check_and_send_pending_batches():
|
|
||||||
"""Check for pending batches that haven't been sent and have been idle for a specified time."""
|
|
||||||
idle_time = timezone.now() - timedelta(minutes=10) # Example idle time
|
|
||||||
pending_batches = EventBatch.objects.filter(sent=False, last_updated__lt=idle_time)
|
|
||||||
for batch in pending_batches:
|
|
||||||
batch.send_notification()
|
|
||||||
batch.sent = True
|
|
||||||
batch.save()
|
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
"""Event API tests"""
|
"""Event API tests"""
|
||||||
from json import loads
|
|
||||||
|
|
||||||
from django.urls import reverse
|
from django.urls import reverse
|
||||||
from rest_framework.test import APITestCase
|
from rest_framework.test import APITestCase
|
||||||
|
@ -12,9 +11,6 @@ from authentik.events.models import (
|
||||||
NotificationSeverity,
|
NotificationSeverity,
|
||||||
TransportMode,
|
TransportMode,
|
||||||
)
|
)
|
||||||
from authentik.events.utils import model_to_dict
|
|
||||||
from authentik.lib.generators import generate_id
|
|
||||||
from authentik.providers.oauth2.models import OAuth2Provider
|
|
||||||
|
|
||||||
|
|
||||||
class TestEventsAPI(APITestCase):
|
class TestEventsAPI(APITestCase):
|
||||||
|
@ -24,25 +20,6 @@ class TestEventsAPI(APITestCase):
|
||||||
self.user = create_test_admin_user()
|
self.user = create_test_admin_user()
|
||||||
self.client.force_login(self.user)
|
self.client.force_login(self.user)
|
||||||
|
|
||||||
def test_filter_model_pk_int(self):
|
|
||||||
"""Test event list with context_model_pk and integer PKs"""
|
|
||||||
provider = OAuth2Provider.objects.create(
|
|
||||||
name=generate_id(),
|
|
||||||
)
|
|
||||||
event = Event.new(EventAction.MODEL_CREATED, model=model_to_dict(provider))
|
|
||||||
event.save()
|
|
||||||
response = self.client.get(
|
|
||||||
reverse("authentik_api:event-list"),
|
|
||||||
data={
|
|
||||||
"context_model_pk": provider.pk,
|
|
||||||
"context_model_app": "authentik_providers_oauth2",
|
|
||||||
"context_model_name": "oauth2provider",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
self.assertEqual(response.status_code, 200)
|
|
||||||
body = loads(response.content)
|
|
||||||
self.assertEqual(body["pagination"]["count"], 1)
|
|
||||||
|
|
||||||
def test_top_n(self):
|
def test_top_n(self):
|
||||||
"""Test top_per_user"""
|
"""Test top_per_user"""
|
||||||
event = Event.new(EventAction.AUTHORIZE_APPLICATION)
|
event = Event.new(EventAction.AUTHORIZE_APPLICATION)
|
||||||
|
|
|
@ -1,24 +0,0 @@
|
||||||
"""Test ASN Wrapper"""
|
|
||||||
from django.test import TestCase
|
|
||||||
|
|
||||||
from authentik.events.context_processors.asn import ASNContextProcessor
|
|
||||||
|
|
||||||
|
|
||||||
class TestASN(TestCase):
|
|
||||||
"""Test ASN Wrapper"""
|
|
||||||
|
|
||||||
def setUp(self) -> None:
|
|
||||||
self.reader = ASNContextProcessor()
|
|
||||||
|
|
||||||
def test_simple(self):
|
|
||||||
"""Test simple asn wrapper"""
|
|
||||||
# IPs from
|
|
||||||
# https://github.com/maxmind/MaxMind-DB/blob/main/source-data/GeoLite2-ASN-Test.json
|
|
||||||
self.assertEqual(
|
|
||||||
self.reader.asn_dict("1.0.0.1"),
|
|
||||||
{
|
|
||||||
"asn": 15169,
|
|
||||||
"as_org": "Google Inc.",
|
|
||||||
"network": "1.0.0.0/24",
|
|
||||||
},
|
|
||||||
)
|
|
|
@ -1,14 +1,14 @@
|
||||||
"""Test GeoIP Wrapper"""
|
"""Test GeoIP Wrapper"""
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
from authentik.events.context_processors.geoip import GeoIPContextProcessor
|
from authentik.events.geo import GeoIPReader
|
||||||
|
|
||||||
|
|
||||||
class TestGeoIP(TestCase):
|
class TestGeoIP(TestCase):
|
||||||
"""Test GeoIP Wrapper"""
|
"""Test GeoIP Wrapper"""
|
||||||
|
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
self.reader = GeoIPContextProcessor()
|
self.reader = GeoIPReader()
|
||||||
|
|
||||||
def test_simple(self):
|
def test_simple(self):
|
||||||
"""Test simple city wrapper"""
|
"""Test simple city wrapper"""
|
|
@ -17,13 +17,12 @@ from django.db.models.base import Model
|
||||||
from django.http.request import HttpRequest
|
from django.http.request import HttpRequest
|
||||||
from django.utils import timezone
|
from django.utils import timezone
|
||||||
from django.views.debug import SafeExceptionReporterFilter
|
from django.views.debug import SafeExceptionReporterFilter
|
||||||
from geoip2.models import ASN, City
|
from geoip2.models import City
|
||||||
from guardian.utils import get_anonymous_user
|
from guardian.utils import get_anonymous_user
|
||||||
|
|
||||||
from authentik.blueprints.v1.common import YAMLTag
|
from authentik.blueprints.v1.common import YAMLTag
|
||||||
from authentik.core.models import User
|
from authentik.core.models import User
|
||||||
from authentik.events.context_processors.asn import ASN_CONTEXT_PROCESSOR
|
from authentik.events.geo import GEOIP_READER
|
||||||
from authentik.events.context_processors.geoip import GEOIP_CONTEXT_PROCESSOR
|
|
||||||
from authentik.policies.types import PolicyRequest
|
from authentik.policies.types import PolicyRequest
|
||||||
|
|
||||||
# Special keys which are *not* cleaned, even when the default filter
|
# Special keys which are *not* cleaned, even when the default filter
|
||||||
|
@ -124,9 +123,7 @@ def sanitize_item(value: Any) -> Any:
|
||||||
if isinstance(value, (HttpRequest, WSGIRequest)):
|
if isinstance(value, (HttpRequest, WSGIRequest)):
|
||||||
return ...
|
return ...
|
||||||
if isinstance(value, City):
|
if isinstance(value, City):
|
||||||
return GEOIP_CONTEXT_PROCESSOR.city_to_dict(value)
|
return GEOIP_READER.city_to_dict(value)
|
||||||
if isinstance(value, ASN):
|
|
||||||
return ASN_CONTEXT_PROCESSOR.asn_to_dict(value)
|
|
||||||
if isinstance(value, Path):
|
if isinstance(value, Path):
|
||||||
return str(value)
|
return str(value)
|
||||||
if isinstance(value, Exception):
|
if isinstance(value, Exception):
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
# Generated by Django 4.2.6 on 2023-10-28 14:24
|
# Generated by Django 4.2.6 on 2023-10-28 14:24
|
||||||
|
|
||||||
from django.apps.registry import Apps
|
from django.apps.registry import Apps
|
||||||
from django.db import migrations, models
|
from django.db import migrations
|
||||||
from django.db.backends.base.schema import BaseDatabaseSchemaEditor
|
from django.db.backends.base.schema import BaseDatabaseSchemaEditor
|
||||||
|
|
||||||
|
|
||||||
|
@ -31,19 +31,4 @@ class Migration(migrations.Migration):
|
||||||
|
|
||||||
operations = [
|
operations = [
|
||||||
migrations.RunPython(set_oobe_flow_authentication),
|
migrations.RunPython(set_oobe_flow_authentication),
|
||||||
migrations.AlterField(
|
|
||||||
model_name="flow",
|
|
||||||
name="authentication",
|
|
||||||
field=models.TextField(
|
|
||||||
choices=[
|
|
||||||
("none", "None"),
|
|
||||||
("require_authenticated", "Require Authenticated"),
|
|
||||||
("require_unauthenticated", "Require Unauthenticated"),
|
|
||||||
("require_superuser", "Require Superuser"),
|
|
||||||
("require_outpost", "Require Outpost"),
|
|
||||||
],
|
|
||||||
default="none",
|
|
||||||
help_text="Required level of authentication and authorization to access a flow.",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
]
|
]
|
||||||
|
|
|
@ -31,7 +31,6 @@ class FlowAuthenticationRequirement(models.TextChoices):
|
||||||
REQUIRE_AUTHENTICATED = "require_authenticated"
|
REQUIRE_AUTHENTICATED = "require_authenticated"
|
||||||
REQUIRE_UNAUTHENTICATED = "require_unauthenticated"
|
REQUIRE_UNAUTHENTICATED = "require_unauthenticated"
|
||||||
REQUIRE_SUPERUSER = "require_superuser"
|
REQUIRE_SUPERUSER = "require_superuser"
|
||||||
REQUIRE_OUTPOST = "require_outpost"
|
|
||||||
|
|
||||||
|
|
||||||
class NotConfiguredAction(models.TextChoices):
|
class NotConfiguredAction(models.TextChoices):
|
||||||
|
|
|
@ -23,7 +23,6 @@ from authentik.flows.models import (
|
||||||
)
|
)
|
||||||
from authentik.lib.config import CONFIG
|
from authentik.lib.config import CONFIG
|
||||||
from authentik.policies.engine import PolicyEngine
|
from authentik.policies.engine import PolicyEngine
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
|
|
||||||
LOGGER = get_logger()
|
LOGGER = get_logger()
|
||||||
PLAN_CONTEXT_PENDING_USER = "pending_user"
|
PLAN_CONTEXT_PENDING_USER = "pending_user"
|
||||||
|
@ -34,7 +33,7 @@ PLAN_CONTEXT_SOURCE = "source"
|
||||||
# Is set by the Flow Planner when a FlowToken was used, and the currently active flow plan
|
# Is set by the Flow Planner when a FlowToken was used, and the currently active flow plan
|
||||||
# was restored.
|
# was restored.
|
||||||
PLAN_CONTEXT_IS_RESTORED = "is_restored"
|
PLAN_CONTEXT_IS_RESTORED = "is_restored"
|
||||||
CACHE_TIMEOUT = CONFIG.get_int("cache.timeout_flows")
|
CACHE_TIMEOUT = CONFIG.get_int("redis.cache_timeout_flows")
|
||||||
CACHE_PREFIX = "goauthentik.io/flows/planner/"
|
CACHE_PREFIX = "goauthentik.io/flows/planner/"
|
||||||
|
|
||||||
|
|
||||||
|
@ -142,10 +141,6 @@ class FlowPlanner:
|
||||||
and not request.user.is_superuser
|
and not request.user.is_superuser
|
||||||
):
|
):
|
||||||
raise FlowNonApplicableException()
|
raise FlowNonApplicableException()
|
||||||
if self.flow.authentication == FlowAuthenticationRequirement.REQUIRE_OUTPOST:
|
|
||||||
outpost_user = ClientIPMiddleware.get_outpost_user(request)
|
|
||||||
if not outpost_user:
|
|
||||||
raise FlowNonApplicableException()
|
|
||||||
|
|
||||||
def plan(
|
def plan(
|
||||||
self, request: HttpRequest, default_context: Optional[dict[str, Any]] = None
|
self, request: HttpRequest, default_context: Optional[dict[str, Any]] = None
|
||||||
|
|
|
@ -472,7 +472,6 @@ class TestFlowExecutor(FlowTestCase):
|
||||||
ident_stage = IdentificationStage.objects.create(
|
ident_stage = IdentificationStage.objects.create(
|
||||||
name="ident",
|
name="ident",
|
||||||
user_fields=[UserFields.E_MAIL],
|
user_fields=[UserFields.E_MAIL],
|
||||||
pretend_user_exists=False,
|
|
||||||
)
|
)
|
||||||
FlowStageBinding.objects.create(
|
FlowStageBinding.objects.create(
|
||||||
target=flow,
|
target=flow,
|
||||||
|
|
|
@ -8,7 +8,6 @@ from django.test import RequestFactory, TestCase
|
||||||
from django.urls import reverse
|
from django.urls import reverse
|
||||||
from guardian.shortcuts import get_anonymous_user
|
from guardian.shortcuts import get_anonymous_user
|
||||||
|
|
||||||
from authentik.blueprints.tests import reconcile_app
|
|
||||||
from authentik.core.models import User
|
from authentik.core.models import User
|
||||||
from authentik.core.tests.utils import create_test_admin_user, create_test_flow
|
from authentik.core.tests.utils import create_test_admin_user, create_test_flow
|
||||||
from authentik.flows.exceptions import EmptyFlowException, FlowNonApplicableException
|
from authentik.flows.exceptions import EmptyFlowException, FlowNonApplicableException
|
||||||
|
@ -16,12 +15,9 @@ from authentik.flows.markers import ReevaluateMarker, StageMarker
|
||||||
from authentik.flows.models import FlowAuthenticationRequirement, FlowDesignation, FlowStageBinding
|
from authentik.flows.models import FlowAuthenticationRequirement, FlowDesignation, FlowStageBinding
|
||||||
from authentik.flows.planner import PLAN_CONTEXT_PENDING_USER, FlowPlanner, cache_key
|
from authentik.flows.planner import PLAN_CONTEXT_PENDING_USER, FlowPlanner, cache_key
|
||||||
from authentik.lib.tests.utils import dummy_get_response
|
from authentik.lib.tests.utils import dummy_get_response
|
||||||
from authentik.outposts.apps import MANAGED_OUTPOST
|
|
||||||
from authentik.outposts.models import Outpost
|
|
||||||
from authentik.policies.dummy.models import DummyPolicy
|
from authentik.policies.dummy.models import DummyPolicy
|
||||||
from authentik.policies.models import PolicyBinding
|
from authentik.policies.models import PolicyBinding
|
||||||
from authentik.policies.types import PolicyResult
|
from authentik.policies.types import PolicyResult
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
from authentik.stages.dummy.models import DummyStage
|
from authentik.stages.dummy.models import DummyStage
|
||||||
|
|
||||||
POLICY_RETURN_FALSE = PropertyMock(return_value=PolicyResult(False))
|
POLICY_RETURN_FALSE = PropertyMock(return_value=PolicyResult(False))
|
||||||
|
@ -72,34 +68,6 @@ class TestFlowPlanner(TestCase):
|
||||||
planner.allow_empty_flows = True
|
planner.allow_empty_flows = True
|
||||||
planner.plan(request)
|
planner.plan(request)
|
||||||
|
|
||||||
@reconcile_app("authentik_outposts")
|
|
||||||
def test_authentication_outpost(self):
|
|
||||||
"""Test flow authentication (outpost)"""
|
|
||||||
flow = create_test_flow()
|
|
||||||
flow.authentication = FlowAuthenticationRequirement.REQUIRE_OUTPOST
|
|
||||||
request = self.request_factory.get(
|
|
||||||
reverse("authentik_api:flow-executor", kwargs={"flow_slug": flow.slug}),
|
|
||||||
)
|
|
||||||
request.user = AnonymousUser()
|
|
||||||
with self.assertRaises(FlowNonApplicableException):
|
|
||||||
planner = FlowPlanner(flow)
|
|
||||||
planner.allow_empty_flows = True
|
|
||||||
planner.plan(request)
|
|
||||||
|
|
||||||
outpost = Outpost.objects.filter(managed=MANAGED_OUTPOST).first()
|
|
||||||
request = self.request_factory.get(
|
|
||||||
reverse("authentik_api:flow-executor", kwargs={"flow_slug": flow.slug}),
|
|
||||||
HTTP_X_AUTHENTIK_OUTPOST_TOKEN=outpost.token.key,
|
|
||||||
HTTP_X_AUTHENTIK_REMOTE_IP="1.2.3.4",
|
|
||||||
)
|
|
||||||
request.user = AnonymousUser()
|
|
||||||
middleware = ClientIPMiddleware(dummy_get_response)
|
|
||||||
middleware(request)
|
|
||||||
|
|
||||||
planner = FlowPlanner(flow)
|
|
||||||
planner.allow_empty_flows = True
|
|
||||||
planner.plan(request)
|
|
||||||
|
|
||||||
@patch(
|
@patch(
|
||||||
"authentik.policies.engine.PolicyEngine.result",
|
"authentik.policies.engine.PolicyEngine.result",
|
||||||
POLICY_RETURN_FALSE,
|
POLICY_RETURN_FALSE,
|
||||||
|
|
|
@ -154,15 +154,7 @@ def generate_avatar_from_name(
|
||||||
|
|
||||||
def avatar_mode_generated(user: "User", mode: str) -> Optional[str]:
|
def avatar_mode_generated(user: "User", mode: str) -> Optional[str]:
|
||||||
"""Wrapper that converts generated avatar to base64 svg"""
|
"""Wrapper that converts generated avatar to base64 svg"""
|
||||||
# By default generate based off of user's display name
|
svg = generate_avatar_from_name(user.name if user.name.strip() != "" else "a k")
|
||||||
name = user.name.strip()
|
|
||||||
if name == "":
|
|
||||||
# Fallback to username
|
|
||||||
name = user.username.strip()
|
|
||||||
# If we still don't have anything, fallback to `a k`
|
|
||||||
if name == "":
|
|
||||||
name = "a k"
|
|
||||||
svg = generate_avatar_from_name(name)
|
|
||||||
return f"data:image/svg+xml;base64,{b64encode(svg.encode('utf-8')).decode('utf-8')}"
|
return f"data:image/svg+xml;base64,{b64encode(svg.encode('utf-8')).decode('utf-8')}"
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,6 +1,4 @@
|
||||||
"""authentik core config loader"""
|
"""authentik core config loader"""
|
||||||
import base64
|
|
||||||
import json
|
|
||||||
import os
|
import os
|
||||||
from collections.abc import Mapping
|
from collections.abc import Mapping
|
||||||
from contextlib import contextmanager
|
from contextlib import contextmanager
|
||||||
|
@ -24,26 +22,6 @@ SEARCH_PATHS = ["authentik/lib/default.yml", "/etc/authentik/config.yml", ""] +
|
||||||
ENV_PREFIX = "AUTHENTIK"
|
ENV_PREFIX = "AUTHENTIK"
|
||||||
ENVIRONMENT = os.getenv(f"{ENV_PREFIX}_ENV", "local")
|
ENVIRONMENT = os.getenv(f"{ENV_PREFIX}_ENV", "local")
|
||||||
|
|
||||||
REDIS_ENV_KEYS = [
|
|
||||||
f"{ENV_PREFIX}_REDIS__HOST",
|
|
||||||
f"{ENV_PREFIX}_REDIS__PORT",
|
|
||||||
f"{ENV_PREFIX}_REDIS__DB",
|
|
||||||
f"{ENV_PREFIX}_REDIS__USERNAME",
|
|
||||||
f"{ENV_PREFIX}_REDIS__PASSWORD",
|
|
||||||
f"{ENV_PREFIX}_REDIS__TLS",
|
|
||||||
f"{ENV_PREFIX}_REDIS__TLS_REQS",
|
|
||||||
]
|
|
||||||
|
|
||||||
DEPRECATIONS = {
|
|
||||||
"geoip": "events.context_processors.geoip",
|
|
||||||
"redis.broker_url": "broker.url",
|
|
||||||
"redis.broker_transport_options": "broker.transport_options",
|
|
||||||
"redis.cache_timeout": "cache.timeout",
|
|
||||||
"redis.cache_timeout_flows": "cache.timeout_flows",
|
|
||||||
"redis.cache_timeout_policies": "cache.timeout_policies",
|
|
||||||
"redis.cache_timeout_reputation": "cache.timeout_reputation",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def get_path_from_dict(root: dict, path: str, sep=".", default=None) -> Any:
|
def get_path_from_dict(root: dict, path: str, sep=".", default=None) -> Any:
|
||||||
"""Recursively walk through `root`, checking each part of `path` separated by `sep`.
|
"""Recursively walk through `root`, checking each part of `path` separated by `sep`.
|
||||||
|
@ -103,18 +81,12 @@ class AttrEncoder(JSONEncoder):
|
||||||
return super().default(o)
|
return super().default(o)
|
||||||
|
|
||||||
|
|
||||||
class UNSET:
|
|
||||||
"""Used to test whether configuration key has not been set."""
|
|
||||||
|
|
||||||
|
|
||||||
class ConfigLoader:
|
class ConfigLoader:
|
||||||
"""Search through SEARCH_PATHS and load configuration. Environment variables starting with
|
"""Search through SEARCH_PATHS and load configuration. Environment variables starting with
|
||||||
`ENV_PREFIX` are also applied.
|
`ENV_PREFIX` are also applied.
|
||||||
|
|
||||||
A variable like AUTHENTIK_POSTGRESQL__HOST would translate to postgresql.host"""
|
A variable like AUTHENTIK_POSTGRESQL__HOST would translate to postgresql.host"""
|
||||||
|
|
||||||
deprecations: dict[tuple[str, str], str] = {}
|
|
||||||
|
|
||||||
def __init__(self, **kwargs):
|
def __init__(self, **kwargs):
|
||||||
super().__init__()
|
super().__init__()
|
||||||
self.__config = {}
|
self.__config = {}
|
||||||
|
@ -141,38 +113,6 @@ class ConfigLoader:
|
||||||
self.update_from_file(env_file)
|
self.update_from_file(env_file)
|
||||||
self.update_from_env()
|
self.update_from_env()
|
||||||
self.update(self.__config, kwargs)
|
self.update(self.__config, kwargs)
|
||||||
self.deprecations = self.check_deprecations()
|
|
||||||
|
|
||||||
def check_deprecations(self) -> dict[str, str]:
|
|
||||||
"""Warn if any deprecated configuration options are used"""
|
|
||||||
|
|
||||||
def _pop_deprecated_key(current_obj, dot_parts, index):
|
|
||||||
"""Recursive function to remove deprecated keys in configuration"""
|
|
||||||
dot_part = dot_parts[index]
|
|
||||||
if index == len(dot_parts) - 1:
|
|
||||||
return current_obj.pop(dot_part)
|
|
||||||
value = _pop_deprecated_key(current_obj[dot_part], dot_parts, index + 1)
|
|
||||||
if not current_obj[dot_part]:
|
|
||||||
current_obj.pop(dot_part)
|
|
||||||
return value
|
|
||||||
|
|
||||||
deprecation_replacements = {}
|
|
||||||
for deprecation, replacement in DEPRECATIONS.items():
|
|
||||||
if self.get(deprecation, default=UNSET) is UNSET:
|
|
||||||
continue
|
|
||||||
message = (
|
|
||||||
f"'{deprecation}' has been deprecated in favor of '{replacement}'! "
|
|
||||||
+ "Please update your configuration."
|
|
||||||
)
|
|
||||||
self.log(
|
|
||||||
"warning",
|
|
||||||
message,
|
|
||||||
)
|
|
||||||
deprecation_replacements[(deprecation, replacement)] = message
|
|
||||||
|
|
||||||
deprecated_attr = _pop_deprecated_key(self.__config, deprecation.split("."), 0)
|
|
||||||
self.set(replacement, deprecated_attr)
|
|
||||||
return deprecation_replacements
|
|
||||||
|
|
||||||
def log(self, level: str, message: str, **kwargs):
|
def log(self, level: str, message: str, **kwargs):
|
||||||
"""Custom Log method, we want to ensure ConfigLoader always logs JSON even when
|
"""Custom Log method, we want to ensure ConfigLoader always logs JSON even when
|
||||||
|
@ -240,10 +180,6 @@ class ConfigLoader:
|
||||||
error=str(exc),
|
error=str(exc),
|
||||||
)
|
)
|
||||||
|
|
||||||
def update_from_dict(self, update: dict):
|
|
||||||
"""Update config from dict"""
|
|
||||||
self.__config.update(update)
|
|
||||||
|
|
||||||
def update_from_env(self):
|
def update_from_env(self):
|
||||||
"""Check environment variables"""
|
"""Check environment variables"""
|
||||||
outer = {}
|
outer = {}
|
||||||
|
@ -252,13 +188,19 @@ class ConfigLoader:
|
||||||
if not key.startswith(ENV_PREFIX):
|
if not key.startswith(ENV_PREFIX):
|
||||||
continue
|
continue
|
||||||
relative_key = key.replace(f"{ENV_PREFIX}_", "", 1).replace("__", ".").lower()
|
relative_key = key.replace(f"{ENV_PREFIX}_", "", 1).replace("__", ".").lower()
|
||||||
|
# Recursively convert path from a.b.c into outer[a][b][c]
|
||||||
|
current_obj = outer
|
||||||
|
dot_parts = relative_key.split(".")
|
||||||
|
for dot_part in dot_parts[:-1]:
|
||||||
|
if dot_part not in current_obj:
|
||||||
|
current_obj[dot_part] = {}
|
||||||
|
current_obj = current_obj[dot_part]
|
||||||
# Check if the value is json, and try to load it
|
# Check if the value is json, and try to load it
|
||||||
try:
|
try:
|
||||||
value = loads(value)
|
value = loads(value)
|
||||||
except JSONDecodeError:
|
except JSONDecodeError:
|
||||||
pass
|
pass
|
||||||
attr_value = Attr(value, Attr.Source.ENV, relative_key)
|
current_obj[dot_parts[-1]] = Attr(value, Attr.Source.ENV, key)
|
||||||
set_path_in_dict(outer, relative_key, attr_value)
|
|
||||||
idx += 1
|
idx += 1
|
||||||
if idx > 0:
|
if idx > 0:
|
||||||
self.log("debug", "Loaded environment variables", count=idx)
|
self.log("debug", "Loaded environment variables", count=idx)
|
||||||
|
@ -299,28 +241,9 @@ class ConfigLoader:
|
||||||
"""Wrapper for get that converts value into boolean"""
|
"""Wrapper for get that converts value into boolean"""
|
||||||
return str(self.get(path, default)).lower() == "true"
|
return str(self.get(path, default)).lower() == "true"
|
||||||
|
|
||||||
def get_dict_from_b64_json(self, path: str, default=None) -> dict:
|
|
||||||
"""Wrapper for get that converts value from Base64 encoded string into dictionary"""
|
|
||||||
config_value = self.get(path)
|
|
||||||
if config_value is None:
|
|
||||||
return {}
|
|
||||||
try:
|
|
||||||
b64decoded_str = base64.b64decode(config_value).decode("utf-8")
|
|
||||||
b64decoded_str = b64decoded_str.strip().lstrip("{").rstrip("}")
|
|
||||||
b64decoded_str = "{" + b64decoded_str + "}"
|
|
||||||
return json.loads(b64decoded_str)
|
|
||||||
except (JSONDecodeError, TypeError, ValueError) as exc:
|
|
||||||
self.log(
|
|
||||||
"warning",
|
|
||||||
f"Ignored invalid configuration for '{path}' due to exception: {str(exc)}",
|
|
||||||
)
|
|
||||||
return default if isinstance(default, dict) else {}
|
|
||||||
|
|
||||||
def set(self, path: str, value: Any, sep="."):
|
def set(self, path: str, value: Any, sep="."):
|
||||||
"""Set value using same syntax as get()"""
|
"""Set value using same syntax as get()"""
|
||||||
if not isinstance(value, Attr):
|
set_path_in_dict(self.raw, path, Attr(value), sep=sep)
|
||||||
value = Attr(value)
|
|
||||||
set_path_in_dict(self.raw, path, value, sep=sep)
|
|
||||||
|
|
||||||
|
|
||||||
CONFIG = ConfigLoader()
|
CONFIG = ConfigLoader()
|
||||||
|
|
|
@ -8,8 +8,6 @@ postgresql:
|
||||||
password: "env://POSTGRES_PASSWORD"
|
password: "env://POSTGRES_PASSWORD"
|
||||||
use_pgbouncer: false
|
use_pgbouncer: false
|
||||||
use_pgpool: false
|
use_pgpool: false
|
||||||
test:
|
|
||||||
name: test_authentik
|
|
||||||
|
|
||||||
listen:
|
listen:
|
||||||
listen_http: 0.0.0.0:9000
|
listen_http: 0.0.0.0:9000
|
||||||
|
@ -30,28 +28,14 @@ listen:
|
||||||
redis:
|
redis:
|
||||||
host: localhost
|
host: localhost
|
||||||
port: 6379
|
port: 6379
|
||||||
db: 0
|
|
||||||
username: ""
|
|
||||||
password: ""
|
password: ""
|
||||||
tls: false
|
tls: false
|
||||||
tls_reqs: "none"
|
tls_reqs: "none"
|
||||||
|
db: 0
|
||||||
# broker:
|
cache_timeout: 300
|
||||||
# url: ""
|
cache_timeout_flows: 300
|
||||||
# transport_options: ""
|
cache_timeout_policies: 300
|
||||||
|
cache_timeout_reputation: 300
|
||||||
cache:
|
|
||||||
# url: ""
|
|
||||||
timeout: 300
|
|
||||||
timeout_flows: 300
|
|
||||||
timeout_policies: 300
|
|
||||||
timeout_reputation: 300
|
|
||||||
|
|
||||||
# channel:
|
|
||||||
# url: ""
|
|
||||||
|
|
||||||
# result_backend:
|
|
||||||
# url: ""
|
|
||||||
|
|
||||||
paths:
|
paths:
|
||||||
media: ./media
|
media: ./media
|
||||||
|
@ -108,10 +92,7 @@ cookie_domain: null
|
||||||
disable_update_check: false
|
disable_update_check: false
|
||||||
disable_startup_analytics: false
|
disable_startup_analytics: false
|
||||||
avatars: env://AUTHENTIK_AUTHENTIK__AVATARS?gravatar,initials
|
avatars: env://AUTHENTIK_AUTHENTIK__AVATARS?gravatar,initials
|
||||||
events:
|
geoip: "/geoip/GeoLite2-City.mmdb"
|
||||||
context_processors:
|
|
||||||
geoip: "/geoip/GeoLite2-City.mmdb"
|
|
||||||
asn: "/geoip/GeoLite2-ASN.mmdb"
|
|
||||||
|
|
||||||
footer_links: []
|
footer_links: []
|
||||||
|
|
||||||
|
|
|
@ -1,32 +1,20 @@
|
||||||
"""Test config loader"""
|
"""Test config loader"""
|
||||||
import base64
|
|
||||||
from json import dumps
|
|
||||||
from os import chmod, environ, unlink, write
|
from os import chmod, environ, unlink, write
|
||||||
from tempfile import mkstemp
|
from tempfile import mkstemp
|
||||||
from unittest import mock
|
|
||||||
|
|
||||||
from django.conf import ImproperlyConfigured
|
from django.conf import ImproperlyConfigured
|
||||||
from django.test import TestCase
|
from django.test import TestCase
|
||||||
|
|
||||||
from authentik.lib.config import ENV_PREFIX, UNSET, Attr, AttrEncoder, ConfigLoader
|
from authentik.lib.config import ENV_PREFIX, ConfigLoader
|
||||||
|
|
||||||
|
|
||||||
class TestConfig(TestCase):
|
class TestConfig(TestCase):
|
||||||
"""Test config loader"""
|
"""Test config loader"""
|
||||||
|
|
||||||
check_deprecations_env_vars = {
|
|
||||||
ENV_PREFIX + "_REDIS__BROKER_URL": "redis://myredis:8327/43",
|
|
||||||
ENV_PREFIX + "_REDIS__BROKER_TRANSPORT_OPTIONS": "bWFzdGVybmFtZT1teW1hc3Rlcg==",
|
|
||||||
ENV_PREFIX + "_REDIS__CACHE_TIMEOUT": "124s",
|
|
||||||
ENV_PREFIX + "_REDIS__CACHE_TIMEOUT_FLOWS": "32m",
|
|
||||||
ENV_PREFIX + "_REDIS__CACHE_TIMEOUT_POLICIES": "3920ns",
|
|
||||||
ENV_PREFIX + "_REDIS__CACHE_TIMEOUT_REPUTATION": "298382us",
|
|
||||||
}
|
|
||||||
|
|
||||||
@mock.patch.dict(environ, {ENV_PREFIX + "_test__test": "bar"})
|
|
||||||
def test_env(self):
|
def test_env(self):
|
||||||
"""Test simple instance"""
|
"""Test simple instance"""
|
||||||
config = ConfigLoader()
|
config = ConfigLoader()
|
||||||
|
environ[ENV_PREFIX + "_test__test"] = "bar"
|
||||||
config.update_from_env()
|
config.update_from_env()
|
||||||
self.assertEqual(config.get("test.test"), "bar")
|
self.assertEqual(config.get("test.test"), "bar")
|
||||||
|
|
||||||
|
@ -39,20 +27,12 @@ class TestConfig(TestCase):
|
||||||
self.assertEqual(config.get("foo.bar"), "baz")
|
self.assertEqual(config.get("foo.bar"), "baz")
|
||||||
self.assertEqual(config.get("foo.bar"), "bar")
|
self.assertEqual(config.get("foo.bar"), "bar")
|
||||||
|
|
||||||
@mock.patch.dict(environ, {"foo": "bar"})
|
|
||||||
def test_uri_env(self):
|
def test_uri_env(self):
|
||||||
"""Test URI parsing (environment)"""
|
"""Test URI parsing (environment)"""
|
||||||
config = ConfigLoader()
|
config = ConfigLoader()
|
||||||
foo_uri = "env://foo"
|
environ["foo"] = "bar"
|
||||||
foo_parsed = config.parse_uri(foo_uri)
|
self.assertEqual(config.parse_uri("env://foo").value, "bar")
|
||||||
self.assertEqual(foo_parsed.value, "bar")
|
self.assertEqual(config.parse_uri("env://foo?bar").value, "bar")
|
||||||
self.assertEqual(foo_parsed.source_type, Attr.Source.URI)
|
|
||||||
self.assertEqual(foo_parsed.source, foo_uri)
|
|
||||||
foo_bar_uri = "env://foo?bar"
|
|
||||||
foo_bar_parsed = config.parse_uri(foo_bar_uri)
|
|
||||||
self.assertEqual(foo_bar_parsed.value, "bar")
|
|
||||||
self.assertEqual(foo_bar_parsed.source_type, Attr.Source.URI)
|
|
||||||
self.assertEqual(foo_bar_parsed.source, foo_bar_uri)
|
|
||||||
|
|
||||||
def test_uri_file(self):
|
def test_uri_file(self):
|
||||||
"""Test URI parsing (file load)"""
|
"""Test URI parsing (file load)"""
|
||||||
|
@ -111,60 +91,3 @@ class TestConfig(TestCase):
|
||||||
config = ConfigLoader()
|
config = ConfigLoader()
|
||||||
config.set("foo", "bar")
|
config.set("foo", "bar")
|
||||||
self.assertEqual(config.get_int("foo", 1234), 1234)
|
self.assertEqual(config.get_int("foo", 1234), 1234)
|
||||||
|
|
||||||
def test_get_dict_from_b64_json(self):
|
|
||||||
"""Test get_dict_from_b64_json"""
|
|
||||||
config = ConfigLoader()
|
|
||||||
test_value = ' { "foo": "bar" } '.encode("utf-8")
|
|
||||||
b64_value = base64.b64encode(test_value)
|
|
||||||
config.set("foo", b64_value)
|
|
||||||
self.assertEqual(config.get_dict_from_b64_json("foo"), {"foo": "bar"})
|
|
||||||
|
|
||||||
def test_get_dict_from_b64_json_missing_brackets(self):
|
|
||||||
"""Test get_dict_from_b64_json with missing brackets"""
|
|
||||||
config = ConfigLoader()
|
|
||||||
test_value = ' "foo": "bar" '.encode("utf-8")
|
|
||||||
b64_value = base64.b64encode(test_value)
|
|
||||||
config.set("foo", b64_value)
|
|
||||||
self.assertEqual(config.get_dict_from_b64_json("foo"), {"foo": "bar"})
|
|
||||||
|
|
||||||
def test_get_dict_from_b64_json_invalid(self):
|
|
||||||
"""Test get_dict_from_b64_json with invalid value"""
|
|
||||||
config = ConfigLoader()
|
|
||||||
config.set("foo", "bar")
|
|
||||||
self.assertEqual(config.get_dict_from_b64_json("foo"), {})
|
|
||||||
|
|
||||||
def test_attr_json_encoder(self):
|
|
||||||
"""Test AttrEncoder"""
|
|
||||||
test_attr = Attr("foo", Attr.Source.ENV, "AUTHENTIK_REDIS__USERNAME")
|
|
||||||
json_attr = dumps(test_attr, indent=4, cls=AttrEncoder)
|
|
||||||
self.assertEqual(json_attr, '"foo"')
|
|
||||||
|
|
||||||
def test_attr_json_encoder_no_attr(self):
|
|
||||||
"""Test AttrEncoder if no Attr is passed"""
|
|
||||||
|
|
||||||
class Test:
|
|
||||||
"""Non Attr class"""
|
|
||||||
|
|
||||||
with self.assertRaises(TypeError):
|
|
||||||
test_obj = Test()
|
|
||||||
dumps(test_obj, indent=4, cls=AttrEncoder)
|
|
||||||
|
|
||||||
@mock.patch.dict(environ, check_deprecations_env_vars)
|
|
||||||
def test_check_deprecations(self):
|
|
||||||
"""Test config key re-write for deprecated env vars"""
|
|
||||||
config = ConfigLoader()
|
|
||||||
config.update_from_env()
|
|
||||||
config.check_deprecations()
|
|
||||||
self.assertEqual(config.get("redis.broker_url", UNSET), UNSET)
|
|
||||||
self.assertEqual(config.get("redis.broker_transport_options", UNSET), UNSET)
|
|
||||||
self.assertEqual(config.get("redis.cache_timeout", UNSET), UNSET)
|
|
||||||
self.assertEqual(config.get("redis.cache_timeout_flows", UNSET), UNSET)
|
|
||||||
self.assertEqual(config.get("redis.cache_timeout_policies", UNSET), UNSET)
|
|
||||||
self.assertEqual(config.get("redis.cache_timeout_reputation", UNSET), UNSET)
|
|
||||||
self.assertEqual(config.get("broker.url"), "redis://myredis:8327/43")
|
|
||||||
self.assertEqual(config.get("broker.transport_options"), "bWFzdGVybmFtZT1teW1hc3Rlcg==")
|
|
||||||
self.assertEqual(config.get("cache.timeout"), "124s")
|
|
||||||
self.assertEqual(config.get("cache.timeout_flows"), "32m")
|
|
||||||
self.assertEqual(config.get("cache.timeout_policies"), "3920ns")
|
|
||||||
self.assertEqual(config.get("cache.timeout_reputation"), "298382us")
|
|
||||||
|
|
|
@ -3,8 +3,8 @@ from django.test import RequestFactory, TestCase
|
||||||
|
|
||||||
from authentik.core.models import Token, TokenIntents, UserTypes
|
from authentik.core.models import Token, TokenIntents, UserTypes
|
||||||
from authentik.core.tests.utils import create_test_admin_user
|
from authentik.core.tests.utils import create_test_admin_user
|
||||||
|
from authentik.lib.utils.http import OUTPOST_REMOTE_IP_HEADER, OUTPOST_TOKEN_HEADER, get_client_ip
|
||||||
from authentik.lib.views import bad_request_message
|
from authentik.lib.views import bad_request_message
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
|
|
||||||
|
|
||||||
class TestHTTP(TestCase):
|
class TestHTTP(TestCase):
|
||||||
|
@ -22,12 +22,12 @@ class TestHTTP(TestCase):
|
||||||
def test_normal(self):
|
def test_normal(self):
|
||||||
"""Test normal request"""
|
"""Test normal request"""
|
||||||
request = self.factory.get("/")
|
request = self.factory.get("/")
|
||||||
self.assertEqual(ClientIPMiddleware.get_client_ip(request), "127.0.0.1")
|
self.assertEqual(get_client_ip(request), "127.0.0.1")
|
||||||
|
|
||||||
def test_forward_for(self):
|
def test_forward_for(self):
|
||||||
"""Test x-forwarded-for request"""
|
"""Test x-forwarded-for request"""
|
||||||
request = self.factory.get("/", HTTP_X_FORWARDED_FOR="127.0.0.2")
|
request = self.factory.get("/", HTTP_X_FORWARDED_FOR="127.0.0.2")
|
||||||
self.assertEqual(ClientIPMiddleware.get_client_ip(request), "127.0.0.2")
|
self.assertEqual(get_client_ip(request), "127.0.0.2")
|
||||||
|
|
||||||
def test_fake_outpost(self):
|
def test_fake_outpost(self):
|
||||||
"""Test faked IP which is overridden by an outpost"""
|
"""Test faked IP which is overridden by an outpost"""
|
||||||
|
@ -38,28 +38,28 @@ class TestHTTP(TestCase):
|
||||||
request = self.factory.get(
|
request = self.factory.get(
|
||||||
"/",
|
"/",
|
||||||
**{
|
**{
|
||||||
ClientIPMiddleware.outpost_remote_ip_header: "1.2.3.4",
|
OUTPOST_REMOTE_IP_HEADER: "1.2.3.4",
|
||||||
ClientIPMiddleware.outpost_token_header: "abc",
|
OUTPOST_TOKEN_HEADER: "abc",
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
self.assertEqual(ClientIPMiddleware.get_client_ip(request), "127.0.0.1")
|
self.assertEqual(get_client_ip(request), "127.0.0.1")
|
||||||
# Invalid, user doesn't have permissions
|
# Invalid, user doesn't have permissions
|
||||||
request = self.factory.get(
|
request = self.factory.get(
|
||||||
"/",
|
"/",
|
||||||
**{
|
**{
|
||||||
ClientIPMiddleware.outpost_remote_ip_header: "1.2.3.4",
|
OUTPOST_REMOTE_IP_HEADER: "1.2.3.4",
|
||||||
ClientIPMiddleware.outpost_token_header: token.key,
|
OUTPOST_TOKEN_HEADER: token.key,
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
self.assertEqual(ClientIPMiddleware.get_client_ip(request), "127.0.0.1")
|
self.assertEqual(get_client_ip(request), "127.0.0.1")
|
||||||
# Valid
|
# Valid
|
||||||
self.user.type = UserTypes.INTERNAL_SERVICE_ACCOUNT
|
self.user.type = UserTypes.INTERNAL_SERVICE_ACCOUNT
|
||||||
self.user.save()
|
self.user.save()
|
||||||
request = self.factory.get(
|
request = self.factory.get(
|
||||||
"/",
|
"/",
|
||||||
**{
|
**{
|
||||||
ClientIPMiddleware.outpost_remote_ip_header: "1.2.3.4",
|
OUTPOST_REMOTE_IP_HEADER: "1.2.3.4",
|
||||||
ClientIPMiddleware.outpost_token_header: token.key,
|
OUTPOST_TOKEN_HEADER: token.key,
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
self.assertEqual(ClientIPMiddleware.get_client_ip(request), "1.2.3.4")
|
self.assertEqual(get_client_ip(request), "1.2.3.4")
|
||||||
|
|
|
@ -1,39 +1,89 @@
|
||||||
"""http helpers"""
|
"""http helpers"""
|
||||||
from uuid import uuid4
|
from typing import Any, Optional
|
||||||
|
|
||||||
from django.conf import settings
|
from django.http import HttpRequest
|
||||||
from requests.sessions import PreparedRequest, Session
|
from requests.sessions import Session
|
||||||
|
from sentry_sdk.hub import Hub
|
||||||
from structlog.stdlib import get_logger
|
from structlog.stdlib import get_logger
|
||||||
|
|
||||||
from authentik import get_full_version
|
from authentik import get_full_version
|
||||||
|
|
||||||
|
OUTPOST_REMOTE_IP_HEADER = "HTTP_X_AUTHENTIK_REMOTE_IP"
|
||||||
|
OUTPOST_TOKEN_HEADER = "HTTP_X_AUTHENTIK_OUTPOST_TOKEN" # nosec
|
||||||
|
DEFAULT_IP = "255.255.255.255"
|
||||||
LOGGER = get_logger()
|
LOGGER = get_logger()
|
||||||
|
|
||||||
|
|
||||||
|
def _get_client_ip_from_meta(meta: dict[str, Any]) -> str:
|
||||||
|
"""Attempt to get the client's IP by checking common HTTP Headers.
|
||||||
|
Returns none if no IP Could be found
|
||||||
|
|
||||||
|
No additional validation is done here as requests are expected to only arrive here
|
||||||
|
via the go proxy, which deals with validating these headers for us"""
|
||||||
|
headers = (
|
||||||
|
"HTTP_X_FORWARDED_FOR",
|
||||||
|
"REMOTE_ADDR",
|
||||||
|
)
|
||||||
|
for _header in headers:
|
||||||
|
if _header in meta:
|
||||||
|
ips: list[str] = meta.get(_header).split(",")
|
||||||
|
return ips[0].strip()
|
||||||
|
return DEFAULT_IP
|
||||||
|
|
||||||
|
|
||||||
|
def _get_outpost_override_ip(request: HttpRequest) -> Optional[str]:
|
||||||
|
"""Get the actual remote IP when set by an outpost. Only
|
||||||
|
allowed when the request is authenticated, by an outpost internal service account"""
|
||||||
|
from authentik.core.models import Token, TokenIntents, UserTypes
|
||||||
|
|
||||||
|
if OUTPOST_REMOTE_IP_HEADER not in request.META or OUTPOST_TOKEN_HEADER not in request.META:
|
||||||
|
return None
|
||||||
|
fake_ip = request.META[OUTPOST_REMOTE_IP_HEADER]
|
||||||
|
token = (
|
||||||
|
Token.filter_not_expired(
|
||||||
|
key=request.META.get(OUTPOST_TOKEN_HEADER), intent=TokenIntents.INTENT_API
|
||||||
|
)
|
||||||
|
.select_related("user")
|
||||||
|
.first()
|
||||||
|
)
|
||||||
|
if not token:
|
||||||
|
LOGGER.warning("Attempted remote-ip override without token", fake_ip=fake_ip)
|
||||||
|
return None
|
||||||
|
user = token.user
|
||||||
|
if user.type != UserTypes.INTERNAL_SERVICE_ACCOUNT:
|
||||||
|
LOGGER.warning(
|
||||||
|
"Remote-IP override: user doesn't have permission",
|
||||||
|
user=user,
|
||||||
|
fake_ip=fake_ip,
|
||||||
|
)
|
||||||
|
return None
|
||||||
|
# Update sentry scope to include correct IP
|
||||||
|
user = Hub.current.scope._user
|
||||||
|
if not user:
|
||||||
|
user = {}
|
||||||
|
user["ip_address"] = fake_ip
|
||||||
|
Hub.current.scope.set_user(user)
|
||||||
|
return fake_ip
|
||||||
|
|
||||||
|
|
||||||
|
def get_client_ip(request: Optional[HttpRequest]) -> str:
|
||||||
|
"""Attempt to get the client's IP by checking common HTTP Headers.
|
||||||
|
Returns none if no IP Could be found"""
|
||||||
|
if not request:
|
||||||
|
return DEFAULT_IP
|
||||||
|
override = _get_outpost_override_ip(request)
|
||||||
|
if override:
|
||||||
|
return override
|
||||||
|
return _get_client_ip_from_meta(request.META)
|
||||||
|
|
||||||
|
|
||||||
def authentik_user_agent() -> str:
|
def authentik_user_agent() -> str:
|
||||||
"""Get a common user agent"""
|
"""Get a common user agent"""
|
||||||
return f"authentik@{get_full_version()}"
|
return f"authentik@{get_full_version()}"
|
||||||
|
|
||||||
|
|
||||||
class DebugSession(Session):
|
|
||||||
"""requests session which logs http requests and responses"""
|
|
||||||
|
|
||||||
def send(self, req: PreparedRequest, *args, **kwargs):
|
|
||||||
request_id = str(uuid4())
|
|
||||||
LOGGER.debug("HTTP request sent", uid=request_id, path=req.path_url, headers=req.headers)
|
|
||||||
resp = super().send(req, *args, **kwargs)
|
|
||||||
LOGGER.debug(
|
|
||||||
"HTTP response received",
|
|
||||||
uid=request_id,
|
|
||||||
status=resp.status_code,
|
|
||||||
body=resp.text,
|
|
||||||
headers=resp.headers,
|
|
||||||
)
|
|
||||||
return resp
|
|
||||||
|
|
||||||
|
|
||||||
def get_http_session() -> Session:
|
def get_http_session() -> Session:
|
||||||
"""Get a requests session with common headers"""
|
"""Get a requests session with common headers"""
|
||||||
session = DebugSession() if settings.DEBUG else Session()
|
session = Session()
|
||||||
session.headers["User-Agent"] = authentik_user_agent()
|
session.headers["User-Agent"] = authentik_user_agent()
|
||||||
return session
|
return session
|
||||||
|
|
|
@ -9,15 +9,14 @@ from rest_framework.fields import BooleanField, CharField, DateTimeField
|
||||||
from rest_framework.relations import PrimaryKeyRelatedField
|
from rest_framework.relations import PrimaryKeyRelatedField
|
||||||
from rest_framework.request import Request
|
from rest_framework.request import Request
|
||||||
from rest_framework.response import Response
|
from rest_framework.response import Response
|
||||||
from rest_framework.serializers import ModelSerializer, ValidationError
|
from rest_framework.serializers import JSONField, ModelSerializer, ValidationError
|
||||||
from rest_framework.viewsets import ModelViewSet
|
from rest_framework.viewsets import ModelViewSet
|
||||||
|
|
||||||
from authentik import get_build_hash
|
from authentik import get_build_hash
|
||||||
from authentik.core.api.providers import ProviderSerializer
|
from authentik.core.api.providers import ProviderSerializer
|
||||||
from authentik.core.api.used_by import UsedByMixin
|
from authentik.core.api.used_by import UsedByMixin
|
||||||
from authentik.core.api.utils import JSONDictField, PassiveSerializer
|
from authentik.core.api.utils import PassiveSerializer, is_dict
|
||||||
from authentik.core.models import Provider
|
from authentik.core.models import Provider
|
||||||
from authentik.enterprise.providers.rac.models import RACProvider
|
|
||||||
from authentik.outposts.api.service_connections import ServiceConnectionSerializer
|
from authentik.outposts.api.service_connections import ServiceConnectionSerializer
|
||||||
from authentik.outposts.apps import MANAGED_OUTPOST, MANAGED_OUTPOST_NAME
|
from authentik.outposts.apps import MANAGED_OUTPOST, MANAGED_OUTPOST_NAME
|
||||||
from authentik.outposts.models import (
|
from authentik.outposts.models import (
|
||||||
|
@ -35,7 +34,7 @@ from authentik.providers.radius.models import RadiusProvider
|
||||||
class OutpostSerializer(ModelSerializer):
|
class OutpostSerializer(ModelSerializer):
|
||||||
"""Outpost Serializer"""
|
"""Outpost Serializer"""
|
||||||
|
|
||||||
config = JSONDictField(source="_config")
|
config = JSONField(validators=[is_dict], source="_config")
|
||||||
# Need to set allow_empty=True for the embedded outpost with no providers
|
# Need to set allow_empty=True for the embedded outpost with no providers
|
||||||
# is checked for other providers in the API Viewset
|
# is checked for other providers in the API Viewset
|
||||||
providers = PrimaryKeyRelatedField(
|
providers = PrimaryKeyRelatedField(
|
||||||
|
@ -64,7 +63,6 @@ class OutpostSerializer(ModelSerializer):
|
||||||
OutpostType.LDAP: LDAPProvider,
|
OutpostType.LDAP: LDAPProvider,
|
||||||
OutpostType.PROXY: ProxyProvider,
|
OutpostType.PROXY: ProxyProvider,
|
||||||
OutpostType.RADIUS: RadiusProvider,
|
OutpostType.RADIUS: RadiusProvider,
|
||||||
OutpostType.RAC: RACProvider,
|
|
||||||
None: Provider,
|
None: Provider,
|
||||||
}
|
}
|
||||||
for provider in providers:
|
for provider in providers:
|
||||||
|
@ -107,7 +105,7 @@ class OutpostSerializer(ModelSerializer):
|
||||||
class OutpostDefaultConfigSerializer(PassiveSerializer):
|
class OutpostDefaultConfigSerializer(PassiveSerializer):
|
||||||
"""Global default outpost config"""
|
"""Global default outpost config"""
|
||||||
|
|
||||||
config = JSONDictField(read_only=True)
|
config = JSONField(read_only=True)
|
||||||
|
|
||||||
|
|
||||||
class OutpostHealthSerializer(PassiveSerializer):
|
class OutpostHealthSerializer(PassiveSerializer):
|
||||||
|
|
|
@ -6,18 +6,16 @@ from typing import Any, Optional
|
||||||
|
|
||||||
from asgiref.sync import async_to_sync
|
from asgiref.sync import async_to_sync
|
||||||
from channels.exceptions import DenyConnection
|
from channels.exceptions import DenyConnection
|
||||||
from channels.generic.websocket import JsonWebsocketConsumer
|
|
||||||
from dacite.core import from_dict
|
from dacite.core import from_dict
|
||||||
from dacite.data import Data
|
from dacite.data import Data
|
||||||
from django.http.request import QueryDict
|
|
||||||
from guardian.shortcuts import get_objects_for_user
|
from guardian.shortcuts import get_objects_for_user
|
||||||
from structlog.stdlib import BoundLogger, get_logger
|
from structlog.stdlib import BoundLogger, get_logger
|
||||||
|
|
||||||
|
from authentik.core.channels import AuthJsonConsumer
|
||||||
from authentik.outposts.apps import GAUGE_OUTPOSTS_CONNECTED, GAUGE_OUTPOSTS_LAST_UPDATE
|
from authentik.outposts.apps import GAUGE_OUTPOSTS_CONNECTED, GAUGE_OUTPOSTS_LAST_UPDATE
|
||||||
from authentik.outposts.models import OUTPOST_HELLO_INTERVAL, Outpost, OutpostState
|
from authentik.outposts.models import OUTPOST_HELLO_INTERVAL, Outpost, OutpostState
|
||||||
|
|
||||||
OUTPOST_GROUP = "group_outpost_%(outpost_pk)s"
|
OUTPOST_GROUP = "group_outpost_%(outpost_pk)s"
|
||||||
OUTPOST_GROUP_INSTANCE = "group_outpost_%(outpost_pk)s_%(instance)s"
|
|
||||||
|
|
||||||
|
|
||||||
class WebsocketMessageInstruction(IntEnum):
|
class WebsocketMessageInstruction(IntEnum):
|
||||||
|
@ -44,23 +42,25 @@ class WebsocketMessage:
|
||||||
args: dict[str, Any] = field(default_factory=dict)
|
args: dict[str, Any] = field(default_factory=dict)
|
||||||
|
|
||||||
|
|
||||||
class OutpostConsumer(JsonWebsocketConsumer):
|
class OutpostConsumer(AuthJsonConsumer):
|
||||||
"""Handler for Outposts that connect over websockets for health checks and live updates"""
|
"""Handler for Outposts that connect over websockets for health checks and live updates"""
|
||||||
|
|
||||||
outpost: Optional[Outpost] = None
|
outpost: Optional[Outpost] = None
|
||||||
logger: BoundLogger
|
logger: BoundLogger
|
||||||
|
|
||||||
instance_uid: Optional[str] = None
|
last_uid: Optional[str] = None
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
super().__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
self.logger = get_logger()
|
self.logger = get_logger()
|
||||||
|
|
||||||
def connect(self):
|
def connect(self):
|
||||||
|
super().connect()
|
||||||
uuid = self.scope["url_route"]["kwargs"]["pk"]
|
uuid = self.scope["url_route"]["kwargs"]["pk"]
|
||||||
user = self.scope["user"]
|
|
||||||
outpost = (
|
outpost = (
|
||||||
get_objects_for_user(user, "authentik_outposts.view_outpost").filter(pk=uuid).first()
|
get_objects_for_user(self.user, "authentik_outposts.view_outpost")
|
||||||
|
.filter(pk=uuid)
|
||||||
|
.first()
|
||||||
)
|
)
|
||||||
if not outpost:
|
if not outpost:
|
||||||
raise DenyConnection()
|
raise DenyConnection()
|
||||||
|
@ -71,19 +71,13 @@ class OutpostConsumer(JsonWebsocketConsumer):
|
||||||
self.logger.warning("runtime error during accept", exc=exc)
|
self.logger.warning("runtime error during accept", exc=exc)
|
||||||
raise DenyConnection()
|
raise DenyConnection()
|
||||||
self.outpost = outpost
|
self.outpost = outpost
|
||||||
query = QueryDict(self.scope["query_string"].decode())
|
self.last_uid = self.channel_name
|
||||||
self.instance_uid = query.get("instance_uuid", self.channel_name)
|
|
||||||
async_to_sync(self.channel_layer.group_add)(
|
async_to_sync(self.channel_layer.group_add)(
|
||||||
OUTPOST_GROUP % {"outpost_pk": str(self.outpost.pk)}, self.channel_name
|
OUTPOST_GROUP % {"outpost_pk": str(self.outpost.pk)}, self.channel_name
|
||||||
)
|
)
|
||||||
async_to_sync(self.channel_layer.group_add)(
|
|
||||||
OUTPOST_GROUP_INSTANCE
|
|
||||||
% {"outpost_pk": str(self.outpost.pk), "instance": self.instance_uid},
|
|
||||||
self.channel_name,
|
|
||||||
)
|
|
||||||
GAUGE_OUTPOSTS_CONNECTED.labels(
|
GAUGE_OUTPOSTS_CONNECTED.labels(
|
||||||
outpost=self.outpost.name,
|
outpost=self.outpost.name,
|
||||||
uid=self.instance_uid,
|
uid=self.last_uid,
|
||||||
expected=self.outpost.config.kubernetes_replicas,
|
expected=self.outpost.config.kubernetes_replicas,
|
||||||
).inc()
|
).inc()
|
||||||
|
|
||||||
|
@ -92,37 +86,34 @@ class OutpostConsumer(JsonWebsocketConsumer):
|
||||||
async_to_sync(self.channel_layer.group_discard)(
|
async_to_sync(self.channel_layer.group_discard)(
|
||||||
OUTPOST_GROUP % {"outpost_pk": str(self.outpost.pk)}, self.channel_name
|
OUTPOST_GROUP % {"outpost_pk": str(self.outpost.pk)}, self.channel_name
|
||||||
)
|
)
|
||||||
if self.instance_uid:
|
if self.outpost and self.last_uid:
|
||||||
async_to_sync(self.channel_layer.group_discard)(
|
|
||||||
OUTPOST_GROUP_INSTANCE
|
|
||||||
% {"outpost_pk": str(self.outpost.pk), "instance": self.instance_uid},
|
|
||||||
self.channel_name,
|
|
||||||
)
|
|
||||||
if self.outpost and self.instance_uid:
|
|
||||||
GAUGE_OUTPOSTS_CONNECTED.labels(
|
GAUGE_OUTPOSTS_CONNECTED.labels(
|
||||||
outpost=self.outpost.name,
|
outpost=self.outpost.name,
|
||||||
uid=self.instance_uid,
|
uid=self.last_uid,
|
||||||
expected=self.outpost.config.kubernetes_replicas,
|
expected=self.outpost.config.kubernetes_replicas,
|
||||||
).dec()
|
).dec()
|
||||||
|
|
||||||
def receive_json(self, content: Data, **kwargs):
|
def receive_json(self, content: Data):
|
||||||
msg = from_dict(WebsocketMessage, content)
|
msg = from_dict(WebsocketMessage, content)
|
||||||
|
uid = msg.args.get("uuid", self.channel_name)
|
||||||
|
self.last_uid = uid
|
||||||
|
|
||||||
if not self.outpost:
|
if not self.outpost:
|
||||||
raise DenyConnection()
|
raise DenyConnection()
|
||||||
|
|
||||||
state = OutpostState.for_instance_uid(self.outpost, self.instance_uid)
|
state = OutpostState.for_instance_uid(self.outpost, uid)
|
||||||
state.last_seen = datetime.now()
|
state.last_seen = datetime.now()
|
||||||
state.hostname = msg.args.pop("hostname", "")
|
state.hostname = msg.args.pop("hostname", "")
|
||||||
|
|
||||||
if msg.instruction == WebsocketMessageInstruction.HELLO:
|
if msg.instruction == WebsocketMessageInstruction.HELLO:
|
||||||
state.version = msg.args.pop("version", None)
|
state.version = msg.args.pop("version", None)
|
||||||
state.build_hash = msg.args.pop("buildHash", "")
|
state.build_hash = msg.args.pop("buildHash", "")
|
||||||
state.args.update(msg.args)
|
state.args = msg.args
|
||||||
elif msg.instruction == WebsocketMessageInstruction.ACK:
|
elif msg.instruction == WebsocketMessageInstruction.ACK:
|
||||||
return
|
return
|
||||||
GAUGE_OUTPOSTS_LAST_UPDATE.labels(
|
GAUGE_OUTPOSTS_LAST_UPDATE.labels(
|
||||||
outpost=self.outpost.name,
|
outpost=self.outpost.name,
|
||||||
uid=self.instance_uid or "",
|
uid=self.last_uid or "",
|
||||||
version=state.version or "",
|
version=state.version or "",
|
||||||
).set_to_current_time()
|
).set_to_current_time()
|
||||||
state.save(timeout=OUTPOST_HELLO_INTERVAL * 1.5)
|
state.save(timeout=OUTPOST_HELLO_INTERVAL * 1.5)
|
||||||
|
|
|
@ -1,6 +1,5 @@
|
||||||
"""k8s utils"""
|
"""k8s utils"""
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Optional
|
|
||||||
|
|
||||||
from kubernetes.client.models.v1_container_port import V1ContainerPort
|
from kubernetes.client.models.v1_container_port import V1ContainerPort
|
||||||
from kubernetes.client.models.v1_service_port import V1ServicePort
|
from kubernetes.client.models.v1_service_port import V1ServicePort
|
||||||
|
@ -38,12 +37,9 @@ def compare_port(
|
||||||
|
|
||||||
|
|
||||||
def compare_ports(
|
def compare_ports(
|
||||||
current: Optional[list[V1ServicePort | V1ContainerPort]],
|
current: list[V1ServicePort | V1ContainerPort], reference: list[V1ServicePort | V1ContainerPort]
|
||||||
reference: Optional[list[V1ServicePort | V1ContainerPort]],
|
|
||||||
):
|
):
|
||||||
"""Compare ports of a list"""
|
"""Compare ports of a list"""
|
||||||
if not current or not reference:
|
|
||||||
raise NeedsRecreate()
|
|
||||||
if len(current) != len(reference):
|
if len(current) != len(reference):
|
||||||
raise NeedsRecreate()
|
raise NeedsRecreate()
|
||||||
for port in reference:
|
for port in reference:
|
||||||
|
|
|
@ -81,10 +81,7 @@ class KubernetesController(BaseController):
|
||||||
def up(self):
|
def up(self):
|
||||||
try:
|
try:
|
||||||
for reconcile_key in self.reconcile_order:
|
for reconcile_key in self.reconcile_order:
|
||||||
reconciler_cls = self.reconcilers.get(reconcile_key)
|
reconciler = self.reconcilers[reconcile_key](self)
|
||||||
if not reconciler_cls:
|
|
||||||
continue
|
|
||||||
reconciler = reconciler_cls(self)
|
|
||||||
reconciler.up()
|
reconciler.up()
|
||||||
|
|
||||||
except (OpenApiException, HTTPError, ServiceConnectionInvalid) as exc:
|
except (OpenApiException, HTTPError, ServiceConnectionInvalid) as exc:
|
||||||
|
@ -98,10 +95,7 @@ class KubernetesController(BaseController):
|
||||||
all_logs += [f"{reconcile_key.title()}: Disabled"]
|
all_logs += [f"{reconcile_key.title()}: Disabled"]
|
||||||
continue
|
continue
|
||||||
with capture_logs() as logs:
|
with capture_logs() as logs:
|
||||||
reconciler_cls = self.reconcilers.get(reconcile_key)
|
reconciler = self.reconcilers[reconcile_key](self)
|
||||||
if not reconciler_cls:
|
|
||||||
continue
|
|
||||||
reconciler = reconciler_cls(self)
|
|
||||||
reconciler.up()
|
reconciler.up()
|
||||||
all_logs += [f"{reconcile_key.title()}: {x['event']}" for x in logs]
|
all_logs += [f"{reconcile_key.title()}: {x['event']}" for x in logs]
|
||||||
return all_logs
|
return all_logs
|
||||||
|
@ -111,10 +105,7 @@ class KubernetesController(BaseController):
|
||||||
def down(self):
|
def down(self):
|
||||||
try:
|
try:
|
||||||
for reconcile_key in self.reconcile_order:
|
for reconcile_key in self.reconcile_order:
|
||||||
reconciler_cls = self.reconcilers.get(reconcile_key)
|
reconciler = self.reconcilers[reconcile_key](self)
|
||||||
if not reconciler_cls:
|
|
||||||
continue
|
|
||||||
reconciler = reconciler_cls(self)
|
|
||||||
self.logger.debug("Tearing down object", name=reconcile_key)
|
self.logger.debug("Tearing down object", name=reconcile_key)
|
||||||
reconciler.down()
|
reconciler.down()
|
||||||
|
|
||||||
|
@ -129,10 +120,7 @@ class KubernetesController(BaseController):
|
||||||
all_logs += [f"{reconcile_key.title()}: Disabled"]
|
all_logs += [f"{reconcile_key.title()}: Disabled"]
|
||||||
continue
|
continue
|
||||||
with capture_logs() as logs:
|
with capture_logs() as logs:
|
||||||
reconciler_cls = self.reconcilers.get(reconcile_key)
|
reconciler = self.reconcilers[reconcile_key](self)
|
||||||
if not reconciler_cls:
|
|
||||||
continue
|
|
||||||
reconciler = reconciler_cls(self)
|
|
||||||
reconciler.down()
|
reconciler.down()
|
||||||
all_logs += [f"{reconcile_key.title()}: {x['event']}" for x in logs]
|
all_logs += [f"{reconcile_key.title()}: {x['event']}" for x in logs]
|
||||||
return all_logs
|
return all_logs
|
||||||
|
@ -142,10 +130,7 @@ class KubernetesController(BaseController):
|
||||||
def get_static_deployment(self) -> str:
|
def get_static_deployment(self) -> str:
|
||||||
documents = []
|
documents = []
|
||||||
for reconcile_key in self.reconcile_order:
|
for reconcile_key in self.reconcile_order:
|
||||||
reconciler_cls = self.reconcilers.get(reconcile_key)
|
reconciler = self.reconcilers[reconcile_key](self)
|
||||||
if not reconciler_cls:
|
|
||||||
continue
|
|
||||||
reconciler = reconciler_cls(self)
|
|
||||||
if reconciler.noop:
|
if reconciler.noop:
|
||||||
continue
|
continue
|
||||||
documents.append(reconciler.get_reference_object().to_dict())
|
documents.append(reconciler.get_reference_object().to_dict())
|
||||||
|
|
|
@ -1,25 +0,0 @@
|
||||||
# Generated by Django 4.2.6 on 2023-10-14 19:23
|
|
||||||
|
|
||||||
from django.db import migrations, models
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
|
||||||
dependencies = [
|
|
||||||
("authentik_outposts", "0020_alter_outpost_type"),
|
|
||||||
]
|
|
||||||
|
|
||||||
operations = [
|
|
||||||
migrations.AlterField(
|
|
||||||
model_name="outpost",
|
|
||||||
name="type",
|
|
||||||
field=models.TextField(
|
|
||||||
choices=[
|
|
||||||
("proxy", "Proxy"),
|
|
||||||
("ldap", "Ldap"),
|
|
||||||
("radius", "Radius"),
|
|
||||||
("rac", "Rac"),
|
|
||||||
],
|
|
||||||
default="proxy",
|
|
||||||
),
|
|
||||||
),
|
|
||||||
]
|
|
|
@ -90,12 +90,11 @@ class OutpostModel(Model):
|
||||||
|
|
||||||
|
|
||||||
class OutpostType(models.TextChoices):
|
class OutpostType(models.TextChoices):
|
||||||
"""Outpost types"""
|
"""Outpost types, currently only the reverse proxy is available"""
|
||||||
|
|
||||||
PROXY = "proxy"
|
PROXY = "proxy"
|
||||||
LDAP = "ldap"
|
LDAP = "ldap"
|
||||||
RADIUS = "radius"
|
RADIUS = "radius"
|
||||||
RAC = "rac"
|
|
||||||
|
|
||||||
|
|
||||||
def default_outpost_config(host: Optional[str] = None):
|
def default_outpost_config(host: Optional[str] = None):
|
||||||
|
@ -460,7 +459,7 @@ class OutpostState:
|
||||||
def for_instance_uid(outpost: Outpost, uid: str) -> "OutpostState":
|
def for_instance_uid(outpost: Outpost, uid: str) -> "OutpostState":
|
||||||
"""Get state for a single instance"""
|
"""Get state for a single instance"""
|
||||||
key = f"{outpost.state_cache_prefix}/{uid}"
|
key = f"{outpost.state_cache_prefix}/{uid}"
|
||||||
default_data = {"uid": uid}
|
default_data = {"uid": uid, "channel_ids": []}
|
||||||
data = cache.get(key, default_data)
|
data = cache.get(key, default_data)
|
||||||
if isinstance(data, str):
|
if isinstance(data, str):
|
||||||
cache.delete(key)
|
cache.delete(key)
|
||||||
|
|
|
@ -17,8 +17,6 @@ from kubernetes.config.kube_config import KUBE_CONFIG_DEFAULT_LOCATION
|
||||||
from structlog.stdlib import get_logger
|
from structlog.stdlib import get_logger
|
||||||
from yaml import safe_load
|
from yaml import safe_load
|
||||||
|
|
||||||
from authentik.enterprise.providers.rac.controllers.docker import RACDockerController
|
|
||||||
from authentik.enterprise.providers.rac.controllers.kubernetes import RACKubernetesController
|
|
||||||
from authentik.events.monitored_tasks import (
|
from authentik.events.monitored_tasks import (
|
||||||
MonitoredTask,
|
MonitoredTask,
|
||||||
TaskResult,
|
TaskResult,
|
||||||
|
@ -73,11 +71,6 @@ def controller_for_outpost(outpost: Outpost) -> Optional[type[BaseController]]:
|
||||||
return RadiusDockerController
|
return RadiusDockerController
|
||||||
if isinstance(service_connection, KubernetesServiceConnection):
|
if isinstance(service_connection, KubernetesServiceConnection):
|
||||||
return RadiusKubernetesController
|
return RadiusKubernetesController
|
||||||
if outpost.type == OutpostType.RAC:
|
|
||||||
if isinstance(service_connection, DockerServiceConnection):
|
|
||||||
return RACDockerController
|
|
||||||
if isinstance(service_connection, KubernetesServiceConnection):
|
|
||||||
return RACKubernetesController
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,6 @@
|
||||||
"""Websocket tests"""
|
"""Websocket tests"""
|
||||||
from dataclasses import asdict
|
from dataclasses import asdict
|
||||||
|
|
||||||
from channels.exceptions import DenyConnection
|
|
||||||
from channels.routing import URLRouter
|
from channels.routing import URLRouter
|
||||||
from channels.testing import WebsocketCommunicator
|
from channels.testing import WebsocketCommunicator
|
||||||
from django.test import TransactionTestCase
|
from django.test import TransactionTestCase
|
||||||
|
@ -36,9 +35,8 @@ class TestOutpostWS(TransactionTestCase):
|
||||||
communicator = WebsocketCommunicator(
|
communicator = WebsocketCommunicator(
|
||||||
URLRouter(websocket.websocket_urlpatterns), f"/ws/outpost/{self.outpost.pk}/"
|
URLRouter(websocket.websocket_urlpatterns), f"/ws/outpost/{self.outpost.pk}/"
|
||||||
)
|
)
|
||||||
with self.assertRaises(DenyConnection):
|
connected, _ = await communicator.connect()
|
||||||
connected, _ = await communicator.connect()
|
self.assertFalse(connected)
|
||||||
self.assertFalse(connected)
|
|
||||||
|
|
||||||
async def test_auth_valid(self):
|
async def test_auth_valid(self):
|
||||||
"""Test auth with token"""
|
"""Test auth with token"""
|
||||||
|
|
|
@ -1,7 +1,6 @@
|
||||||
"""Outpost Websocket URLS"""
|
"""Outpost Websocket URLS"""
|
||||||
from django.urls import path
|
from django.urls import path
|
||||||
|
|
||||||
from authentik.core.channels import TokenOutpostMiddleware
|
|
||||||
from authentik.outposts.api.outposts import OutpostViewSet
|
from authentik.outposts.api.outposts import OutpostViewSet
|
||||||
from authentik.outposts.api.service_connections import (
|
from authentik.outposts.api.service_connections import (
|
||||||
DockerServiceConnectionViewSet,
|
DockerServiceConnectionViewSet,
|
||||||
|
@ -12,10 +11,7 @@ from authentik.outposts.consumer import OutpostConsumer
|
||||||
from authentik.root.middleware import ChannelsLoggingMiddleware
|
from authentik.root.middleware import ChannelsLoggingMiddleware
|
||||||
|
|
||||||
websocket_urlpatterns = [
|
websocket_urlpatterns = [
|
||||||
path(
|
path("ws/outpost/<uuid:pk>/", ChannelsLoggingMiddleware(OutpostConsumer.as_asgi())),
|
||||||
"ws/outpost/<uuid:pk>/",
|
|
||||||
ChannelsLoggingMiddleware(TokenOutpostMiddleware(OutpostConsumer.as_asgi())),
|
|
||||||
),
|
|
||||||
]
|
]
|
||||||
|
|
||||||
api_urlpatterns = [
|
api_urlpatterns = [
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
"""Serializer for policy execution"""
|
"""Serializer for policy execution"""
|
||||||
from rest_framework.fields import BooleanField, CharField, DictField, ListField
|
from rest_framework.fields import BooleanField, CharField, DictField, JSONField, ListField
|
||||||
from rest_framework.relations import PrimaryKeyRelatedField
|
from rest_framework.relations import PrimaryKeyRelatedField
|
||||||
|
|
||||||
from authentik.core.api.utils import JSONDictField, PassiveSerializer
|
from authentik.core.api.utils import PassiveSerializer, is_dict
|
||||||
from authentik.core.models import User
|
from authentik.core.models import User
|
||||||
|
|
||||||
|
|
||||||
|
@ -10,7 +10,7 @@ class PolicyTestSerializer(PassiveSerializer):
|
||||||
"""Test policy execution for a user with context"""
|
"""Test policy execution for a user with context"""
|
||||||
|
|
||||||
user = PrimaryKeyRelatedField(queryset=User.objects.all())
|
user = PrimaryKeyRelatedField(queryset=User.objects.all())
|
||||||
context = JSONDictField(required=False)
|
context = JSONField(required=False, validators=[is_dict])
|
||||||
|
|
||||||
|
|
||||||
class PolicyTestResultSerializer(PassiveSerializer):
|
class PolicyTestResultSerializer(PassiveSerializer):
|
||||||
|
|
|
@ -7,9 +7,9 @@ from structlog.stdlib import get_logger
|
||||||
|
|
||||||
from authentik.flows.planner import PLAN_CONTEXT_SSO
|
from authentik.flows.planner import PLAN_CONTEXT_SSO
|
||||||
from authentik.lib.expression.evaluator import BaseEvaluator
|
from authentik.lib.expression.evaluator import BaseEvaluator
|
||||||
|
from authentik.lib.utils.http import get_client_ip
|
||||||
from authentik.policies.exceptions import PolicyException
|
from authentik.policies.exceptions import PolicyException
|
||||||
from authentik.policies.types import PolicyRequest, PolicyResult
|
from authentik.policies.types import PolicyRequest, PolicyResult
|
||||||
from authentik.root.middleware import ClientIPMiddleware
|
|
||||||
|
|
||||||
LOGGER = get_logger()
|
LOGGER = get_logger()
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
|
@ -49,7 +49,7 @@ class PolicyEvaluator(BaseEvaluator):
|
||||||
"""Update context based on http request"""
|
"""Update context based on http request"""
|
||||||
# update website/docs/expressions/_objects.md
|
# update website/docs/expressions/_objects.md
|
||||||
# update website/docs/expressions/_functions.md
|
# update website/docs/expressions/_functions.md
|
||||||
self._context["ak_client_ip"] = ip_address(ClientIPMiddleware.get_client_ip(request))
|
self._context["ak_client_ip"] = ip_address(get_client_ip(request))
|
||||||
self._context["http_request"] = request
|
self._context["http_request"] = request
|
||||||
|
|
||||||
def handle_error(self, exc: Exception, expression_source: str):
|
def handle_error(self, exc: Exception, expression_source: str):
|
||||||
|
|
|
@ -20,7 +20,7 @@ from authentik.policies.types import CACHE_PREFIX, PolicyRequest, PolicyResult
|
||||||
LOGGER = get_logger()
|
LOGGER = get_logger()
|
||||||
|
|
||||||
FORK_CTX = get_context("fork")
|
FORK_CTX = get_context("fork")
|
||||||
CACHE_TIMEOUT = CONFIG.get_int("cache.timeout_policies")
|
CACHE_TIMEOUT = CONFIG.get_int("redis.cache_timeout_policies")
|
||||||
PROCESS_CLASS = FORK_CTX.Process
|
PROCESS_CLASS = FORK_CTX.Process
|
||||||
|
|
||||||
|
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show more
Reference in a new issue