mirror of
https://github.com/chatmail/relay.git
synced 2026-05-11 08:24:37 +00:00
Compare commits
2 Commits
j4n/docker
...
postfix-lo
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
0d301f9807 | ||
|
|
a5dffdf2e6 |
1
.github/ISSUE_TEMPLATE/bug_report.md
vendored
1
.github/ISSUE_TEMPLATE/bug_report.md
vendored
@@ -12,7 +12,6 @@ Please fill out as much of this form as you can (leaving out stuff that is not a
|
|||||||
|
|
||||||
- Server OS (Operating System) - preferably Debian 12:
|
- Server OS (Operating System) - preferably Debian 12:
|
||||||
- On which OS you run cmdeploy:
|
- On which OS you run cmdeploy:
|
||||||
- chatmail/relay version: `git rev-parse HEAD`
|
|
||||||
|
|
||||||
## Expected behavior
|
## Expected behavior
|
||||||
|
|
||||||
|
|||||||
4
.github/ISSUE_TEMPLATE/config.yml
vendored
4
.github/ISSUE_TEMPLATE/config.yml
vendored
@@ -1 +1,5 @@
|
|||||||
blank_issues_enabled: true
|
blank_issues_enabled: true
|
||||||
|
contact_links:
|
||||||
|
- name: Mutual Help Chat Group
|
||||||
|
url: https://i.delta.chat/#C2846EB4C1CB8DF84B1818F5E3A638FC3FBDC981&a=stalebot1%40nine.testrun.org&g=Chatmail%20Mutual%20Help&x=7sFF7Ik50pWv6J1z7RVC5527&i=d7s1HvOsk5UrSf9AoqRZggg4&s=XmX_9BAW6-g5Ao5E8PyaeKNB
|
||||||
|
about: If you have troubles setting up the relay server, feel free to ask here.
|
||||||
|
|||||||
54
.github/workflows/ci.yaml
vendored
54
.github/workflows/ci.yaml
vendored
@@ -1,35 +1,16 @@
|
|||||||
name: Run unit-tests and container-based deploy+test verification
|
name: CI
|
||||||
|
|
||||||
on:
|
on:
|
||||||
# Triggers when a PR is merged into main or a direct push occurs
|
|
||||||
push:
|
|
||||||
branches: [ "main" ]
|
|
||||||
|
|
||||||
# Triggers for any PR (and its subsequent commits) targeting the main branch
|
|
||||||
pull_request:
|
pull_request:
|
||||||
branches: [ "main" ]
|
push:
|
||||||
|
|
||||||
permissions: {}
|
|
||||||
|
|
||||||
# Newest push wins: Prevents multiple runs from clashing and wasting runner efforts
|
|
||||||
concurrency:
|
|
||||||
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
|
|
||||||
cancel-in-progress: true
|
|
||||||
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
tox:
|
tox:
|
||||||
name: isolated chatmaild tests
|
name: isolated chatmaild tests
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v6
|
- uses: actions/checkout@v4
|
||||||
# Checkout pull request HEAD commit instead of merge commit
|
|
||||||
# Otherwise `test_deployed_state` will be unhappy.
|
|
||||||
with:
|
|
||||||
ref: ${{ github.event.pull_request.head.sha }}
|
|
||||||
persist-credentials: false
|
|
||||||
- name: download filtermail
|
|
||||||
run: curl -L https://github.com/chatmail/filtermail/releases/download/v0.6.4/filtermail-x86_64 -o /usr/local/bin/filtermail && chmod +x /usr/local/bin/filtermail
|
|
||||||
- name: run chatmaild tests
|
- name: run chatmaild tests
|
||||||
working-directory: chatmaild
|
working-directory: chatmaild
|
||||||
run: pipx run tox
|
run: pipx run tox
|
||||||
@@ -38,10 +19,7 @@ jobs:
|
|||||||
name: deploy-chatmail tests
|
name: deploy-chatmail tests
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v6
|
- uses: actions/checkout@v4
|
||||||
with:
|
|
||||||
ref: ${{ github.event.pull_request.head.sha }}
|
|
||||||
persist-credentials: false
|
|
||||||
|
|
||||||
- name: initenv
|
- name: initenv
|
||||||
run: scripts/initenv.sh
|
run: scripts/initenv.sh
|
||||||
@@ -55,23 +33,5 @@ jobs:
|
|||||||
- name: run deploy-chatmail offline tests
|
- name: run deploy-chatmail offline tests
|
||||||
run: pytest --pyargs cmdeploy
|
run: pytest --pyargs cmdeploy
|
||||||
|
|
||||||
lxc-test:
|
# all other cmdeploy commands require a staging server
|
||||||
name: LXC deploy and test
|
# see https://github.com/deltachat/chatmail/issues/100
|
||||||
uses: chatmail/cmlxc/.github/workflows/lxc-test.yml@v0.10.0
|
|
||||||
with:
|
|
||||||
cmlxc_commands: |
|
|
||||||
cmlxc init
|
|
||||||
# single cmdeploy relay test
|
|
||||||
cmlxc -v deploy-cmdeploy --source ./repo cm0
|
|
||||||
cmlxc -v test-mini cm0
|
|
||||||
cmlxc -v test-cmdeploy cm0
|
|
||||||
|
|
||||||
# cross cmdeploy relay test
|
|
||||||
cmlxc -v deploy-cmdeploy --source ./repo --ipv4-only cm1
|
|
||||||
cmlxc -v test-cmdeploy cm0 cm1
|
|
||||||
|
|
||||||
# cross cmdeploy/madmail relay tests
|
|
||||||
cmlxc -v deploy-madmail mad0
|
|
||||||
cmlxc -v test-cmdeploy cm0 mad0
|
|
||||||
cmlxc -v test-mini cm0 mad0
|
|
||||||
cmlxc -v test-mini mad0 cm0
|
|
||||||
|
|||||||
37
.github/workflows/docker-dispatch.yaml
vendored
37
.github/workflows/docker-dispatch.yaml
vendored
@@ -1,37 +0,0 @@
|
|||||||
# Notify the docker repo to build and test a new image after relay CI passes.
|
|
||||||
#
|
|
||||||
# Sends a repository_dispatch event to chatmail/docker with the relay ref
|
|
||||||
# and short SHA, which triggers docker-ci.yaml to build, push to GHCR,
|
|
||||||
# and run integration tests via cmlxc.
|
|
||||||
|
|
||||||
name: Trigger Docker build
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
branches: [main]
|
|
||||||
workflow_dispatch:
|
|
||||||
|
|
||||||
permissions: {}
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
dispatch:
|
|
||||||
name: Dispatch build to chatmail/docker
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
if: github.repository == 'chatmail/relay'
|
|
||||||
steps:
|
|
||||||
- name: Compute short SHA
|
|
||||||
id: sha
|
|
||||||
run: echo "short=$(echo '${{ github.sha }}' | cut -c1-7)" >> "$GITHUB_OUTPUT"
|
|
||||||
|
|
||||||
- name: Send repository_dispatch
|
|
||||||
uses: peter-evans/repository-dispatch@ff45666b9427631e3450c54a1bcbee4d9ff4d7c0 # v3
|
|
||||||
with:
|
|
||||||
token: ${{ secrets.CHATMAIL_DOCKER_DISPATCH_TOKEN }}
|
|
||||||
repository: chatmail/docker
|
|
||||||
event-type: relay-updated
|
|
||||||
client-payload: >-
|
|
||||||
{
|
|
||||||
"relay_ref": "${{ github.ref_name }}",
|
|
||||||
"relay_sha": "${{ github.sha }}",
|
|
||||||
"relay_sha_short": "${{ steps.sha.outputs.short }}"
|
|
||||||
}
|
|
||||||
61
.github/workflows/docs-preview.yaml
vendored
61
.github/workflows/docs-preview.yaml
vendored
@@ -1,61 +0,0 @@
|
|||||||
name: documentation preview
|
|
||||||
|
|
||||||
on:
|
|
||||||
pull_request:
|
|
||||||
paths:
|
|
||||||
- 'doc/**'
|
|
||||||
- 'scripts/build-docs.sh'
|
|
||||||
- '.github/workflows/docs-preview.yaml'
|
|
||||||
|
|
||||||
permissions: {}
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
scripts:
|
|
||||||
name: build
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
environment:
|
|
||||||
name: 'staging.chatmail.at/doc/relay/'
|
|
||||||
url: https://staging.chatmail.at/doc/relay/${{ steps.prepare.outputs.prid }}
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v4
|
|
||||||
with:
|
|
||||||
persist-credentials: false
|
|
||||||
|
|
||||||
- name: initenv
|
|
||||||
run: scripts/initenv.sh
|
|
||||||
|
|
||||||
- name: append venv/bin to PATH
|
|
||||||
run: echo `pwd`/venv/bin >>$GITHUB_PATH
|
|
||||||
|
|
||||||
- name: build documentation
|
|
||||||
working-directory: doc
|
|
||||||
run: sphinx-build source build
|
|
||||||
|
|
||||||
- name: build documentation second time (for TOC)
|
|
||||||
working-directory: doc
|
|
||||||
run: sphinx-build source build
|
|
||||||
|
|
||||||
- name: Get Pullrequest ID
|
|
||||||
id: prepare
|
|
||||||
run: |
|
|
||||||
export PULLREQUEST_ID=$(echo "${GITHUB_REF}" | cut -d "/" -f3)
|
|
||||||
echo "prid=$PULLREQUEST_ID" >> $GITHUB_OUTPUT
|
|
||||||
if [ $(expr length "${{ secrets.USERNAME }}") -gt "1" ]; then echo "uploadtoserver=true" >> $GITHUB_OUTPUT; fi
|
|
||||||
- run: |
|
|
||||||
echo "baseurl: /${STEPS_PREPARE_OUTPUTS_PRID}" >> _config.yml
|
|
||||||
env:
|
|
||||||
STEPS_PREPARE_OUTPUTS_PRID: ${{ steps.prepare.outputs.prid }}
|
|
||||||
|
|
||||||
- name: Upload preview
|
|
||||||
run: |
|
|
||||||
mkdir -p "$HOME/.ssh"
|
|
||||||
echo "${{ secrets.CHATMAIL_STAGING_SSHKEY }}" > "$HOME/.ssh/key"
|
|
||||||
chmod 600 "$HOME/.ssh/key"
|
|
||||||
rsync -rILvh -e "ssh -i $HOME/.ssh/key -o StrictHostKeyChecking=no" $GITHUB_WORKSPACE/doc/build/ "${{ secrets.USERNAME }}@chatmail.at:/var/www/html/staging.chatmail.at/doc/relay/${STEPS_PREPARE_OUTPUTS_PRID}/"
|
|
||||||
env:
|
|
||||||
STEPS_PREPARE_OUTPUTS_PRID: ${{ steps.prepare.outputs.prid }}
|
|
||||||
|
|
||||||
- name: check links
|
|
||||||
working-directory: doc
|
|
||||||
run: sphinx-build --builder linkcheck source build
|
|
||||||
|
|
||||||
51
.github/workflows/docs.yaml
vendored
51
.github/workflows/docs.yaml
vendored
@@ -1,51 +0,0 @@
|
|||||||
name: build and upload documentation
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
branches:
|
|
||||||
- main
|
|
||||||
- 'missytake/docs-ci'
|
|
||||||
paths:
|
|
||||||
- 'doc/**'
|
|
||||||
- 'scripts/build-docs.sh'
|
|
||||||
- '.github/workflows/docs.yaml'
|
|
||||||
|
|
||||||
permissions: {}
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
scripts:
|
|
||||||
name: build
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
environment:
|
|
||||||
name: 'chatmail.at/doc/relay/'
|
|
||||||
url: https://chatmail.at/doc/relay/
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v4
|
|
||||||
with:
|
|
||||||
persist-credentials: false
|
|
||||||
|
|
||||||
- name: initenv
|
|
||||||
run: scripts/initenv.sh
|
|
||||||
|
|
||||||
- name: append venv/bin to PATH
|
|
||||||
run: echo `pwd`/venv/bin >>$GITHUB_PATH
|
|
||||||
|
|
||||||
- name: build documentation
|
|
||||||
working-directory: doc
|
|
||||||
run: sphinx-build source build
|
|
||||||
|
|
||||||
- name: build documentation second time (for TOC)
|
|
||||||
working-directory: doc
|
|
||||||
run: sphinx-build source build
|
|
||||||
|
|
||||||
- name: check links
|
|
||||||
working-directory: doc
|
|
||||||
run: sphinx-build --builder linkcheck source build
|
|
||||||
|
|
||||||
- name: upload documentation
|
|
||||||
run: |
|
|
||||||
mkdir -p "$HOME/.ssh"
|
|
||||||
echo "${{ secrets.CHATMAIL_STAGING_SSHKEY }}" > "$HOME/.ssh/key"
|
|
||||||
chmod 600 "$HOME/.ssh/key"
|
|
||||||
rsync -rILvh -e "ssh -i $HOME/.ssh/key -o StrictHostKeyChecking=no" $GITHUB_WORKSPACE/doc/build/ "${{ secrets.USERNAME }}@chatmail.at:/var/www/html/chatmail.at/doc/relay/"
|
|
||||||
|
|
||||||
100
.github/workflows/test-and-deploy-ipv4only.yaml
vendored
Normal file
100
.github/workflows/test-and-deploy-ipv4only.yaml
vendored
Normal file
@@ -0,0 +1,100 @@
|
|||||||
|
name: deploy on staging-ipv4.testrun.org, and run tests
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
pull_request:
|
||||||
|
paths-ignore:
|
||||||
|
- 'scripts/**'
|
||||||
|
- '**/README.md'
|
||||||
|
- 'CHANGELOG.md'
|
||||||
|
- 'LICENSE'
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
deploy:
|
||||||
|
name: deploy on staging-ipv4.testrun.org, and run tests
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
timeout-minutes: 30
|
||||||
|
concurrency:
|
||||||
|
group: ci-ipv4-${{ github.workflow }}-${{ github.ref }}
|
||||||
|
cancel-in-progress: ${{ !contains(github.ref, '$GITHUB_REF') }}
|
||||||
|
steps:
|
||||||
|
- uses: jsok/serialize-workflow-action@v1
|
||||||
|
with:
|
||||||
|
github-token: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: prepare SSH
|
||||||
|
run: |
|
||||||
|
mkdir ~/.ssh
|
||||||
|
echo "${{ secrets.STAGING_SSH_KEY }}" >> ~/.ssh/id_ed25519
|
||||||
|
chmod 600 ~/.ssh/id_ed25519
|
||||||
|
ssh-keyscan staging-ipv4.testrun.org > ~/.ssh/known_hosts
|
||||||
|
# save previous acme & dkim state
|
||||||
|
rsync -avz root@staging-ipv4.testrun.org:/var/lib/acme acme-ipv4 || true
|
||||||
|
rsync -avz root@staging-ipv4.testrun.org:/etc/dkimkeys dkimkeys-ipv4 || true
|
||||||
|
# store previous acme & dkim state on ns.testrun.org, if it contains useful certs
|
||||||
|
if [ -f dkimkeys-ipv4/dkimkeys/opendkim.private ]; then rsync -avz -e "ssh -o StrictHostKeyChecking=accept-new" dkimkeys-ipv4 root@ns.testrun.org:/tmp/ || true; fi
|
||||||
|
if [ "$(ls -A acme-ipv4/acme/certs)" ]; then rsync -avz -e "ssh -o StrictHostKeyChecking=accept-new" acme-ipv4 root@ns.testrun.org:/tmp/ || true; fi
|
||||||
|
# make sure CAA record isn't set
|
||||||
|
scp -o StrictHostKeyChecking=accept-new .github/workflows/staging-ipv4.testrun.org-default.zone root@ns.testrun.org:/etc/nsd/staging-ipv4.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org sed -i '/CAA/d' /etc/nsd/staging-ipv4.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org nsd-checkzone staging-ipv4.testrun.org /etc/nsd/staging-ipv4.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org systemctl reload nsd
|
||||||
|
|
||||||
|
- name: rebuild staging-ipv4.testrun.org to have a clean VPS
|
||||||
|
run: |
|
||||||
|
curl -X POST \
|
||||||
|
-H "Authorization: Bearer ${{ secrets.HETZNER_API_TOKEN }}" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d '{"image":"debian-12"}' \
|
||||||
|
"https://api.hetzner.cloud/v1/servers/${{ secrets.STAGING_IPV4_SERVER_ID }}/actions/rebuild"
|
||||||
|
|
||||||
|
- run: scripts/initenv.sh
|
||||||
|
|
||||||
|
- name: append venv/bin to PATH
|
||||||
|
run: echo venv/bin >>$GITHUB_PATH
|
||||||
|
|
||||||
|
- name: upload TLS cert after rebuilding
|
||||||
|
run: |
|
||||||
|
echo " --- wait until staging-ipv4.testrun.org VPS is rebuilt --- "
|
||||||
|
rm ~/.ssh/known_hosts
|
||||||
|
while ! ssh -o ConnectTimeout=180 -o StrictHostKeyChecking=accept-new -v root@staging-ipv4.testrun.org id -u ; do sleep 1 ; done
|
||||||
|
ssh -o StrictHostKeyChecking=accept-new -v root@staging-ipv4.testrun.org id -u
|
||||||
|
# download acme & dkim state from ns.testrun.org
|
||||||
|
rsync -e "ssh -o StrictHostKeyChecking=accept-new" -avz root@ns.testrun.org:/tmp/acme-ipv4/acme acme-restore || true
|
||||||
|
rsync -avz root@ns.testrun.org:/tmp/dkimkeys-ipv4/dkimkeys dkimkeys-restore || true
|
||||||
|
# restore acme & dkim state to staging2.testrun.org
|
||||||
|
rsync -avz acme-restore/acme root@staging-ipv4.testrun.org:/var/lib/ || true
|
||||||
|
rsync -avz dkimkeys-restore/dkimkeys root@staging-ipv4.testrun.org:/etc/ || true
|
||||||
|
ssh -o StrictHostKeyChecking=accept-new -v root@staging-ipv4.testrun.org chown root:root -R /var/lib/acme || true
|
||||||
|
|
||||||
|
- name: run formatting checks
|
||||||
|
run: cmdeploy fmt -v
|
||||||
|
|
||||||
|
- name: run deploy-chatmail offline tests
|
||||||
|
run: pytest --pyargs cmdeploy
|
||||||
|
|
||||||
|
- run: |
|
||||||
|
cmdeploy init staging-ipv4.testrun.org
|
||||||
|
sed -i 's#disable_ipv6 = False#disable_ipv6 = True#' chatmail.ini
|
||||||
|
|
||||||
|
- run: cmdeploy run
|
||||||
|
|
||||||
|
- name: set DNS entries
|
||||||
|
run: |
|
||||||
|
ssh -o StrictHostKeyChecking=accept-new -v root@staging-ipv4.testrun.org chown opendkim:opendkim -R /etc/dkimkeys
|
||||||
|
cmdeploy dns --zonefile staging-generated.zone
|
||||||
|
cat staging-generated.zone >> .github/workflows/staging-ipv4.testrun.org-default.zone
|
||||||
|
cat .github/workflows/staging-ipv4.testrun.org-default.zone
|
||||||
|
scp .github/workflows/staging-ipv4.testrun.org-default.zone root@ns.testrun.org:/etc/nsd/staging-ipv4.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org nsd-checkzone staging-ipv4.testrun.org /etc/nsd/staging-ipv4.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org systemctl reload nsd
|
||||||
|
|
||||||
|
- name: cmdeploy test
|
||||||
|
run: CHATMAIL_DOMAIN2=nine.testrun.org cmdeploy test --slow
|
||||||
|
|
||||||
|
- name: cmdeploy dns
|
||||||
|
run: cmdeploy dns -v
|
||||||
|
|
||||||
98
.github/workflows/test-and-deploy.yaml
vendored
Normal file
98
.github/workflows/test-and-deploy.yaml
vendored
Normal file
@@ -0,0 +1,98 @@
|
|||||||
|
name: deploy on staging2.testrun.org, and run tests
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
pull_request:
|
||||||
|
paths-ignore:
|
||||||
|
- 'scripts/**'
|
||||||
|
- '**/README.md'
|
||||||
|
- 'CHANGELOG.md'
|
||||||
|
- 'LICENSE'
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
deploy:
|
||||||
|
name: deploy on staging2.testrun.org, and run tests
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
timeout-minutes: 30
|
||||||
|
concurrency:
|
||||||
|
group: ci-${{ github.workflow }}-${{ github.ref }}
|
||||||
|
cancel-in-progress: ${{ !contains(github.ref, '$GITHUB_REF') }}
|
||||||
|
steps:
|
||||||
|
- uses: jsok/serialize-workflow-action@v1
|
||||||
|
with:
|
||||||
|
github-token: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: prepare SSH
|
||||||
|
run: |
|
||||||
|
mkdir ~/.ssh
|
||||||
|
echo "${{ secrets.STAGING_SSH_KEY }}" >> ~/.ssh/id_ed25519
|
||||||
|
chmod 600 ~/.ssh/id_ed25519
|
||||||
|
ssh-keyscan staging2.testrun.org > ~/.ssh/known_hosts
|
||||||
|
# save previous acme & dkim state
|
||||||
|
rsync -avz root@staging2.testrun.org:/var/lib/acme . || true
|
||||||
|
rsync -avz root@staging2.testrun.org:/etc/dkimkeys . || true
|
||||||
|
# store previous acme & dkim state on ns.testrun.org, if it contains useful certs
|
||||||
|
if [ -f dkimkeys/opendkim.private ]; then rsync -avz -e "ssh -o StrictHostKeyChecking=accept-new" dkimkeys root@ns.testrun.org:/tmp/ || true; fi
|
||||||
|
if [ "$(ls -A acme/certs)" ]; then rsync -avz -e "ssh -o StrictHostKeyChecking=accept-new" acme root@ns.testrun.org:/tmp/ || true; fi
|
||||||
|
# make sure CAA record isn't set
|
||||||
|
scp -o StrictHostKeyChecking=accept-new .github/workflows/staging.testrun.org-default.zone root@ns.testrun.org:/etc/nsd/staging2.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org sed -i '/CAA/d' /etc/nsd/staging2.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org nsd-checkzone staging2.testrun.org /etc/nsd/staging2.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org systemctl reload nsd
|
||||||
|
|
||||||
|
- name: rebuild staging2.testrun.org to have a clean VPS
|
||||||
|
run: |
|
||||||
|
curl -X POST \
|
||||||
|
-H "Authorization: Bearer ${{ secrets.HETZNER_API_TOKEN }}" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d '{"image":"debian-12"}' \
|
||||||
|
"https://api.hetzner.cloud/v1/servers/${{ secrets.STAGING_SERVER_ID }}/actions/rebuild"
|
||||||
|
|
||||||
|
- run: scripts/initenv.sh
|
||||||
|
|
||||||
|
- name: append venv/bin to PATH
|
||||||
|
run: echo venv/bin >>$GITHUB_PATH
|
||||||
|
|
||||||
|
- name: upload TLS cert after rebuilding
|
||||||
|
run: |
|
||||||
|
echo " --- wait until staging2.testrun.org VPS is rebuilt --- "
|
||||||
|
rm ~/.ssh/known_hosts
|
||||||
|
while ! ssh -o ConnectTimeout=180 -o StrictHostKeyChecking=accept-new -v root@staging2.testrun.org id -u ; do sleep 1 ; done
|
||||||
|
ssh -o StrictHostKeyChecking=accept-new -v root@staging2.testrun.org id -u
|
||||||
|
# download acme & dkim state from ns.testrun.org
|
||||||
|
rsync -e "ssh -o StrictHostKeyChecking=accept-new" -avz root@ns.testrun.org:/tmp/acme acme-restore || true
|
||||||
|
rsync -avz root@ns.testrun.org:/tmp/dkimkeys dkimkeys-restore || true
|
||||||
|
# restore acme & dkim state to staging2.testrun.org
|
||||||
|
rsync -avz acme-restore/acme root@staging2.testrun.org:/var/lib/ || true
|
||||||
|
rsync -avz dkimkeys-restore/dkimkeys root@staging2.testrun.org:/etc/ || true
|
||||||
|
ssh -o StrictHostKeyChecking=accept-new -v root@staging2.testrun.org chown root:root -R /var/lib/acme || true
|
||||||
|
|
||||||
|
- name: run formatting checks
|
||||||
|
run: cmdeploy fmt -v
|
||||||
|
|
||||||
|
- name: run deploy-chatmail offline tests
|
||||||
|
run: pytest --pyargs cmdeploy
|
||||||
|
|
||||||
|
- run: cmdeploy init staging2.testrun.org
|
||||||
|
|
||||||
|
- run: cmdeploy run --verbose
|
||||||
|
|
||||||
|
- name: set DNS entries
|
||||||
|
run: |
|
||||||
|
ssh -o StrictHostKeyChecking=accept-new root@staging2.testrun.org chown opendkim:opendkim -R /etc/dkimkeys
|
||||||
|
cmdeploy dns --zonefile staging-generated.zone --verbose
|
||||||
|
cat staging-generated.zone >> .github/workflows/staging.testrun.org-default.zone
|
||||||
|
cat .github/workflows/staging.testrun.org-default.zone
|
||||||
|
scp .github/workflows/staging.testrun.org-default.zone root@ns.testrun.org:/etc/nsd/staging2.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org nsd-checkzone staging2.testrun.org /etc/nsd/staging2.testrun.org.zone
|
||||||
|
ssh root@ns.testrun.org systemctl reload nsd
|
||||||
|
|
||||||
|
- name: cmdeploy test
|
||||||
|
run: CHATMAIL_DOMAIN2=nine.testrun.org cmdeploy test --slow
|
||||||
|
|
||||||
|
- name: cmdeploy dns
|
||||||
|
run: cmdeploy dns -v
|
||||||
|
|
||||||
26
.github/workflows/zizmor-scan.yml
vendored
26
.github/workflows/zizmor-scan.yml
vendored
@@ -1,26 +0,0 @@
|
|||||||
name: GitHub Actions Security Analysis with zizmor
|
|
||||||
|
|
||||||
on:
|
|
||||||
push:
|
|
||||||
branches: ["main"]
|
|
||||||
pull_request:
|
|
||||||
branches: ["**"]
|
|
||||||
|
|
||||||
permissions: {}
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
zizmor:
|
|
||||||
name: Run zizmor
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
permissions:
|
|
||||||
security-events: write # Required for upload-sarif (used by zizmor-action) to upload SARIF files.
|
|
||||||
contents: read
|
|
||||||
actions: read
|
|
||||||
steps:
|
|
||||||
- name: Checkout repository
|
|
||||||
uses: actions/checkout@v6
|
|
||||||
with:
|
|
||||||
persist-credentials: false
|
|
||||||
|
|
||||||
- name: Run zizmor
|
|
||||||
uses: zizmorcore/zizmor-action@b1d7e1fb5de872772f31590499237e7cce841e8e # v0.5.3
|
|
||||||
7
.github/zizmor.yml
vendored
7
.github/zizmor.yml
vendored
@@ -1,7 +0,0 @@
|
|||||||
rules:
|
|
||||||
unpinned-uses:
|
|
||||||
config:
|
|
||||||
policies:
|
|
||||||
actions/*: ref-pin
|
|
||||||
dependabot/*: ref-pin
|
|
||||||
chatmail/*: ref-pin
|
|
||||||
2
.gitignore
vendored
2
.gitignore
vendored
@@ -4,7 +4,7 @@ __pycache__/
|
|||||||
*$py.class
|
*$py.class
|
||||||
*.swp
|
*.swp
|
||||||
*qr-*.png
|
*qr-*.png
|
||||||
chatmail*.ini
|
chatmail.ini
|
||||||
|
|
||||||
|
|
||||||
# C extensions
|
# C extensions
|
||||||
|
|||||||
269
CHANGELOG.md
269
CHANGELOG.md
@@ -1,271 +1,10 @@
|
|||||||
# Changelog for chatmail deployment
|
# Changelog for chatmail deployment
|
||||||
|
|
||||||
## 1.10.0 2026-04-30
|
## untagged
|
||||||
|
|
||||||
* start mtail after networking is fully up <https://github.com/chatmail/relay/pull/942>
|
|
||||||
* support specifying custom filtermail binary through environment variable <https://github.com/chatmail/relay/pull/941>
|
|
||||||
* add automated zizmor scanning of github workflows <https://github.com/chatmail/relay/pull/938>
|
|
||||||
* added dispatch for *automated builds of chatmail relay docker images* <https://github.com/chatmail/relay/pull/934>
|
|
||||||
* do not bind SMTP client sockets to public addresses <https://github.com/chatmail/relay/pull/932>
|
|
||||||
* underline in docs that scripts/initenv.sh should be used for building the docs <https://github.com/chatmail/relay/pull/933>
|
|
||||||
* automatic oldest-first message removal from mailboxes to always stay under max_mailbox_size <https://github.com/chatmail/relay/pull/929>
|
|
||||||
* remove --slow from cmdeploy test <https://github.com/chatmail/relay/pull/931>
|
|
||||||
* handle missing inotify sysctl keys in containers <https://github.com/chatmail/relay/pull/930>
|
|
||||||
* replace resolvconf with static resolv.conf <https://github.com/chatmail/relay/pull/928>
|
|
||||||
* disable fsync for LMTP and IMAP services <https://github.com/chatmail/relay/pull/925>
|
|
||||||
* re-use cmlxc workflow, replacing CI with hetzner staging servers with local lxc containers <https://github.com/chatmail/relay/pull/917>
|
|
||||||
* explicitly install resolvconf <https://github.com/chatmail/relay/pull/924>
|
|
||||||
* detect stale dovecot binary and force restart in activate() <https://github.com/chatmail/relay/pull/922>
|
|
||||||
* Rename filtermail_http_port to filtermail_http_port_incoming <https://github.com/chatmail/relay/pull/921>
|
|
||||||
* consolidated is_in_container() check https://github.com/chatmail/relay/pull/920>
|
|
||||||
* restart dovecot after package replacement (rebase, test condense) <https://github.com/chatmail/relay/pull/913>
|
|
||||||
* Set permissions on dovecot pin prefs <https://github.com/chatmail/relay/pull/915>
|
|
||||||
* Route `/mxdeliv/` to configurable port <https://github.com/chatmail/relay/pull/901>
|
|
||||||
* fix VM detection, automated testing fixes, use newer chatmail-turn and move to standard BIND DNS zone format <https://github.com/chatmail/relay/pull/912>
|
|
||||||
* Upgrade to filtermail 0.6.1 <https://github.com/chatmail/relay/pull/910>
|
|
||||||
* pin dovecot packages to prevent apt upgrades <https://github.com/chatmail/relay/pull/908>
|
|
||||||
* add rpc server to cmdeploy along with client <https://github.com/chatmail/relay/pull/906>
|
|
||||||
* remove unused deps from chatmaild <https://github.com/chatmail/relay/pull/905>
|
|
||||||
* set default smtp_tls_security_level to "verify" unconditionally <https://github.com/chatmail/relay/pull/902>
|
|
||||||
* featprefer IPv4 in SMTP client <https://github.com/chatmail/relay/pull/900>
|
|
||||||
* Install dovecot .deb packages atomically <https://github.com/chatmail/relay/pull/899>
|
|
||||||
* stop installing cron package <https://github.com/chatmail/relay/pull/898>
|
|
||||||
* Rewrite dovecot install logic, update <https://github.com/chatmail/relay/pull/862>
|
|
||||||
* fix a test and some linting fixes <https://github.com/chatmail/relay/pull/897>
|
|
||||||
* Disable IP verification on domain-literal addresses <https://github.com/chatmail/relay/pull/895>
|
|
||||||
* disable installing recommended packages globally on the relay <https://github.com/chatmail/relay/pull/887>
|
|
||||||
* multiple bug fixes across chatmaild and cmdeploy <https://github.com/chatmail/relay/pull/883>
|
|
||||||
* remove /metrics from the website <https://github.com/chatmail/relay/pull/703>
|
|
||||||
* add Prometheus textfile output to fsreport <https://github.com/chatmail/relay/pull/881>
|
|
||||||
* chown opendkim: private key <https://github.com/chatmail/relay/pull/879>
|
|
||||||
* make sure chatmail-metadata was started <https://github.com/chatmail/relay/pull/882>
|
|
||||||
* dovecot update url <https://github.com/chatmail/relay/pull/880>
|
|
||||||
* upgrade to filtermail v0.5.2 <https://github.com/chatmail/relay/pull/876>
|
|
||||||
* download dovecot packages from github release <https://github.com/chatmail/relay/pull/875>
|
|
||||||
* replace DKIM verification with filtermail v0.5 <https://github.com/chatmail/relay/pull/831>
|
|
||||||
* remove CFFI deltachat bindings usage, and consolidate test support with rpc-bindings <https://github.com/chatmail/relay/pull/872>
|
|
||||||
* prepare chatmaild/cmdeploy changes for Docker support <https://github.com/chatmail/relay/pull/857>
|
|
||||||
* stabilize online benchmark timing adding rate-limit-aware cooldown between iterations <https://github.com/chatmail/relay/pull/867>
|
|
||||||
* move rate-limit cooldown to benchmark fixture <https://github.com/chatmail/relay/pull/868>
|
|
||||||
* reconfigure acmetool from redirector to proxy mode <https://github.com/chatmail/relay/pull/861>
|
|
||||||
* make tests work with `--ssh-host localhost` <https://github.com/chatmail/relay/pull/856>
|
|
||||||
* mark f-string with f prefix in test_expunged <https://github.com/chatmail/relay/pull/863>
|
|
||||||
* install also if dovecot.service=False in SystemdEnabled Fact <https://github.com/chatmail/relay/pull/841>
|
|
||||||
* Introduce support for self-signed chatmail relays <https://github.com/chatmail/relay/pull/855>
|
|
||||||
* Strip Received headers before delivery <https://github.com/chatmail/relay/pull/849>
|
|
||||||
* upgrade to filtermail v0.3 <https://github.com/chatmail/relay/pull/850>
|
|
||||||
* fix link to Maddy and update madmail URL <https://github.com/chatmail/relay/pull/847>
|
|
||||||
* accept self-signed certificates for IP-only relays <https://github.com/chatmail/relay/pull/846>
|
|
||||||
* enforce sending from public IP addresses <https://github.com/chatmail/relay/pull/845>
|
|
||||||
* port check: check addresses, fix single services <https://github.com/chatmail/relay/pull/844>
|
|
||||||
* remediates issue with improper concat on resolver injection <https://github.com/chatmail/relay/pull/834>
|
|
||||||
* ipv6 boolean not being respected during operations <https://github.com/chatmail/relay/pull/832>
|
|
||||||
* upgrade to filtermail v0.2 by <https://github.com/chatmail/relay/pull/825>
|
|
||||||
* fix link to filtermail <https://github.com/chatmail/relay/pull/824>
|
|
||||||
* print timestamps when sending messages <https://github.com/chatmail/relay/pull/823>
|
|
||||||
* fix flaky test_exceed_rate_limit <https://github.com/chatmail/relay/pull/822>
|
|
||||||
* Replace filtermail with rust reimplementation <https://github.com/chatmail/relay/pull/808>
|
|
||||||
* Set default internal SMTP ports in Config <https://github.com/chatmail/relay/pull/819>
|
|
||||||
* separate metrics for incoming and outgoing messages <https://github.com/chatmail/relay/pull/820>
|
|
||||||
* disable appending the Received header <https://github.com/chatmail/relay/pull/815>
|
|
||||||
* fail on errors in postfix/dovecot config <https://github.com/chatmail/relay/pull/813>
|
|
||||||
* tweak idle/hibernate metrics some more <https://github.com/chatmail/relay/pull/811>
|
|
||||||
* add config flag to export statistics <https://github.com/chatmail/relay/pull/806>
|
|
||||||
* add --website-only option to run subcommand <https://github.com/chatmail/relay/pull/768>
|
|
||||||
* Strip DKIM-Signature header before LMTP <https://github.com/chatmail/relay/pull/803>
|
|
||||||
* properly make sure that postfix gets restarted on failure <https://github.com/chatmail/relay/pull/802>
|
|
||||||
* expire.py: use absolute path to maildirsize <https://github.com/chatmail/relay/pull/807>
|
|
||||||
* pin Dovecot documentation URLs to version 2.3 <https://github.com/chatmail/relay/pull/800>
|
|
||||||
* try to use "build machine" and "deployment server" consistently <https://github.com/chatmail/relay/pull/797>
|
|
||||||
* adds instructions for migrating control machines <https://github.com/chatmail/relay/pull/795>
|
|
||||||
* use consistent naming schema in getting started <https://github.com/chatmail/relay/pull/793>
|
|
||||||
* remove jsok/serialize-workflow-action dependency <https://github.com/chatmail/relay/pull/790>
|
|
||||||
* streamline migration guide wording, provide titled steps <https://github.com/chatmail/relay/pull/789>
|
|
||||||
* increases default max mailbox size <https://github.com/chatmail/relay/pull/792>
|
|
||||||
* use daemon_name for OpenDKIM sign-verify decision instead of IP <https://github.com/chatmail/relay/pull/784>
|
|
||||||
|
|
||||||
## 1.9.0 2025-12-18
|
|
||||||
|
|
||||||
### Documentation
|
|
||||||
|
|
||||||
- Add RELEASE.md and CONTRIBUTING.md
|
|
||||||
- README update, mention Chatmail Cookbook project
|
|
||||||
|
|
||||||
### Bug Fixes
|
|
||||||
|
|
||||||
- Expire messages also from IMAP subfolders
|
|
||||||
- Use absolute path instead of relative path in message expiration script
|
|
||||||
- Restart Postfix and Dovecot automatically on failure
|
|
||||||
- acmetool: Use a fixed name and `reconcile` instead of `want`
|
|
||||||
|
|
||||||
### Features
|
|
||||||
|
|
||||||
- Report DKIM error code in SMTP response
|
|
||||||
- Remove development notice from the web pages
|
|
||||||
|
|
||||||
### Miscellaneous Tasks
|
|
||||||
|
|
||||||
- Update the heading in the CHANGELOG.md
|
|
||||||
- Setup git-cliff
|
|
||||||
- Run tests against ci-chatmail.testrun.org instead of nine.testrun.org
|
|
||||||
- Cleanup remaining echobot code, remove echobot user from deployment and passthrough recipients
|
|
||||||
|
|
||||||
## 1.8.0 2025-12-12
|
|
||||||
|
|
||||||
- Add imap_compress option to chatmail.ini
|
|
||||||
([#760](https://github.com/chatmail/relay/pull/760))
|
|
||||||
|
|
||||||
- Remove echobot from relays
|
|
||||||
([#753](https://github.com/chatmail/relay/pull/753))
|
|
||||||
|
|
||||||
- Fix `cmdeploy webdev`
|
|
||||||
([#743](https://github.com/chatmail/relay/pull/743))
|
|
||||||
|
|
||||||
- Add robots.txt to exclude all web crawlers
|
|
||||||
([#732](https://github.com/chatmail/relay/pull/732))
|
|
||||||
|
|
||||||
- acmetool: accept new Let's Encrypt ToS: https://letsencrypt.org/documents/LE-SA-v1.6-August-18-2025.pdf
|
|
||||||
([#729](https://github.com/chatmail/relay/pull/729))
|
|
||||||
|
|
||||||
- Organized cmdeploy into install, configure, and activate stages
|
|
||||||
([#695](https://github.com/chatmail/relay/pull/695))
|
|
||||||
|
|
||||||
- docs: move readme.md docs to sphinx documentation rendered at https://chatmail.at/doc/relay
|
|
||||||
([#711](https://github.com/chatmail/relay/pull/711))
|
|
||||||
|
|
||||||
- acmetool: replace cronjob with a systemd timer
|
|
||||||
([#719](https://github.com/chatmail/relay/pull/719))
|
|
||||||
|
|
||||||
- remove xstore@testrun.org from default passthrough recipients
|
|
||||||
([#722](https://github.com/chatmail/relay/pull/722))
|
|
||||||
|
|
||||||
- don't deploy the website if there are merge conflicts in the www folder
|
|
||||||
([#714](https://github.com/chatmail/relay/pull/714))
|
|
||||||
|
|
||||||
- acmetool: use ECDSA keys instead of RSA
|
|
||||||
([#689](https://github.com/chatmail/relay/pull/689))
|
|
||||||
|
|
||||||
- Require TLS 1.2 for outgoing SMTP connections
|
|
||||||
([#685](https://github.com/chatmail/relay/pull/685), [#730](https://github.com/chatmail/relay/pull/730))
|
|
||||||
|
|
||||||
- require STARTTLS for incoming port 25 connections
|
|
||||||
([#684](https://github.com/chatmail/relay/pull/684), [#730](https://github.com/chatmail/relay/pull/730))
|
|
||||||
|
|
||||||
- filtermail: run CPU-intensive handle_DATA in a thread pool executor
|
|
||||||
([#676](https://github.com/chatmail/relay/pull/676))
|
|
||||||
|
|
||||||
- don't use the complicated logging module in filtermail to exclude a potential source of errors.
|
|
||||||
([#674](https://github.com/chatmail/relay/pull/674))
|
|
||||||
|
|
||||||
- Specify nginx.conf to only handle `mail_domain`, www, and mta-sts domains
|
|
||||||
([#636](https://github.com/chatmail/relay/pull/636))
|
|
||||||
|
|
||||||
- Setup TURN server
|
|
||||||
([#621](https://github.com/chatmail/relay/pull/621))
|
|
||||||
|
|
||||||
- cmdeploy: make --ssh-host work with localhost
|
|
||||||
([#659](https://github.com/chatmail/relay/pull/659))
|
|
||||||
|
|
||||||
- Update iroh-relay to 0.35.0
|
|
||||||
([#650](https://github.com/chatmail/relay/pull/650))
|
|
||||||
|
|
||||||
- filtermail: accept mails from Protonmail
|
|
||||||
([#616](https://github.com/chatmail/relay/pull/616))
|
|
||||||
|
|
||||||
- Ignore all RCPT TO: parameters
|
|
||||||
([#651](https://github.com/chatmail/relay/pull/651))
|
|
||||||
|
|
||||||
- Increase opendkim DNS Timeout from 5 to 60 seconds
|
|
||||||
([#672](https://github.com/chatmail/relay/pull/672))
|
|
||||||
|
|
||||||
- Add config parameter for Let's Encrypt ACME email
|
|
||||||
([#663](https://github.com/chatmail/relay/pull/663))
|
|
||||||
|
|
||||||
- Use max username length in newemail.py, not min
|
|
||||||
([#648](https://github.com/chatmail/relay/pull/648))
|
|
||||||
|
|
||||||
- Add startup for `fcgiwrap.service` because sometimes it did not start automatically.
|
|
||||||
([#657](https://github.com/chatmail/relay/pull/657))
|
|
||||||
|
|
||||||
- Add `cmdeploy init --force` command for recreating chatmail.ini
|
|
||||||
([#656](https://github.com/chatmail/relay/pull/656))
|
|
||||||
|
|
||||||
- Increase maxproc for reinjecting ports from 10 to 100
|
|
||||||
([#646](https://github.com/chatmail/relay/pull/646))
|
|
||||||
|
|
||||||
- Allow ports 143 and 993 to be used by `dovecot` process
|
|
||||||
([#639](https://github.com/chatmail/relay/pull/639))
|
|
||||||
|
|
||||||
- Add `--skip-dns-check` argument to `cmdeploy run` command, which disables DNS record checking before installation.
|
|
||||||
([#661](https://github.com/chatmail/relay/pull/661))
|
|
||||||
|
|
||||||
- Rework expiry of message files and mailboxes in Python
|
|
||||||
to only do a single iteration over sometimes millions of messages
|
|
||||||
instead of doing "find" commands that iterate 9 times over the messages.
|
|
||||||
Provide an "fsreport" CLI for more fine grained analysis of message files.
|
|
||||||
([#637](https://github.com/chatmail/relay/pull/637))
|
|
||||||
|
|
||||||
|
|
||||||
## 1.7.0 2025-09-11
|
|
||||||
|
|
||||||
- Make www upload path configurable
|
|
||||||
([#618](https://github.com/chatmail/relay/pull/618))
|
|
||||||
|
|
||||||
- Check whether GCC is installed in initenv.sh
|
|
||||||
([#608](https://github.com/chatmail/relay/pull/608))
|
|
||||||
|
|
||||||
- Expire push notification tokens after 90 days
|
|
||||||
([#583](https://github.com/chatmail/relay/pull/583))
|
|
||||||
|
|
||||||
- Use official `mtail` binary instead of `mtail` package
|
|
||||||
([#581](https://github.com/chatmail/relay/pull/581))
|
|
||||||
|
|
||||||
- dovecot: install from download.delta.chat instead of openSUSE Build Service
|
|
||||||
([#590](https://github.com/chatmail/relay/pull/590))
|
|
||||||
|
|
||||||
- Reconfigure Dovecot imap-login service to high-performance mode
|
|
||||||
([#578](https://github.com/chatmail/relay/pull/578))
|
|
||||||
|
|
||||||
- Set timezone to improve dovecot performance
|
|
||||||
([#584](https://github.com/chatmail/relay/pull/584))
|
|
||||||
|
|
||||||
- Increase nginx connection limits
|
|
||||||
([#576](https://github.com/chatmail/relay/pull/576))
|
|
||||||
|
|
||||||
- If `dns-utils` needs to be installed before cmdeploy run, apt update to make sure it works
|
|
||||||
([#560](https://github.com/chatmail/relay/pull/560))
|
|
||||||
|
|
||||||
- filtermail: respect config message size limit
|
|
||||||
([#572](https://github.com/chatmail/relay/pull/572))
|
|
||||||
|
|
||||||
- Don't deploy if one of the ports used for chatmail relay services is occupied by an unexpected process
|
|
||||||
([#568](https://github.com/chatmail/relay/pull/568))
|
|
||||||
|
|
||||||
- Add config value after how many days large files are deleted
|
|
||||||
([#555](https://github.com/chatmail/relay/pull/555))
|
|
||||||
|
|
||||||
- cmdeploy: push relay version to /etc/chatmail-version
|
|
||||||
([#573](https://github.com/chatmail/relay/pull/573))
|
|
||||||
|
|
||||||
- filtermail: allow partial body length in OpenPGP payloads
|
|
||||||
([#570](https://github.com/chatmail/relay/pull/570))
|
|
||||||
|
|
||||||
- chatmaild: allow echobot to receive unencrypted messages by default
|
|
||||||
([#556](https://github.com/chatmail/relay/pull/556))
|
|
||||||
|
|
||||||
|
|
||||||
## 1.6.0 2025-04-11
|
|
||||||
|
|
||||||
- Handle Port-25 connect errors more gracefully (common with VPNs)
|
|
||||||
([#552](https://github.com/chatmail/relay/pull/552))
|
|
||||||
|
|
||||||
- Avoid "acmetool not found" during initial run
|
- Avoid "acmetool not found" during initial run
|
||||||
([#550](https://github.com/chatmail/relay/pull/550))
|
([#550](https://github.com/chatmail/relay/pull/550))
|
||||||
|
|
||||||
- Fix timezone handling such that client/servers do not need to use
|
|
||||||
same timezone.
|
|
||||||
([#553](https://github.com/chatmail/relay/pull/553))
|
|
||||||
|
|
||||||
- Enforce end-to-end encryption for incoming messages.
|
- Enforce end-to-end encryption for incoming messages.
|
||||||
New user address mailboxes now get a `enforceE2EEincoming` file
|
New user address mailboxes now get a `enforceE2EEincoming` file
|
||||||
which prohibits incoming cleartext messages from other domains.
|
which prohibits incoming cleartext messages from other domains.
|
||||||
@@ -278,12 +17,6 @@
|
|||||||
- Enforce end-to-end encryption between local addresses
|
- Enforce end-to-end encryption between local addresses
|
||||||
([#535](https://github.com/chatmail/server/pull/535))
|
([#535](https://github.com/chatmail/server/pull/535))
|
||||||
|
|
||||||
- unbound: check that port 53 is not occupied by a different process
|
|
||||||
([#537](https://github.com/chatmail/server/pull/537))
|
|
||||||
|
|
||||||
- unbound: before unbound is there, use 9.9.9.9 for resolving
|
|
||||||
([#518](https://github.com/chatmail/relay/pull/518))
|
|
||||||
|
|
||||||
- Limit the bind for the HTTPS server on 8443 to 127.0.0.1
|
- Limit the bind for the HTTPS server on 8443 to 127.0.0.1
|
||||||
([#522](https://github.com/chatmail/server/pull/522))
|
([#522](https://github.com/chatmail/server/pull/522))
|
||||||
([#532](https://github.com/chatmail/server/pull/532))
|
([#532](https://github.com/chatmail/server/pull/532))
|
||||||
|
|||||||
@@ -1,7 +0,0 @@
|
|||||||
# Contributing to the chatmail relay
|
|
||||||
|
|
||||||
Commit messages follow the [Conventional Commits] notation.
|
|
||||||
We use [git-cliff] to generate the changelog from commit messages before the release.
|
|
||||||
|
|
||||||
[Conventional Commits]: https://www.conventionalcommits.org/
|
|
||||||
[git-cliff]: https://git-cliff.org/
|
|
||||||
535
README.md
535
README.md
@@ -1,20 +1,535 @@
|
|||||||
|
|
||||||
# Chatmail relays for end-to-end encrypted email
|
<img width="800px" src="www/src/collage-top.png"/>
|
||||||
|
|
||||||
|
# Chatmail relays for end-to-end encrypted e-mail
|
||||||
|
|
||||||
Chatmail relay servers are interoperable Mail Transport Agents (MTAs) designed for:
|
Chatmail relay servers are interoperable Mail Transport Agents (MTAs) designed for:
|
||||||
|
|
||||||
- **Zero State:** no private data or metadata collected, messages are auto-deleted, low disk usage
|
- **Convenience:** Low friction instant onboarding
|
||||||
|
|
||||||
- **Instant/Realtime:** sub-second message delivery, realtime P2P
|
- **Privacy:** No name, phone numbers, email required or collected
|
||||||
streaming, privacy-preserving Push Notifications for Apple, Google, and Huawei;
|
|
||||||
|
|
||||||
- **Security Enforcement**: only strict TLS, DKIM and OpenPGP with minimized metadata accepted
|
- **End-to-End Encryption enforced**: only OpenPGP messages with metadata minimization allowed
|
||||||
|
|
||||||
- **Reliable Federation and Decentralization:** No spam or IP reputation checks, federating
|
- **Instant:** Privacy-preserving Push Notifications for Apple, Google, and Huawei
|
||||||
depends on established IETF standards and protocols.
|
|
||||||
|
|
||||||
This repository contains everything needed to setup a ready-to-use chatmail relay on an ssh-reachable host.
|
- **Speed:** Message delivery in half a second, with optional P2P realtime connections
|
||||||
For getting started and more information please refer to the web version of this repositories' documentation at
|
|
||||||
|
|
||||||
[https://chatmail.at/doc/relay](https://chatmail.at/doc/relay)
|
- **Transport Security:** Strict TLS and DKIM enforced
|
||||||
|
|
||||||
|
- **Reliability:** No spam or IP reputation checks; rate-limits are suitable for realtime chats
|
||||||
|
|
||||||
|
- **Efficiency:** Messages are only stored for transit and removed automatically
|
||||||
|
|
||||||
|
This repository contains everything needed to setup a ready-to-use chatmail relay
|
||||||
|
comprised of a minimal setup of the battle-tested
|
||||||
|
[Postfix SMTP](https://www.postfix.org) and [Dovecot IMAP](https://www.dovecot.org) MTAs/MDAs.
|
||||||
|
|
||||||
|
The automated setup is designed and optimized for providing chatmail addresses
|
||||||
|
for immediate permission-free onboarding through chat apps and bots.
|
||||||
|
Chatmail addresses are automatically created at first login,
|
||||||
|
after which the initially specified password is required
|
||||||
|
for sending and receiving messages through them.
|
||||||
|
|
||||||
|
Please see [this list of known apps and client projects](https://chatmail.at/clients.html)
|
||||||
|
and [this list of known public 3rd party chatmail relay servers](https://chatmail.at/relays).
|
||||||
|
|
||||||
|
|
||||||
|
## Minimal requirements, Prerequisites
|
||||||
|
|
||||||
|
You will need the following:
|
||||||
|
|
||||||
|
- Control over a domain through a DNS provider of your choice.
|
||||||
|
|
||||||
|
- A Debian 12 server with reachable SMTP/SUBMISSIONS/IMAPS/HTTPS ports.
|
||||||
|
IPv6 is encouraged if available.
|
||||||
|
Chatmail relay servers only require 1GB RAM, one CPU, and perhaps 10GB storage for a
|
||||||
|
few thousand active chatmail addresses.
|
||||||
|
|
||||||
|
- Key-based SSH authentication to the root user.
|
||||||
|
You must add a passphrase-protected private key to your local ssh-agent
|
||||||
|
because you can't type in your passphrase during deployment.
|
||||||
|
(An ed25519 private key is required due to an [upstream bug in paramiko](https://github.com/paramiko/paramiko/issues/2191))
|
||||||
|
|
||||||
|
|
||||||
|
## Getting started
|
||||||
|
|
||||||
|
We use `chat.example.org` as the chatmail domain in the following steps.
|
||||||
|
Please substitute it with your own domain.
|
||||||
|
|
||||||
|
1. Setup the initial DNS records.
|
||||||
|
The following is an example in the familiar BIND zone file format with
|
||||||
|
a TTL of 1 hour (3600 seconds).
|
||||||
|
Please substitute your domain and IP addresses.
|
||||||
|
|
||||||
|
```
|
||||||
|
chat.example.com. 3600 IN A 198.51.100.5
|
||||||
|
chat.example.com. 3600 IN AAAA 2001:db8::5
|
||||||
|
www.chat.example.com. 3600 IN CNAME chat.example.com.
|
||||||
|
mta-sts.chat.example.com. 3600 IN CNAME chat.example.com.
|
||||||
|
```
|
||||||
|
|
||||||
|
2. Clone the repository and bootstrap the Python virtualenv.
|
||||||
|
|
||||||
|
```
|
||||||
|
git clone https://github.com/chatmail/relay
|
||||||
|
cd relay
|
||||||
|
scripts/initenv.sh
|
||||||
|
```
|
||||||
|
|
||||||
|
3. Create chatmail configuration file `chatmail.ini`:
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy init chat.example.org # <-- use your domain
|
||||||
|
```
|
||||||
|
|
||||||
|
4. Verify that SSH root login works:
|
||||||
|
|
||||||
|
```
|
||||||
|
ssh root@chat.example.org # <-- use your domain
|
||||||
|
```
|
||||||
|
|
||||||
|
|
||||||
|
5. Deploy the remote chatmail relay server:
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy run
|
||||||
|
```
|
||||||
|
This script will check that you have all necessary DNS records.
|
||||||
|
If DNS records are missing, it will recommend
|
||||||
|
which you should configure at your DNS provider
|
||||||
|
(it can take some time until they are public).
|
||||||
|
|
||||||
|
### Other helpful commands:
|
||||||
|
|
||||||
|
To check the status of your remotely running chatmail service:
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy status
|
||||||
|
```
|
||||||
|
|
||||||
|
To display and check all recommended DNS records:
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy dns
|
||||||
|
```
|
||||||
|
|
||||||
|
To test whether your chatmail service is working correctly:
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy test
|
||||||
|
```
|
||||||
|
|
||||||
|
To measure the performance of your chatmail service:
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy bench
|
||||||
|
```
|
||||||
|
|
||||||
|
## Overview of this repository
|
||||||
|
|
||||||
|
This repository has four directories:
|
||||||
|
|
||||||
|
- [cmdeploy](https://github.com/chatmail/relay/tree/main/cmdeploy)
|
||||||
|
is a collection of configuration files
|
||||||
|
and a [pyinfra](https://pyinfra.com)-based deployment script.
|
||||||
|
|
||||||
|
- [chatmaild](https://github.com/chatmail/relay/tree/main/chatmaild)
|
||||||
|
is a Python package containing several small services
|
||||||
|
which handle authentication,
|
||||||
|
trigger push notifications on new messages,
|
||||||
|
ensure that outbound mails are encrypted,
|
||||||
|
delete inactive users,
|
||||||
|
and some other minor things.
|
||||||
|
chatmaild can also be installed as a stand-alone Python package.
|
||||||
|
|
||||||
|
- [www](https://github.com/chatmail/relay/tree/main/www)
|
||||||
|
contains the html, css, and markdown files
|
||||||
|
which make up a chatmail relay's web page.
|
||||||
|
Edit them before deploying to make your chatmail relay stand out.
|
||||||
|
|
||||||
|
- [scripts](https://github.com/chatmail/relay/tree/main/scripts)
|
||||||
|
offers two convenience tools for beginners;
|
||||||
|
`initenv.sh` installs the necessary dependencies to a local virtual environment,
|
||||||
|
and the `scripts/cmdeploy` script enables you
|
||||||
|
to run the `cmdeploy` command line tool in the local virtual environment.
|
||||||
|
|
||||||
|
### cmdeploy
|
||||||
|
|
||||||
|
The `cmdeploy/src/cmdeploy/cmdeploy.py` command line tool
|
||||||
|
helps with setting up and managing the chatmail service.
|
||||||
|
`cmdeploy init` creates the `chatmail.ini` config file.
|
||||||
|
`cmdeploy run` uses a [pyinfra](https://pyinfra.com/)-based [script](`cmdeploy/src/cmdeploy/__init__.py`)
|
||||||
|
to automatically install or upgrade all chatmail components on a relay,
|
||||||
|
according to the `chatmail.ini` config.
|
||||||
|
|
||||||
|
The components of chatmail are:
|
||||||
|
|
||||||
|
- [Postfix SMTP MTA](https://www.postfix.org) accepts and relays messages
|
||||||
|
(both from your users and from the wider e-mail MTA network)
|
||||||
|
|
||||||
|
- [Dovecot IMAP MDA](https://www.dovecot.org) stores messages for your users until they download them
|
||||||
|
|
||||||
|
- [Nginx](https://nginx.org/) shows the web page with your privacy policy and additional information
|
||||||
|
|
||||||
|
- [acmetool](https://hlandau.github.io/acmetool/) manages TLS certificates for Dovecot, Postfix, and Nginx
|
||||||
|
|
||||||
|
- [OpenDKIM](http://www.opendkim.org/) for signing messages with DKIM and rejecting inbound messages without DKIM
|
||||||
|
|
||||||
|
- [mtail](https://google.github.io/mtail/) for collecting anonymized metrics in case you have monitoring
|
||||||
|
|
||||||
|
- [Iroh relay](https://www.iroh.computer/docs/concepts/relay)
|
||||||
|
which helps client devices to establish Peer-to-Peer connections
|
||||||
|
|
||||||
|
- and the chatmaild services, explained in the next section:
|
||||||
|
|
||||||
|
### chatmaild
|
||||||
|
|
||||||
|
`chatmaild` implements various systemd-controlled services
|
||||||
|
that integrate with Dovecot and Postfix to achieve instant-onboarding and
|
||||||
|
only relaying OpenPGP end-to-end messages encrypted messages.
|
||||||
|
A short overview of `chatmaild` services:
|
||||||
|
|
||||||
|
- [`doveauth`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/doveauth.py)
|
||||||
|
implements create-on-login address semantics and is used
|
||||||
|
by Dovecot during IMAP login and by Postfix during SMTP/SUBMISSION login
|
||||||
|
which in turn uses [Dovecot SASL](https://doc.dovecot.org/configuration_manual/authentication/dict/#complete-example-for-authenticating-via-a-unix-socket)
|
||||||
|
to authenticate logins.
|
||||||
|
|
||||||
|
- [`filtermail`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/filtermail.py)
|
||||||
|
prevents unencrypted email from leaving or entering the chatmail service
|
||||||
|
and is integrated into Postfix's outbound and inbound mail pipelines.
|
||||||
|
|
||||||
|
- [`chatmail-metadata`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/metadata.py) is contacted by a
|
||||||
|
[Dovecot lua script](https://github.com/chatmail/relay/blob/main/cmdeploy/src/cmdeploy/dovecot/push_notification.lua)
|
||||||
|
to store user-specific relay-side config.
|
||||||
|
On new messages,
|
||||||
|
it [passes the user's push notification token](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/notifier.py)
|
||||||
|
to [notifications.delta.chat](https://delta.chat/help#instant-delivery)
|
||||||
|
so the push notifications on the user's phone can be triggered
|
||||||
|
by Apple/Google/Huawei.
|
||||||
|
|
||||||
|
- [`delete_inactive_users`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/delete_inactive_users.py)
|
||||||
|
deletes users if they have not logged in for a very long time.
|
||||||
|
The timeframe can be configured in `chatmail.ini`.
|
||||||
|
|
||||||
|
- [`lastlogin`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/lastlogin.py)
|
||||||
|
is contacted by Dovecot when a user logs in
|
||||||
|
and stores the date of the login.
|
||||||
|
|
||||||
|
- [`echobot`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/echo.py)
|
||||||
|
is a small bot for test purposes.
|
||||||
|
It simply echoes back messages from users.
|
||||||
|
|
||||||
|
- [`chatmail-metrics`](https://github.com/chatmail/relay/blob/main/chatmaild/src/chatmaild/metrics.py)
|
||||||
|
collects some metrics and displays them at `https://example.org/metrics`.
|
||||||
|
|
||||||
|
### Home page and getting started for users
|
||||||
|
|
||||||
|
`cmdeploy run` also creates default static web pages and deploys them
|
||||||
|
to a Nginx web server with:
|
||||||
|
|
||||||
|
- a default `index.html` along with a QR code that users can click to
|
||||||
|
create an address on your chatmail relay
|
||||||
|
|
||||||
|
- a default `info.html` that is linked from the home page
|
||||||
|
|
||||||
|
- a default `policy.html` that is linked from the home page
|
||||||
|
|
||||||
|
All `.html` files are generated
|
||||||
|
by the according markdown `.md` file in the `www/src` directory.
|
||||||
|
|
||||||
|
|
||||||
|
### Refining the web pages
|
||||||
|
|
||||||
|
```
|
||||||
|
scripts/cmdeploy webdev
|
||||||
|
```
|
||||||
|
|
||||||
|
This starts a local live development cycle for chatmail web pages:
|
||||||
|
|
||||||
|
- uses the `www/src/page-layout.html` file for producing static
|
||||||
|
HTML pages from `www/src/*.md` files
|
||||||
|
|
||||||
|
- continously builds the web presence reading files from `www/src` directory
|
||||||
|
and generating HTML files and copying assets to the `www/build` directory.
|
||||||
|
|
||||||
|
- Starts a browser window automatically where you can "refresh" as needed.
|
||||||
|
|
||||||
|
## Mailbox directory layout
|
||||||
|
|
||||||
|
Fresh chatmail addresses have a mailbox directory that contains:
|
||||||
|
|
||||||
|
- a `password` file with the salted password required for authenticating
|
||||||
|
whether a login may use the address to send/receive messages.
|
||||||
|
If you modify the password file manually, you effectively block the user.
|
||||||
|
|
||||||
|
- `enforceE2EEincoming` is a default-created file with each address.
|
||||||
|
If present the file indicates that this chatmail address rejects incoming cleartext messages.
|
||||||
|
If absent the address accepts incoming cleartext messages.
|
||||||
|
|
||||||
|
- `dovecot*`, `cur`, `new` and `tmp` represent IMAP/mailbox state.
|
||||||
|
If the address is only used by one device, the Maildir directories
|
||||||
|
will typically be empty unless the user of that address hasn't been online
|
||||||
|
for a while.
|
||||||
|
|
||||||
|
|
||||||
|
## Emergency Commands to disable automatic address creation
|
||||||
|
|
||||||
|
If you need to stop address creation,
|
||||||
|
e.g. because some script is wildly creating addresses,
|
||||||
|
login with ssh and run:
|
||||||
|
|
||||||
|
```
|
||||||
|
touch /etc/chatmail-nocreate
|
||||||
|
```
|
||||||
|
|
||||||
|
Chatmail address creation will be denied while this file is present.
|
||||||
|
|
||||||
|
### Ports
|
||||||
|
|
||||||
|
[Postfix](http://www.postfix.org/) listens on ports 25 (SMTP) and 587 (SUBMISSION) and 465 (SUBMISSIONS).
|
||||||
|
[Dovecot](https://www.dovecot.org/) listens on ports 143 (IMAP) and 993 (IMAPS).
|
||||||
|
[Nginx](https://www.nginx.com/) listens on port 8443 (HTTPS-ALT) and 443 (HTTPS).
|
||||||
|
Port 443 multiplexes HTTPS, IMAP and SMTP using ALPN to redirect connections to ports 8443, 465 or 993.
|
||||||
|
[acmetool](https://hlandau.github.io/acmetool/) listens on port 80 (HTTP).
|
||||||
|
|
||||||
|
chatmail-core based apps will, however, discover all ports and configurations
|
||||||
|
automatically by reading the [autoconfig XML file](https://www.ietf.org/archive/id/draft-bucksch-autoconfig-00.html) from the chatmail relay server.
|
||||||
|
|
||||||
|
## Email authentication
|
||||||
|
|
||||||
|
Chatmail relays enforce [DKIM](https://www.rfc-editor.org/rfc/rfc6376)
|
||||||
|
to authenticate incoming emails.
|
||||||
|
Incoming emails must have a valid DKIM signature with
|
||||||
|
Signing Domain Identifier (SDID, `d=` parameter in the DKIM-Signature header)
|
||||||
|
equal to the `From:` header domain.
|
||||||
|
This property is checked by OpenDKIM screen policy script
|
||||||
|
before validating the signatures.
|
||||||
|
This correpsonds to strict [DMARC](https://www.rfc-editor.org/rfc/rfc7489) alignment (`adkim=s`),
|
||||||
|
but chatmail does not rely on DMARC and does not consult the sender policy published in DMARC records.
|
||||||
|
Other legacy authentication mechanisms such as [iprev](https://www.rfc-editor.org/rfc/rfc8601#section-2.7.3)
|
||||||
|
and [SPF](https://www.rfc-editor.org/rfc/rfc7208) are also not taken into account.
|
||||||
|
If there is no valid DKIM signature on the incoming email,
|
||||||
|
the sender receives a "5.7.1 No valid DKIM signature found" error.
|
||||||
|
|
||||||
|
Outgoing emails must be sent over authenticated connection
|
||||||
|
with envelope MAIL FROM (return path) corresponding to the login.
|
||||||
|
This is ensured by Postfix which maps login username
|
||||||
|
to MAIL FROM with
|
||||||
|
[`smtpd_sender_login_maps`](https://www.postfix.org/postconf.5.html#smtpd_sender_login_maps)
|
||||||
|
and rejects incorrectly authenticated emails with [`reject_sender_login_mismatch`](reject_sender_login_mismatch) policy.
|
||||||
|
`From:` header must correspond to envelope MAIL FROM,
|
||||||
|
this is ensured by `filtermail` proxy.
|
||||||
|
|
||||||
|
## TLS requirements
|
||||||
|
|
||||||
|
Postfix is configured to require valid TLS
|
||||||
|
by setting [`smtp_tls_security_level`](https://www.postfix.org/postconf.5.html#smtp_tls_security_level) to `verify`.
|
||||||
|
If emails don't arrive at your chatmail relay server,
|
||||||
|
the problem is likely that your relay does not have a valid TLS certificate.
|
||||||
|
|
||||||
|
You can test it by resolving `MX` records of your relay domain
|
||||||
|
and then connecting to MX relays (e.g `mx.example.org`) with
|
||||||
|
`openssl s_client -connect mx.example.org:25 -verify_hostname mx.example.org -verify_return_error -starttls smtp`
|
||||||
|
from the host that has open port 25 to verify that certificate is valid.
|
||||||
|
|
||||||
|
When providing a TLS certificate to your chatmail relay server,
|
||||||
|
make sure to provide the full certificate chain
|
||||||
|
and not just the last certificate.
|
||||||
|
|
||||||
|
If you are running an Exim server and don't see incoming connections
|
||||||
|
from a chatmail relay server in the logs,
|
||||||
|
make sure `smtp_no_mail` log item is enabled in the config
|
||||||
|
with `log_selector = +smtp_no_mail`.
|
||||||
|
By default Exim does not log sessions that are closed
|
||||||
|
before sending the `MAIL` command.
|
||||||
|
This happens if certificate is not recognized as valid by Postfix,
|
||||||
|
so you might think that connection is not established
|
||||||
|
while actually it is a problem with your TLS certificate.
|
||||||
|
|
||||||
|
## Migrating a chatmail relay to a new host
|
||||||
|
|
||||||
|
If you want to migrate chatmail relay from an old machine
|
||||||
|
to a new machine,
|
||||||
|
you can use these steps.
|
||||||
|
They were tested with a Linux laptop;
|
||||||
|
you might need to adjust some of the steps to your environment.
|
||||||
|
|
||||||
|
Let's assume that your `mail_domain` is `mail.example.org`,
|
||||||
|
all involved machines run Debian 12,
|
||||||
|
your old site's IP address is `13.37.13.37`,
|
||||||
|
and your new site's IP address is `13.12.23.42`.
|
||||||
|
|
||||||
|
Note, you should lower the TTLs of your DNS records to a value
|
||||||
|
such as 300 (5 minutes) so the migration happens as smoothly as possible.
|
||||||
|
|
||||||
|
During the guide you might get a warning about changed SSH Host keys;
|
||||||
|
in this case, just run `ssh-keygen -R "mail.example.org"` as recommended.
|
||||||
|
|
||||||
|
1. First, disable mail services on the old site.
|
||||||
|
|
||||||
|
```
|
||||||
|
cmdeploy run --disable-mail --ssh-host 13.37.13.37
|
||||||
|
```
|
||||||
|
|
||||||
|
Now your users will notice the migration
|
||||||
|
and will not be able to send or receive messages
|
||||||
|
until the migration is completed.
|
||||||
|
|
||||||
|
2. Now we want to copy `/home/vmail`, `/var/lib/acme`, `/etc/dkimkeys`, `/run/echobot`, and `/var/spool/postfix` to the new site.
|
||||||
|
Login to the old site while forwarding your SSH agent
|
||||||
|
so you can copy directly from the old to the new site with your SSH key:
|
||||||
|
```
|
||||||
|
ssh -A root@13.37.13.37
|
||||||
|
tar c - /home/vmail/mail /var/lib/acme /etc/dkimkeys /run/echobot /var/spool/postfix | ssh root@13.12.23.42 "tar x -C /"
|
||||||
|
```
|
||||||
|
|
||||||
|
This transfers all addresses, the TLS certificate, DKIM keys (so DKIM DNS record remains valid), and the echobot's password so it continues to function.
|
||||||
|
It also preserves the Postfix mail spool so any messages pending delivery will still be delivered.
|
||||||
|
|
||||||
|
3. Install chatmail on the new machine:
|
||||||
|
|
||||||
|
```
|
||||||
|
cmdeploy run --disable-mail --ssh-host 13.12.23.42
|
||||||
|
```
|
||||||
|
Postfix and Dovecot are disabled for now; we will enable them later.
|
||||||
|
We first need to make the new site fully operational.
|
||||||
|
|
||||||
|
3. On the new site, run the following to ensure the ownership is correct in case UIDs/GIDs changed:
|
||||||
|
|
||||||
|
```
|
||||||
|
chown root: -R /var/lib/acme
|
||||||
|
chown opendkim: -R /etc/dkimkeys
|
||||||
|
chown vmail: -R /home/vmail/mail
|
||||||
|
chown echobot: -R /run/echobot
|
||||||
|
```
|
||||||
|
|
||||||
|
4. Now, update DNS entries.
|
||||||
|
|
||||||
|
If other MTAs try to deliver messages to your chatmail domain they may fail intermittently,
|
||||||
|
as DNS catches up with the new site settings
|
||||||
|
but normally will retry delivering messages
|
||||||
|
for at least a week, so messages will not be lost.
|
||||||
|
|
||||||
|
5. Finally, you can execute `cmdeploy run --ssh-host 13.12.23.42` to turn on chatmail on the new relay.
|
||||||
|
Your users will be able to use the chatmail relay as soon as the DNS changes have propagated.
|
||||||
|
Voilà!
|
||||||
|
|
||||||
|
## Setting up a reverse proxy
|
||||||
|
|
||||||
|
A chatmail relay MTA does not track or depend on the client IP address
|
||||||
|
for its operation, so it can be run behind a reverse proxy.
|
||||||
|
This will not even affect incoming mail authentication
|
||||||
|
as DKIM only checks the cryptographic signature
|
||||||
|
of the message and does not use the IP address as the input.
|
||||||
|
|
||||||
|
For example, you may want to self-host your chatmail relay
|
||||||
|
and only use hosted VPS to provide a public IP address
|
||||||
|
for client connections and incoming mail.
|
||||||
|
You can connect chatmail relay to VPS
|
||||||
|
using a tunnel protocol
|
||||||
|
such as [WireGuard](https://www.wireguard.com/)
|
||||||
|
and setup a reverse proxy on a VPS
|
||||||
|
to forward connections to the chatmail relay
|
||||||
|
over the tunnel.
|
||||||
|
You can also setup multiple reverse proxies
|
||||||
|
for your chatmail relay in different networks
|
||||||
|
to ensure your relay is reachable even when
|
||||||
|
one of the IPs becomes inaccessible due to
|
||||||
|
hosting or routing problems.
|
||||||
|
|
||||||
|
Note that your chatmail relay still needs
|
||||||
|
to be able to make outgoing connections on port 25
|
||||||
|
to send messages outside.
|
||||||
|
|
||||||
|
To setup a reverse proxy
|
||||||
|
(or rather Destination NAT, DNAT)
|
||||||
|
for your chatmail relay,
|
||||||
|
put the following configuration in `/etc/nftables.conf`:
|
||||||
|
```
|
||||||
|
#!/usr/sbin/nft -f
|
||||||
|
|
||||||
|
flush ruleset
|
||||||
|
|
||||||
|
define wan = eth0
|
||||||
|
|
||||||
|
# Which ports to proxy.
|
||||||
|
#
|
||||||
|
# Note that SSH is not proxied
|
||||||
|
# so it is possible to log into the proxy server
|
||||||
|
# and not the original one.
|
||||||
|
define ports = { smtp, http, https, imap, imaps, submission, submissions }
|
||||||
|
|
||||||
|
# The host we want to proxy to.
|
||||||
|
define ipv4_address = AAA.BBB.CCC.DDD
|
||||||
|
define ipv6_address = [XXX::1]
|
||||||
|
|
||||||
|
table ip nat {
|
||||||
|
chain prerouting {
|
||||||
|
type nat hook prerouting priority dstnat; policy accept;
|
||||||
|
iif $wan tcp dport $ports dnat to $ipv4_address
|
||||||
|
}
|
||||||
|
|
||||||
|
chain postrouting {
|
||||||
|
type nat hook postrouting priority 0;
|
||||||
|
|
||||||
|
oifname $wan masquerade
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
table ip6 nat {
|
||||||
|
chain prerouting {
|
||||||
|
type nat hook prerouting priority dstnat; policy accept;
|
||||||
|
iif $wan tcp dport $ports dnat to $ipv6_address
|
||||||
|
}
|
||||||
|
|
||||||
|
chain postrouting {
|
||||||
|
type nat hook postrouting priority 0;
|
||||||
|
|
||||||
|
oifname $wan masquerade
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
table inet filter {
|
||||||
|
chain input {
|
||||||
|
type filter hook input priority filter; policy drop;
|
||||||
|
|
||||||
|
# Accept ICMP.
|
||||||
|
# It is especially important to accept ICMPv6 ND messages,
|
||||||
|
# otherwise IPv6 connectivity breaks.
|
||||||
|
icmp type { echo-request } accept
|
||||||
|
icmpv6 type { echo-request, nd-neighbor-solicit, nd-router-advert, nd-neighbor-advert } accept
|
||||||
|
|
||||||
|
# Allow incoming SSH connections.
|
||||||
|
tcp dport { ssh } accept
|
||||||
|
|
||||||
|
ct state established accept
|
||||||
|
}
|
||||||
|
chain forward {
|
||||||
|
type filter hook forward priority filter; policy drop;
|
||||||
|
|
||||||
|
ct state established accept
|
||||||
|
ip daddr $ipv4_address counter accept
|
||||||
|
ip6 daddr $ipv6_address counter accept
|
||||||
|
}
|
||||||
|
chain output {
|
||||||
|
type filter hook output priority filter;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Run `systemctl enable nftables.service`
|
||||||
|
to ensure configuration is reloaded when the proxy relay reboots.
|
||||||
|
|
||||||
|
Uncomment in `/etc/sysctl.conf` the following two lines:
|
||||||
|
|
||||||
|
```
|
||||||
|
net.ipv4.ip_forward=1
|
||||||
|
net.ipv6.conf.all.forwarding=1
|
||||||
|
```
|
||||||
|
|
||||||
|
Then reboot the relay or do `sysctl -p` and `nft -f /etc/nftables.conf`.
|
||||||
|
|
||||||
|
Once proxy relay is set up,
|
||||||
|
you can add its IP address to the DNS.
|
||||||
|
|||||||
15
RELEASE.md
15
RELEASE.md
@@ -1,15 +0,0 @@
|
|||||||
# Releasing a new version of chatmail relay
|
|
||||||
|
|
||||||
For example, to release version 1.9.0 of chatmail relay, do the following steps.
|
|
||||||
|
|
||||||
1. Update the changelog: `git cliff --unreleased --tag 1.9.0 --prepend CHANGELOG.md` or `git cliff -u -t 1.9.0 -p CHANGELOG.md`.
|
|
||||||
|
|
||||||
2. Open the changelog in the editor, edit it if required.
|
|
||||||
|
|
||||||
3. Commit the changes to the changelog with a commit message `chore(release): prepare for 1.9.0`.
|
|
||||||
|
|
||||||
3. Tag the release: `git tag --annotate 1.9.0`.
|
|
||||||
|
|
||||||
4. Push the release tag: `git push origin 1.9.0`.
|
|
||||||
|
|
||||||
5. Create a GitHub release: `gh release create 1.9.0`.
|
|
||||||
@@ -4,9 +4,12 @@ build-backend = "setuptools.build_meta"
|
|||||||
|
|
||||||
[project]
|
[project]
|
||||||
name = "chatmaild"
|
name = "chatmaild"
|
||||||
version = "0.3"
|
version = "0.2"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
|
"aiosmtpd",
|
||||||
"iniconfig",
|
"iniconfig",
|
||||||
|
"deltachat-rpc-server",
|
||||||
|
"deltachat-rpc-client",
|
||||||
"filelock",
|
"filelock",
|
||||||
"requests",
|
"requests",
|
||||||
"crypt-r >= 3.13.1 ; python_version >= '3.11'",
|
"crypt-r >= 3.13.1 ; python_version >= '3.11'",
|
||||||
@@ -21,11 +24,11 @@ where = ['src']
|
|||||||
[project.scripts]
|
[project.scripts]
|
||||||
doveauth = "chatmaild.doveauth:main"
|
doveauth = "chatmaild.doveauth:main"
|
||||||
chatmail-metadata = "chatmaild.metadata:main"
|
chatmail-metadata = "chatmaild.metadata:main"
|
||||||
chatmail-expire = "chatmaild.expire:daily_expire_main"
|
filtermail = "chatmaild.filtermail:main"
|
||||||
chatmail-quota-expire = "chatmaild.expire:quota_expire_main"
|
echobot = "chatmaild.echo:main"
|
||||||
chatmail-fsreport = "chatmaild.fsreport:main"
|
chatmail-metrics = "chatmaild.metrics:main"
|
||||||
|
delete_inactive_users = "chatmaild.delete_inactive_users:main"
|
||||||
lastlogin = "chatmaild.lastlogin:main"
|
lastlogin = "chatmaild.lastlogin:main"
|
||||||
turnserver = "chatmaild.turnserver:main"
|
|
||||||
|
|
||||||
[project.entry-points.pytest11]
|
[project.entry-points.pytest11]
|
||||||
"chatmaild.testplugin" = "chatmaild.tests.plugin"
|
"chatmaild.testplugin" = "chatmaild.tests.plugin"
|
||||||
@@ -45,9 +48,6 @@ lint.select = [
|
|||||||
"PLE", # Pylint Error
|
"PLE", # Pylint Error
|
||||||
"PLW", # Pylint Warning
|
"PLW", # Pylint Warning
|
||||||
]
|
]
|
||||||
lint.ignore = [
|
|
||||||
"PLC0415" # import-outside-top-level
|
|
||||||
]
|
|
||||||
|
|
||||||
[tool.tox]
|
[tool.tox]
|
||||||
legacy_tox_ini = """
|
legacy_tox_ini = """
|
||||||
@@ -67,8 +67,5 @@ commands =
|
|||||||
[testenv]
|
[testenv]
|
||||||
deps = pytest
|
deps = pytest
|
||||||
pdbpp
|
pdbpp
|
||||||
pytest-localserver
|
|
||||||
aiosmtpd
|
|
||||||
execnet
|
|
||||||
commands = pytest -v -rsXx {posargs}
|
commands = pytest -v -rsXx {posargs}
|
||||||
"""
|
"""
|
||||||
|
|||||||
@@ -4,6 +4,8 @@ import iniconfig
|
|||||||
|
|
||||||
from chatmaild.user import User
|
from chatmaild.user import User
|
||||||
|
|
||||||
|
echobot_password_path = Path("/run/echobot/password")
|
||||||
|
|
||||||
|
|
||||||
def read_config(inipath):
|
def read_config(inipath):
|
||||||
assert Path(inipath).exists(), inipath
|
assert Path(inipath).exists(), inipath
|
||||||
@@ -20,38 +22,27 @@ class Config:
|
|||||||
def __init__(self, inipath, params):
|
def __init__(self, inipath, params):
|
||||||
self._inipath = inipath
|
self._inipath = inipath
|
||||||
self.mail_domain = params["mail_domain"]
|
self.mail_domain = params["mail_domain"]
|
||||||
self.max_user_send_per_minute = int(params.get("max_user_send_per_minute", 60))
|
self.max_user_send_per_minute = int(params["max_user_send_per_minute"])
|
||||||
self.max_user_send_burst_size = int(params.get("max_user_send_burst_size", 10))
|
|
||||||
self.max_mailbox_size = params["max_mailbox_size"]
|
self.max_mailbox_size = params["max_mailbox_size"]
|
||||||
self.max_message_size = int(params.get("max_message_size", "31457280"))
|
self.max_message_size = int(params.get("max_message_size", "31457280"))
|
||||||
self.delete_mails_after = params["delete_mails_after"]
|
self.delete_mails_after = params["delete_mails_after"]
|
||||||
self.delete_large_after = params["delete_large_after"]
|
|
||||||
self.delete_inactive_users_after = int(params["delete_inactive_users_after"])
|
self.delete_inactive_users_after = int(params["delete_inactive_users_after"])
|
||||||
self.username_min_length = int(params["username_min_length"])
|
self.username_min_length = int(params["username_min_length"])
|
||||||
self.username_max_length = int(params["username_max_length"])
|
self.username_max_length = int(params["username_max_length"])
|
||||||
self.password_min_length = int(params["password_min_length"])
|
self.password_min_length = int(params["password_min_length"])
|
||||||
self.passthrough_senders = params["passthrough_senders"].split()
|
self.passthrough_senders = params["passthrough_senders"].split()
|
||||||
self.passthrough_recipients = params["passthrough_recipients"].split()
|
self.passthrough_recipients = params["passthrough_recipients"].split()
|
||||||
self.www_folder = params.get("www_folder", "")
|
self.filtermail_smtp_port = int(params["filtermail_smtp_port"])
|
||||||
self.filtermail_smtp_port = int(params.get("filtermail_smtp_port", "10080"))
|
|
||||||
self.filtermail_smtp_port_incoming = int(
|
self.filtermail_smtp_port_incoming = int(
|
||||||
params.get("filtermail_smtp_port_incoming", "10081")
|
params["filtermail_smtp_port_incoming"]
|
||||||
)
|
)
|
||||||
self.filtermail_http_port_incoming = int(
|
self.postfix_reinject_port = int(params["postfix_reinject_port"])
|
||||||
params.get("filtermail_http_port_incoming", "10082")
|
|
||||||
)
|
|
||||||
self.filtermail_lmtp_port_transport = int(
|
|
||||||
params.get("filtermail_lmtp_port_transport", "10083")
|
|
||||||
)
|
|
||||||
self.postfix_reinject_port = int(params.get("postfix_reinject_port", "10025"))
|
|
||||||
self.postfix_reinject_port_incoming = int(
|
self.postfix_reinject_port_incoming = int(
|
||||||
params.get("postfix_reinject_port_incoming", "10026")
|
params["postfix_reinject_port_incoming"]
|
||||||
)
|
)
|
||||||
self.mtail_address = params.get("mtail_address")
|
self.mtail_address = params.get("mtail_address")
|
||||||
self.disable_ipv6 = params.get("disable_ipv6", "false").lower() == "true"
|
self.disable_ipv6 = params.get("disable_ipv6", "false").lower() == "true"
|
||||||
self.acme_email = params.get("acme_email", "")
|
|
||||||
self.imap_rawlog = params.get("imap_rawlog", "false").lower() == "true"
|
self.imap_rawlog = params.get("imap_rawlog", "false").lower() == "true"
|
||||||
self.imap_compress = params.get("imap_compress", "false").lower() == "true"
|
|
||||||
if "iroh_relay" not in params:
|
if "iroh_relay" not in params:
|
||||||
self.iroh_relay = "https://" + params["mail_domain"]
|
self.iroh_relay = "https://" + params["mail_domain"]
|
||||||
self.enable_iroh_relay = True
|
self.enable_iroh_relay = True
|
||||||
@@ -63,31 +54,6 @@ class Config:
|
|||||||
self.privacy_pdo = params.get("privacy_pdo")
|
self.privacy_pdo = params.get("privacy_pdo")
|
||||||
self.privacy_supervisor = params.get("privacy_supervisor")
|
self.privacy_supervisor = params.get("privacy_supervisor")
|
||||||
|
|
||||||
# TLS certificate management.
|
|
||||||
# If tls_external_cert_and_key is set, use externally managed certs.
|
|
||||||
# Otherwise derived from the domain name:
|
|
||||||
# - Domains starting with "_" use self-signed certificates
|
|
||||||
# - All other domains use ACME.
|
|
||||||
external = params.get("tls_external_cert_and_key", "").strip()
|
|
||||||
|
|
||||||
if external:
|
|
||||||
parts = external.split()
|
|
||||||
if len(parts) != 2:
|
|
||||||
raise ValueError(
|
|
||||||
"tls_external_cert_and_key must have two space-separated"
|
|
||||||
" paths: CERT_PATH KEY_PATH"
|
|
||||||
)
|
|
||||||
self.tls_cert_mode = "external"
|
|
||||||
self.tls_cert_path, self.tls_key_path = parts
|
|
||||||
elif self.mail_domain.startswith("_"):
|
|
||||||
self.tls_cert_mode = "self"
|
|
||||||
self.tls_cert_path = "/etc/ssl/certs/mailserver.pem"
|
|
||||||
self.tls_key_path = "/etc/ssl/private/mailserver.key"
|
|
||||||
else:
|
|
||||||
self.tls_cert_mode = "acme"
|
|
||||||
self.tls_cert_path = f"/var/lib/acme/live/{self.mail_domain}/fullchain"
|
|
||||||
self.tls_key_path = f"/var/lib/acme/live/{self.mail_domain}/privkey"
|
|
||||||
|
|
||||||
# deprecated option
|
# deprecated option
|
||||||
mbdir = params.get("mailboxes_dir", f"/home/vmail/mail/{self.mail_domain}")
|
mbdir = params.get("mailboxes_dir", f"/home/vmail/mail/{self.mail_domain}")
|
||||||
self.mailboxes_dir = Path(mbdir.strip())
|
self.mailboxes_dir = Path(mbdir.strip())
|
||||||
@@ -95,34 +61,22 @@ class Config:
|
|||||||
# old unused option (except for first migration from sqlite to maildir store)
|
# old unused option (except for first migration from sqlite to maildir store)
|
||||||
self.passdb_path = Path(params.get("passdb_path", "/home/vmail/passdb.sqlite"))
|
self.passdb_path = Path(params.get("passdb_path", "/home/vmail/passdb.sqlite"))
|
||||||
|
|
||||||
@property
|
|
||||||
def max_mailbox_size_mb(self):
|
|
||||||
"""Return max_mailbox_size as an integer in megabytes."""
|
|
||||||
return parse_size_mb(self.max_mailbox_size)
|
|
||||||
|
|
||||||
def _getbytefile(self):
|
def _getbytefile(self):
|
||||||
return open(self._inipath, "rb")
|
return open(self._inipath, "rb")
|
||||||
|
|
||||||
def get_user(self, addr) -> User:
|
def get_user(self, addr):
|
||||||
if not addr or "@" not in addr or "/" in addr:
|
if not addr or "@" not in addr or "/" in addr:
|
||||||
raise ValueError(f"invalid address {addr!r}")
|
raise ValueError(f"invalid address {addr!r}")
|
||||||
|
|
||||||
maildir = self.mailboxes_dir.joinpath(addr)
|
maildir = self.mailboxes_dir.joinpath(addr)
|
||||||
|
if addr.startswith("echo@"):
|
||||||
|
password_path = echobot_password_path
|
||||||
|
else:
|
||||||
password_path = maildir.joinpath("password")
|
password_path = maildir.joinpath("password")
|
||||||
|
|
||||||
return User(maildir, addr, password_path, uid="vmail", gid="vmail")
|
return User(maildir, addr, password_path, uid="vmail", gid="vmail")
|
||||||
|
|
||||||
|
|
||||||
def parse_size_mb(limit):
|
|
||||||
"""Parse a size string like ``500M`` or ``2G`` and return megabytes."""
|
|
||||||
value = limit.strip().upper().removesuffix("B")
|
|
||||||
if value.endswith("G"):
|
|
||||||
return int(value[:-1]) * 1024
|
|
||||||
if value.endswith("M"):
|
|
||||||
return int(value[:-1])
|
|
||||||
return int(value)
|
|
||||||
|
|
||||||
|
|
||||||
def write_initial_config(inipath, mail_domain, overrides):
|
def write_initial_config(inipath, mail_domain, overrides):
|
||||||
"""Write out default config file, using the specified config value overrides."""
|
"""Write out default config file, using the specified config value overrides."""
|
||||||
content = get_default_config_content(mail_domain, **overrides)
|
content = get_default_config_content(mail_domain, **overrides)
|
||||||
@@ -161,7 +115,7 @@ def get_default_config_content(mail_domain, **overrides):
|
|||||||
lines = []
|
lines = []
|
||||||
for line in content.split("\n"):
|
for line in content.split("\n"):
|
||||||
for key, value in privacy.items():
|
for key, value in privacy.items():
|
||||||
value_lines = value.format(mail_domain=mail_domain).strip().split("\n")
|
value_lines = value.strip().split("\n")
|
||||||
if not line.startswith(f"{key} =") or not value_lines:
|
if not line.startswith(f"{key} =") or not value_lines:
|
||||||
continue
|
continue
|
||||||
if len(value_lines) == 1:
|
if len(value_lines) == 1:
|
||||||
|
|||||||
31
chatmaild/src/chatmaild/delete_inactive_users.py
Normal file
31
chatmaild/src/chatmaild/delete_inactive_users.py
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
"""
|
||||||
|
Remove inactive users
|
||||||
|
"""
|
||||||
|
|
||||||
|
import os
|
||||||
|
import shutil
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
|
|
||||||
|
from .config import read_config
|
||||||
|
|
||||||
|
|
||||||
|
def delete_inactive_users(config):
|
||||||
|
cutoff_date = time.time() - config.delete_inactive_users_after * 86400
|
||||||
|
for addr in os.listdir(config.mailboxes_dir):
|
||||||
|
try:
|
||||||
|
user = config.get_user(addr)
|
||||||
|
except ValueError:
|
||||||
|
continue
|
||||||
|
|
||||||
|
read_timestamp = user.get_last_login_timestamp()
|
||||||
|
if read_timestamp and read_timestamp < cutoff_date:
|
||||||
|
path = config.mailboxes_dir.joinpath(addr)
|
||||||
|
assert path == user.maildir
|
||||||
|
shutil.rmtree(path, ignore_errors=True)
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
(cfgpath,) = sys.argv[1:]
|
||||||
|
config = read_config(cfgpath)
|
||||||
|
delete_inactive_users(config)
|
||||||
@@ -22,7 +22,7 @@ class DictProxy:
|
|||||||
wfile.flush()
|
wfile.flush()
|
||||||
|
|
||||||
def handle_dovecot_request(self, msg, transactions):
|
def handle_dovecot_request(self, msg, transactions):
|
||||||
# see https://doc.dovecot.org/2.3/developer_manual/design/dict_protocol/#dovecot-dict-protocol
|
# see https://doc.dovecot.org/developer_manual/design/dict_protocol/#dovecot-dict-protocol
|
||||||
short_command = msg[0]
|
short_command = msg[0]
|
||||||
parts = msg[1:].split("\t")
|
parts = msg[1:].split("\t")
|
||||||
|
|
||||||
|
|||||||
@@ -1,11 +1,8 @@
|
|||||||
import json
|
import json
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import re
|
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
import filelock
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import crypt_r
|
import crypt_r
|
||||||
except ImportError:
|
except ImportError:
|
||||||
@@ -16,11 +13,10 @@ from .dictproxy import DictProxy
|
|||||||
from .migrate_db import migrate_from_db_to_maildir
|
from .migrate_db import migrate_from_db_to_maildir
|
||||||
|
|
||||||
NOCREATE_FILE = "/etc/chatmail-nocreate"
|
NOCREATE_FILE = "/etc/chatmail-nocreate"
|
||||||
VALID_LOCALPART_RE = re.compile(r"^[a-z0-9._-]+$")
|
|
||||||
|
|
||||||
|
|
||||||
def encrypt_password(password: str):
|
def encrypt_password(password: str):
|
||||||
# https://doc.dovecot.org/2.3/configuration_manual/authentication/password_schemes/
|
# https://doc.dovecot.org/configuration_manual/authentication/password_schemes/
|
||||||
passhash = crypt_r.crypt(password, crypt_r.METHOD_SHA512)
|
passhash = crypt_r.crypt(password, crypt_r.METHOD_SHA512)
|
||||||
return "{SHA512-CRYPT}" + passhash
|
return "{SHA512-CRYPT}" + passhash
|
||||||
|
|
||||||
@@ -44,6 +40,10 @@ def is_allowed_to_create(config: Config, user, cleartext_password) -> bool:
|
|||||||
return False
|
return False
|
||||||
localpart, domain = parts
|
localpart, domain = parts
|
||||||
|
|
||||||
|
if localpart == "echo":
|
||||||
|
# echobot account should not be created in the database
|
||||||
|
return False
|
||||||
|
|
||||||
if (
|
if (
|
||||||
len(localpart) > config.username_max_length
|
len(localpart) > config.username_max_length
|
||||||
or len(localpart) < config.username_min_length
|
or len(localpart) < config.username_min_length
|
||||||
@@ -56,10 +56,6 @@ def is_allowed_to_create(config: Config, user, cleartext_password) -> bool:
|
|||||||
)
|
)
|
||||||
return False
|
return False
|
||||||
|
|
||||||
if not VALID_LOCALPART_RE.match(localpart):
|
|
||||||
logging.warning("localpart %r contains invalid characters", localpart)
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
@@ -148,11 +144,6 @@ class AuthDictProxy(DictProxy):
|
|||||||
if not is_allowed_to_create(self.config, addr, cleartext_password):
|
if not is_allowed_to_create(self.config, addr, cleartext_password):
|
||||||
return
|
return
|
||||||
|
|
||||||
lock = filelock.FileLock(str(user.password_path) + ".lock", timeout=5)
|
|
||||||
with lock:
|
|
||||||
userdata = user.get_userdb_dict()
|
|
||||||
if userdata:
|
|
||||||
return userdata
|
|
||||||
user.set_password(encrypt_password(cleartext_password))
|
user.set_password(encrypt_password(cleartext_password))
|
||||||
print(f"Created address: {addr}", file=sys.stderr)
|
print(f"Created address: {addr}", file=sys.stderr)
|
||||||
return user.get_userdb_dict()
|
return user.get_userdb_dict()
|
||||||
|
|||||||
109
chatmaild/src/chatmaild/echo.py
Normal file
109
chatmaild/src/chatmaild/echo.py
Normal file
@@ -0,0 +1,109 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
"""Advanced echo bot example.
|
||||||
|
|
||||||
|
it will echo back any message that has non-empty text and also supports the /help command.
|
||||||
|
"""
|
||||||
|
|
||||||
|
import logging
|
||||||
|
import os
|
||||||
|
import subprocess
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
from deltachat_rpc_client import Bot, DeltaChat, EventType, Rpc, events
|
||||||
|
|
||||||
|
from chatmaild.config import echobot_password_path, read_config
|
||||||
|
from chatmaild.doveauth import encrypt_password
|
||||||
|
from chatmaild.newemail import create_newemail_dict
|
||||||
|
|
||||||
|
hooks = events.HookCollection()
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.RawEvent)
|
||||||
|
def log_event(event):
|
||||||
|
if event.kind == EventType.INFO:
|
||||||
|
logging.info(event.msg)
|
||||||
|
elif event.kind == EventType.WARNING:
|
||||||
|
logging.warning(event.msg)
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.RawEvent(EventType.ERROR))
|
||||||
|
def log_error(event):
|
||||||
|
logging.error("%s", event.msg)
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.MemberListChanged)
|
||||||
|
def on_memberlist_changed(event):
|
||||||
|
logging.info(
|
||||||
|
"member %s was %s", event.member, "added" if event.member_added else "removed"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.GroupImageChanged)
|
||||||
|
def on_group_image_changed(event):
|
||||||
|
logging.info("group image %s", "deleted" if event.image_deleted else "changed")
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.GroupNameChanged)
|
||||||
|
def on_group_name_changed(event):
|
||||||
|
logging.info(f"group name changed, old name: {event.old_name}")
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.NewMessage(func=lambda e: not e.command))
|
||||||
|
def echo(event):
|
||||||
|
snapshot = event.message_snapshot
|
||||||
|
if snapshot.is_info:
|
||||||
|
# Ignore info messages
|
||||||
|
return
|
||||||
|
if snapshot.text or snapshot.file:
|
||||||
|
snapshot.chat.send_message(text=snapshot.text, file=snapshot.file)
|
||||||
|
|
||||||
|
|
||||||
|
@hooks.on(events.NewMessage(command="/help"))
|
||||||
|
def help_command(event):
|
||||||
|
snapshot = event.message_snapshot
|
||||||
|
snapshot.chat.send_text("Send me any message and I will echo it back")
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
logging.basicConfig(level=logging.INFO)
|
||||||
|
path = os.environ.get("PATH")
|
||||||
|
venv_path = sys.argv[0].strip("echobot")
|
||||||
|
os.environ["PATH"] = path + ":" + venv_path
|
||||||
|
with Rpc() as rpc:
|
||||||
|
deltachat = DeltaChat(rpc)
|
||||||
|
system_info = deltachat.get_system_info()
|
||||||
|
logging.info(f"Running deltachat core {system_info.deltachat_core_version}")
|
||||||
|
|
||||||
|
accounts = deltachat.get_all_accounts()
|
||||||
|
account = accounts[0] if accounts else deltachat.add_account()
|
||||||
|
|
||||||
|
bot = Bot(account, hooks)
|
||||||
|
|
||||||
|
config = read_config(sys.argv[1])
|
||||||
|
addr = "echo@" + config.mail_domain
|
||||||
|
|
||||||
|
# Create password file
|
||||||
|
if bot.is_configured():
|
||||||
|
password = bot.account.get_config("mail_pw")
|
||||||
|
else:
|
||||||
|
password = create_newemail_dict(config)["password"]
|
||||||
|
|
||||||
|
echobot_password_path.write_text(encrypt_password(password))
|
||||||
|
# Give the user which doveauth runs as access to the password file.
|
||||||
|
subprocess.check_call(
|
||||||
|
["/usr/bin/setfacl", "-m", "user:vmail:r", echobot_password_path],
|
||||||
|
)
|
||||||
|
|
||||||
|
if not bot.is_configured():
|
||||||
|
bot.configure(addr, password)
|
||||||
|
|
||||||
|
# write invite link to working directory
|
||||||
|
invitelink = bot.account.get_qr_code()
|
||||||
|
Path("invite-link.txt").write_text(invitelink)
|
||||||
|
|
||||||
|
bot.run_forever()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
@@ -1,292 +0,0 @@
|
|||||||
"""
|
|
||||||
Expire old messages and addresses.
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
import shutil
|
|
||||||
import sys
|
|
||||||
import time
|
|
||||||
from argparse import ArgumentParser
|
|
||||||
from collections import namedtuple
|
|
||||||
from datetime import datetime
|
|
||||||
from pathlib import Path
|
|
||||||
from stat import S_ISREG
|
|
||||||
|
|
||||||
from chatmaild.config import read_config
|
|
||||||
|
|
||||||
FileEntry = namedtuple("FileEntry", ("path", "mtime", "size"))
|
|
||||||
QuotaFileEntry = namedtuple("QuotaFileEntry", ("mtime", "quota_size", "path"))
|
|
||||||
|
|
||||||
# Quota cleanup factor of max_mailbox_size. The mailbox is reset to this size.
|
|
||||||
QUOTA_CLEANUP_FACTOR = 0.7
|
|
||||||
|
|
||||||
# e.g. "cur/1775324677.M448978P3029757.exam,S=3235,W=3305:2,S"
|
|
||||||
_dovecot_fn_rex = re.compile(r".+/(\d+)\..+,S=(\d+)")
|
|
||||||
|
|
||||||
|
|
||||||
def iter_mailboxes(basedir, maxnum):
|
|
||||||
if not os.path.exists(basedir):
|
|
||||||
print_info(f"no mailboxes found at: {basedir}")
|
|
||||||
return
|
|
||||||
|
|
||||||
for name in os_listdir_if_exists(basedir)[:maxnum]:
|
|
||||||
if "@" in name:
|
|
||||||
yield MailboxStat(basedir + "/" + name)
|
|
||||||
|
|
||||||
|
|
||||||
def get_file_entry(path):
|
|
||||||
"""return a FileEntry or None if the path does not exist or is not a regular file."""
|
|
||||||
try:
|
|
||||||
st = os.stat(path)
|
|
||||||
except FileNotFoundError:
|
|
||||||
return None
|
|
||||||
if not S_ISREG(st.st_mode):
|
|
||||||
return None
|
|
||||||
return FileEntry(path, st.st_mtime, st.st_size)
|
|
||||||
|
|
||||||
|
|
||||||
def os_listdir_if_exists(path):
|
|
||||||
"""return a list of names obtained from os.listdir or an empty list if the path does not exist."""
|
|
||||||
try:
|
|
||||||
return os.listdir(path)
|
|
||||||
except FileNotFoundError:
|
|
||||||
return []
|
|
||||||
|
|
||||||
|
|
||||||
class MailboxStat:
|
|
||||||
last_login = None
|
|
||||||
|
|
||||||
def __init__(self, basedir):
|
|
||||||
self.basedir = str(basedir)
|
|
||||||
self.messages = []
|
|
||||||
self.extrafiles = []
|
|
||||||
self.scandir(self.basedir)
|
|
||||||
|
|
||||||
def scandir(self, folderdir):
|
|
||||||
for name in os_listdir_if_exists(folderdir):
|
|
||||||
path = f"{folderdir}/{name}"
|
|
||||||
if name in ("cur", "new", "tmp"):
|
|
||||||
for msg_name in os_listdir_if_exists(path):
|
|
||||||
entry = get_file_entry(f"{path}/{msg_name}")
|
|
||||||
if entry is not None:
|
|
||||||
self.messages.append(entry)
|
|
||||||
elif os.path.isdir(path):
|
|
||||||
self.scandir(path)
|
|
||||||
else:
|
|
||||||
entry = get_file_entry(path)
|
|
||||||
if entry is not None:
|
|
||||||
self.extrafiles.append(entry)
|
|
||||||
if name == "password":
|
|
||||||
self.last_login = entry.mtime
|
|
||||||
self.extrafiles.sort(key=lambda x: -x.size)
|
|
||||||
|
|
||||||
|
|
||||||
def parse_dovecot_filename(relpath):
|
|
||||||
m = _dovecot_fn_rex.match(relpath)
|
|
||||||
if not m:
|
|
||||||
return None
|
|
||||||
return QuotaFileEntry(int(m.group(1)), int(m.group(2)), relpath)
|
|
||||||
|
|
||||||
|
|
||||||
def scan_mailbox_messages(mbox):
|
|
||||||
messages = []
|
|
||||||
for sub in ("cur", "new"):
|
|
||||||
for name in os_listdir_if_exists(mbox / sub):
|
|
||||||
if entry := parse_dovecot_filename(f"{sub}/{name}"):
|
|
||||||
messages.append(entry)
|
|
||||||
return messages
|
|
||||||
|
|
||||||
|
|
||||||
def expire_to_target(mbox, target_bytes):
|
|
||||||
messages = scan_mailbox_messages(mbox)
|
|
||||||
total_size = sum(m.quota_size for m in messages)
|
|
||||||
# Keep recent 24 hours of messages protected from expiry because
|
|
||||||
# likely something is wrong with interactions on that address
|
|
||||||
# and quota-full signal can help the address owner's device to notice it
|
|
||||||
undeletable_messages_cutoff = time.time() - (3600 * 24)
|
|
||||||
removed = 0
|
|
||||||
for entry in sorted(messages):
|
|
||||||
if total_size <= target_bytes:
|
|
||||||
break
|
|
||||||
if entry.mtime > undeletable_messages_cutoff:
|
|
||||||
break
|
|
||||||
(mbox / entry.path).unlink(missing_ok=True)
|
|
||||||
total_size -= entry.quota_size
|
|
||||||
removed += 1
|
|
||||||
|
|
||||||
return removed
|
|
||||||
|
|
||||||
|
|
||||||
def print_info(msg):
|
|
||||||
print(msg, file=sys.stderr)
|
|
||||||
|
|
||||||
|
|
||||||
class Expiry:
|
|
||||||
def __init__(self, config, dry, now, verbose):
|
|
||||||
self.config = config
|
|
||||||
self.dry = dry
|
|
||||||
self.now = now
|
|
||||||
self.verbose = verbose
|
|
||||||
self.del_mboxes = 0
|
|
||||||
self.all_mboxes = 0
|
|
||||||
self.del_files = 0
|
|
||||||
self.all_files = 0
|
|
||||||
self.start = time.time()
|
|
||||||
|
|
||||||
def remove_mailbox(self, mboxdir):
|
|
||||||
if self.verbose:
|
|
||||||
print_info(f"removing {mboxdir}")
|
|
||||||
if not self.dry:
|
|
||||||
shutil.rmtree(mboxdir)
|
|
||||||
self.del_mboxes += 1
|
|
||||||
|
|
||||||
def remove_file(self, path, mtime=None):
|
|
||||||
if self.verbose:
|
|
||||||
if mtime is not None:
|
|
||||||
date = datetime.fromtimestamp(mtime).strftime("%b %d")
|
|
||||||
print_info(f"removing {date} {path}")
|
|
||||||
else:
|
|
||||||
print_info(f"removing {path}")
|
|
||||||
if not self.dry:
|
|
||||||
try:
|
|
||||||
os.unlink(path)
|
|
||||||
except FileNotFoundError:
|
|
||||||
print_info(f"file not found/vanished {path}")
|
|
||||||
self.del_files += 1
|
|
||||||
|
|
||||||
def process_mailbox_stat(self, mbox):
|
|
||||||
cutoff_without_login = (
|
|
||||||
self.now - int(self.config.delete_inactive_users_after) * 86400
|
|
||||||
)
|
|
||||||
cutoff_mails = self.now - int(self.config.delete_mails_after) * 86400
|
|
||||||
cutoff_large_mails = self.now - int(self.config.delete_large_after) * 86400
|
|
||||||
|
|
||||||
self.all_mboxes += 1
|
|
||||||
changed = False
|
|
||||||
if mbox.last_login and mbox.last_login < cutoff_without_login:
|
|
||||||
self.remove_mailbox(mbox.basedir)
|
|
||||||
return
|
|
||||||
|
|
||||||
mboxname = os.path.basename(mbox.basedir)
|
|
||||||
if self.verbose:
|
|
||||||
date = datetime.fromtimestamp(mbox.last_login) if mbox.last_login else None
|
|
||||||
if date:
|
|
||||||
print_info(f"checking mailbox {date.strftime('%b %d')} {mboxname}")
|
|
||||||
else:
|
|
||||||
print_info(f"checking mailbox (no last_login) {mboxname}")
|
|
||||||
self.all_files += len(mbox.messages)
|
|
||||||
for message in mbox.messages:
|
|
||||||
if message.mtime < cutoff_mails:
|
|
||||||
self.remove_file(message.path, mtime=message.mtime)
|
|
||||||
elif message.size > 200000 and message.mtime < cutoff_large_mails:
|
|
||||||
# we only remove noticed large files (not unnoticed ones in new/)
|
|
||||||
parts = message.path.split("/")
|
|
||||||
if len(parts) >= 2 and parts[-2] == "cur":
|
|
||||||
self.remove_file(message.path, mtime=message.mtime)
|
|
||||||
else:
|
|
||||||
continue
|
|
||||||
changed = True
|
|
||||||
|
|
||||||
target_bytes = (
|
|
||||||
self.config.max_mailbox_size_mb * 1024 * 1024 * QUOTA_CLEANUP_FACTOR
|
|
||||||
)
|
|
||||||
removed = expire_to_target(Path(mbox.basedir), target_bytes)
|
|
||||||
if removed:
|
|
||||||
changed = True
|
|
||||||
self.del_files += removed
|
|
||||||
if self.verbose:
|
|
||||||
print_info(
|
|
||||||
f"quota-expire: removed {removed} message(s) from {mboxname}"
|
|
||||||
)
|
|
||||||
|
|
||||||
if changed:
|
|
||||||
self.remove_file(f"{mbox.basedir}/maildirsize")
|
|
||||||
|
|
||||||
def get_summary(self):
|
|
||||||
return (
|
|
||||||
f"Removed {self.del_mboxes} out of {self.all_mboxes} mailboxes "
|
|
||||||
f"and {self.del_files} out of {self.all_files} files in existing mailboxes "
|
|
||||||
f"in {time.time() - self.start:2.2f} seconds"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def daily_expire_main(args=None):
|
|
||||||
"""Expire mailboxes and messages according to chatmail config"""
|
|
||||||
parser = ArgumentParser(description=daily_expire_main.__doc__)
|
|
||||||
ini = "/usr/local/lib/chatmaild/chatmail.ini"
|
|
||||||
parser.add_argument(
|
|
||||||
"chatmail_ini",
|
|
||||||
action="store",
|
|
||||||
nargs="?",
|
|
||||||
help=f"path pointing to chatmail.ini file, default: {ini}",
|
|
||||||
default=ini,
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--days", action="store", help="assume date to be days older than now"
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
"--maxnum",
|
|
||||||
default=None,
|
|
||||||
action="store",
|
|
||||||
help="maximum number of mailboxes to iterate on",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"-v",
|
|
||||||
dest="verbose",
|
|
||||||
action="store_true",
|
|
||||||
help="print out removed files and mailboxes",
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
"--remove",
|
|
||||||
dest="remove",
|
|
||||||
action="store_true",
|
|
||||||
help="actually remove all expired files and dirs",
|
|
||||||
)
|
|
||||||
args = parser.parse_args(args)
|
|
||||||
|
|
||||||
config = read_config(args.chatmail_ini)
|
|
||||||
now = datetime.utcnow().timestamp()
|
|
||||||
if args.days:
|
|
||||||
now = now - 86400 * int(args.days)
|
|
||||||
|
|
||||||
maxnum = int(args.maxnum) if args.maxnum else None
|
|
||||||
exp = Expiry(config, dry=not args.remove, now=now, verbose=args.verbose)
|
|
||||||
for mailbox in iter_mailboxes(str(config.mailboxes_dir), maxnum=maxnum):
|
|
||||||
exp.process_mailbox_stat(mailbox)
|
|
||||||
print(exp.get_summary())
|
|
||||||
|
|
||||||
|
|
||||||
def quota_expire_main(args=None):
|
|
||||||
"""Remove mailbox messages to stay within a megabyte target.
|
|
||||||
|
|
||||||
This entry point is called by dovecot when a quota threshold is passed.
|
|
||||||
"""
|
|
||||||
|
|
||||||
parser = ArgumentParser(description=quota_expire_main.__doc__)
|
|
||||||
parser.add_argument(
|
|
||||||
"target_mb",
|
|
||||||
type=int,
|
|
||||||
help="target mailbox size in megabytes",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"mailbox_path",
|
|
||||||
type=Path,
|
|
||||||
help="path to a user mailbox",
|
|
||||||
)
|
|
||||||
args = parser.parse_args(args)
|
|
||||||
|
|
||||||
target_bytes = args.target_mb * 1024 * 1024
|
|
||||||
|
|
||||||
removed_count = expire_to_target(args.mailbox_path, target_bytes)
|
|
||||||
if removed_count:
|
|
||||||
(args.mailbox_path / "maildirsize").unlink(missing_ok=True)
|
|
||||||
print(
|
|
||||||
f"quota-expire: removed {removed_count} message(s)"
|
|
||||||
f" from {args.mailbox_path.name}",
|
|
||||||
file=sys.stderr,
|
|
||||||
)
|
|
||||||
return 0
|
|
||||||
342
chatmaild/src/chatmaild/filtermail.py
Normal file
342
chatmaild/src/chatmaild/filtermail.py
Normal file
@@ -0,0 +1,342 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
import asyncio
|
||||||
|
import base64
|
||||||
|
import binascii
|
||||||
|
import logging
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
|
from email import policy
|
||||||
|
from email.parser import BytesParser
|
||||||
|
from email.utils import parseaddr
|
||||||
|
from smtplib import SMTP as SMTPClient
|
||||||
|
|
||||||
|
from aiosmtpd.controller import Controller
|
||||||
|
from aiosmtpd.smtp import SMTP
|
||||||
|
|
||||||
|
from .config import read_config
|
||||||
|
|
||||||
|
ENCRYPTION_NEEDED_523 = "523 Encryption Needed: Invalid Unencrypted Mail"
|
||||||
|
|
||||||
|
|
||||||
|
def check_openpgp_payload(payload: bytes):
|
||||||
|
"""Checks the OpenPGP payload.
|
||||||
|
|
||||||
|
OpenPGP payload must consist only of PKESK and SKESK packets
|
||||||
|
terminated by a single SEIPD packet.
|
||||||
|
|
||||||
|
Returns True if OpenPGP payload is correct,
|
||||||
|
False otherwise.
|
||||||
|
|
||||||
|
May raise IndexError while trying to read OpenPGP packet header
|
||||||
|
if it is truncated.
|
||||||
|
"""
|
||||||
|
i = 0
|
||||||
|
while i < len(payload):
|
||||||
|
# Only OpenPGP format is allowed.
|
||||||
|
if payload[i] & 0xC0 != 0xC0:
|
||||||
|
return False
|
||||||
|
|
||||||
|
packet_type_id = payload[i] & 0x3F
|
||||||
|
i += 1
|
||||||
|
if payload[i] < 192:
|
||||||
|
# One-octet length.
|
||||||
|
body_len = payload[i]
|
||||||
|
i += 1
|
||||||
|
elif payload[i] < 224:
|
||||||
|
# Two-octet length.
|
||||||
|
body_len = ((payload[i] - 192) << 8) + payload[i + 1] + 192
|
||||||
|
i += 2
|
||||||
|
elif payload[i] == 255:
|
||||||
|
# Five-octet length.
|
||||||
|
body_len = (
|
||||||
|
(payload[i + 1] << 24)
|
||||||
|
| (payload[i + 2] << 16)
|
||||||
|
| (payload[i + 3] << 8)
|
||||||
|
| payload[i + 4]
|
||||||
|
)
|
||||||
|
i += 5
|
||||||
|
else:
|
||||||
|
# Partial body length is not allowed.
|
||||||
|
return False
|
||||||
|
|
||||||
|
i += body_len
|
||||||
|
|
||||||
|
if i == len(payload):
|
||||||
|
# Last packet should be
|
||||||
|
# Symmetrically Encrypted and Integrity Protected Data Packet (SEIPD)
|
||||||
|
#
|
||||||
|
# This is the only place where this function may return `True`.
|
||||||
|
return packet_type_id == 18
|
||||||
|
elif packet_type_id not in [1, 3]:
|
||||||
|
# All packets except the last one must be either
|
||||||
|
# Public-Key Encrypted Session Key Packet (PKESK)
|
||||||
|
# or
|
||||||
|
# Symmetric-Key Encrypted Session Key Packet (SKESK)
|
||||||
|
return False
|
||||||
|
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def check_armored_payload(payload: str):
|
||||||
|
prefix = "-----BEGIN PGP MESSAGE-----\r\n\r\n"
|
||||||
|
if not payload.startswith(prefix):
|
||||||
|
return False
|
||||||
|
payload = payload.removeprefix(prefix)
|
||||||
|
|
||||||
|
while payload.endswith("\r\n"):
|
||||||
|
payload = payload.removesuffix("\r\n")
|
||||||
|
suffix = "-----END PGP MESSAGE-----"
|
||||||
|
if not payload.endswith(suffix):
|
||||||
|
return False
|
||||||
|
payload = payload.removesuffix(suffix)
|
||||||
|
|
||||||
|
# Remove CRC24.
|
||||||
|
payload = payload.rpartition("=")[0]
|
||||||
|
|
||||||
|
try:
|
||||||
|
payload = base64.b64decode(payload)
|
||||||
|
except binascii.Error:
|
||||||
|
return False
|
||||||
|
|
||||||
|
try:
|
||||||
|
return check_openpgp_payload(payload)
|
||||||
|
except IndexError:
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def is_securejoin(message):
|
||||||
|
if message.get("secure-join") not in ["vc-request", "vg-request"]:
|
||||||
|
return False
|
||||||
|
if not message.is_multipart():
|
||||||
|
return False
|
||||||
|
parts_count = 0
|
||||||
|
for part in message.iter_parts():
|
||||||
|
parts_count += 1
|
||||||
|
if parts_count > 1:
|
||||||
|
return False
|
||||||
|
if part.is_multipart():
|
||||||
|
return False
|
||||||
|
if part.get_content_type() != "text/plain":
|
||||||
|
return False
|
||||||
|
|
||||||
|
payload = part.get_payload().strip().lower()
|
||||||
|
if payload not in ("secure-join: vc-request", "secure-join: vg-request"):
|
||||||
|
return False
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
def check_encrypted(message):
|
||||||
|
"""Check that the message is an OpenPGP-encrypted message.
|
||||||
|
|
||||||
|
MIME structure of the message must correspond to <https://www.rfc-editor.org/rfc/rfc3156>.
|
||||||
|
"""
|
||||||
|
if not message.is_multipart():
|
||||||
|
return False
|
||||||
|
if message.get_content_type() != "multipart/encrypted":
|
||||||
|
return False
|
||||||
|
parts_count = 0
|
||||||
|
for part in message.iter_parts():
|
||||||
|
# We explicitly check Content-Type of each part later,
|
||||||
|
# but this is to be absolutely sure `get_payload()` returns string and not list.
|
||||||
|
if part.is_multipart():
|
||||||
|
return False
|
||||||
|
|
||||||
|
if parts_count == 0:
|
||||||
|
if part.get_content_type() != "application/pgp-encrypted":
|
||||||
|
return False
|
||||||
|
|
||||||
|
payload = part.get_payload()
|
||||||
|
if payload.strip() != "Version: 1":
|
||||||
|
return False
|
||||||
|
elif parts_count == 1:
|
||||||
|
if part.get_content_type() != "application/octet-stream":
|
||||||
|
return False
|
||||||
|
|
||||||
|
if not check_armored_payload(part.get_payload()):
|
||||||
|
return False
|
||||||
|
else:
|
||||||
|
return False
|
||||||
|
parts_count += 1
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
async def asyncmain_beforequeue(config, mode):
|
||||||
|
if mode == "outgoing":
|
||||||
|
port = config.filtermail_smtp_port
|
||||||
|
handler = OutgoingBeforeQueueHandler(config)
|
||||||
|
else:
|
||||||
|
port = config.filtermail_smtp_port_incoming
|
||||||
|
handler = IncomingBeforeQueueHandler(config)
|
||||||
|
HackedController(handler, hostname="127.0.0.1", port=port).start()
|
||||||
|
|
||||||
|
|
||||||
|
def recipient_matches_passthrough(recipient, passthrough_recipients):
|
||||||
|
for addr in passthrough_recipients:
|
||||||
|
if recipient == addr:
|
||||||
|
return True
|
||||||
|
if addr[0] == "@" and recipient.endswith(addr):
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
class HackedController(Controller):
|
||||||
|
def factory(self):
|
||||||
|
return SMTPDiscardRCPTO_options(self.handler, **self.SMTP_kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
class SMTPDiscardRCPTO_options(SMTP):
|
||||||
|
def _getparams(self, params):
|
||||||
|
# aiosmtpd's SMTP daemon fails to handle a request if there are RCPT TO options
|
||||||
|
# We just ignore them for our incoming filtermail purposes
|
||||||
|
if len(params) == 1 and params[0].startswith("ORCPT"):
|
||||||
|
return {}
|
||||||
|
return super()._getparams(params)
|
||||||
|
|
||||||
|
|
||||||
|
class OutgoingBeforeQueueHandler:
|
||||||
|
def __init__(self, config):
|
||||||
|
self.config = config
|
||||||
|
self.send_rate_limiter = SendRateLimiter()
|
||||||
|
|
||||||
|
async def handle_MAIL(self, server, session, envelope, address, mail_options):
|
||||||
|
logging.info(f"handle_MAIL from {address}")
|
||||||
|
envelope.mail_from = address
|
||||||
|
max_sent = self.config.max_user_send_per_minute
|
||||||
|
if not self.send_rate_limiter.is_sending_allowed(address, max_sent):
|
||||||
|
return f"450 4.7.1: Too much mail from {address}"
|
||||||
|
|
||||||
|
parts = envelope.mail_from.split("@")
|
||||||
|
if len(parts) != 2:
|
||||||
|
return f"500 Invalid from address <{envelope.mail_from!r}>"
|
||||||
|
|
||||||
|
return "250 OK"
|
||||||
|
|
||||||
|
async def handle_DATA(self, server, session, envelope):
|
||||||
|
logging.info("handle_DATA before-queue")
|
||||||
|
error = self.check_DATA(envelope)
|
||||||
|
if error:
|
||||||
|
return error
|
||||||
|
logging.info("re-injecting the mail that passed checks")
|
||||||
|
client = SMTPClient("localhost", self.config.postfix_reinject_port)
|
||||||
|
client.sendmail(
|
||||||
|
envelope.mail_from, envelope.rcpt_tos, envelope.original_content
|
||||||
|
)
|
||||||
|
return "250 OK"
|
||||||
|
|
||||||
|
def check_DATA(self, envelope):
|
||||||
|
"""the central filtering function for e-mails."""
|
||||||
|
logging.info(f"Processing DATA message from {envelope.mail_from}")
|
||||||
|
|
||||||
|
message = BytesParser(policy=policy.default).parsebytes(envelope.content)
|
||||||
|
mail_encrypted = check_encrypted(message)
|
||||||
|
|
||||||
|
_, from_addr = parseaddr(message.get("from").strip())
|
||||||
|
|
||||||
|
if envelope.mail_from.lower() != from_addr.lower():
|
||||||
|
return f"500 Invalid FROM <{from_addr!r}> for <{envelope.mail_from!r}>"
|
||||||
|
|
||||||
|
if mail_encrypted or is_securejoin(message):
|
||||||
|
print("Outgoing: Filtering encrypted mail.", file=sys.stderr)
|
||||||
|
return
|
||||||
|
|
||||||
|
print("Outgoing: Filtering unencrypted mail.", file=sys.stderr)
|
||||||
|
|
||||||
|
if envelope.mail_from in self.config.passthrough_senders:
|
||||||
|
return
|
||||||
|
|
||||||
|
# allow self-sent Autocrypt Setup Message
|
||||||
|
if envelope.rcpt_tos == [from_addr]:
|
||||||
|
if message.get("subject") == "Autocrypt Setup Message":
|
||||||
|
if message.get_content_type() == "multipart/mixed":
|
||||||
|
return
|
||||||
|
|
||||||
|
passthrough_recipients = self.config.passthrough_recipients
|
||||||
|
|
||||||
|
for recipient in envelope.rcpt_tos:
|
||||||
|
if recipient_matches_passthrough(recipient, passthrough_recipients):
|
||||||
|
continue
|
||||||
|
|
||||||
|
print("Rejected unencrypted mail.", file=sys.stderr)
|
||||||
|
return ENCRYPTION_NEEDED_523
|
||||||
|
|
||||||
|
|
||||||
|
class IncomingBeforeQueueHandler:
|
||||||
|
def __init__(self, config):
|
||||||
|
self.config = config
|
||||||
|
|
||||||
|
async def handle_DATA(self, server, session, envelope):
|
||||||
|
logging.info("handle_DATA before-queue")
|
||||||
|
error = self.check_DATA(envelope)
|
||||||
|
if error:
|
||||||
|
return error
|
||||||
|
logging.info("re-injecting the mail that passed checks")
|
||||||
|
|
||||||
|
# the smtp daemon on reinject_port_incoming gives it to dkim milter
|
||||||
|
# which looks at source address to determine whether to verify or sign
|
||||||
|
client = SMTPClient(
|
||||||
|
"localhost",
|
||||||
|
self.config.postfix_reinject_port_incoming,
|
||||||
|
source_address=("127.0.0.2", 0),
|
||||||
|
)
|
||||||
|
client.sendmail(
|
||||||
|
envelope.mail_from, envelope.rcpt_tos, envelope.original_content
|
||||||
|
)
|
||||||
|
return "250 OK"
|
||||||
|
|
||||||
|
def check_DATA(self, envelope):
|
||||||
|
"""the central filtering function for e-mails."""
|
||||||
|
logging.info(f"Processing DATA message from {envelope.mail_from}")
|
||||||
|
|
||||||
|
message = BytesParser(policy=policy.default).parsebytes(envelope.content)
|
||||||
|
mail_encrypted = check_encrypted(message)
|
||||||
|
|
||||||
|
if mail_encrypted or is_securejoin(message):
|
||||||
|
print("Incoming: Filtering encrypted mail.", file=sys.stderr)
|
||||||
|
return
|
||||||
|
|
||||||
|
print("Incoming: Filtering unencrypted mail.", file=sys.stderr)
|
||||||
|
|
||||||
|
# we want cleartext mailer-daemon messages to pass through
|
||||||
|
# chatmail core will typically not display them as normal messages
|
||||||
|
if message.get("auto-submitted"):
|
||||||
|
_, from_addr = parseaddr(message.get("from").strip())
|
||||||
|
if from_addr.lower().startswith("mailer-daemon@"):
|
||||||
|
if message.get_content_type() == "multipart/report":
|
||||||
|
return
|
||||||
|
|
||||||
|
for recipient in envelope.rcpt_tos:
|
||||||
|
user = self.config.get_user(recipient)
|
||||||
|
if user is None or user.is_incoming_cleartext_ok():
|
||||||
|
continue
|
||||||
|
|
||||||
|
print("Rejected unencrypted mail.", file=sys.stderr)
|
||||||
|
return ENCRYPTION_NEEDED_523
|
||||||
|
|
||||||
|
|
||||||
|
class SendRateLimiter:
|
||||||
|
def __init__(self):
|
||||||
|
self.addr2timestamps = {}
|
||||||
|
|
||||||
|
def is_sending_allowed(self, mail_from, max_send_per_minute):
|
||||||
|
last = self.addr2timestamps.setdefault(mail_from, [])
|
||||||
|
now = time.time()
|
||||||
|
last[:] = [ts for ts in last if ts >= (now - 60)]
|
||||||
|
if len(last) <= max_send_per_minute:
|
||||||
|
last.append(now)
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
args = sys.argv[1:]
|
||||||
|
assert len(args) == 2
|
||||||
|
config = read_config(args[0])
|
||||||
|
mode = args[1]
|
||||||
|
logging.basicConfig(level=logging.WARN)
|
||||||
|
loop = asyncio.new_event_loop()
|
||||||
|
asyncio.set_event_loop(loop)
|
||||||
|
assert mode in ["incoming", "outgoing"]
|
||||||
|
task = asyncmain_beforequeue(config, mode)
|
||||||
|
loop.create_task(task)
|
||||||
|
logging.info("entering serving loop")
|
||||||
|
loop.run_forever()
|
||||||
@@ -1,287 +0,0 @@
|
|||||||
"""
|
|
||||||
command line tool to analyze mailbox message storage
|
|
||||||
|
|
||||||
example invocation:
|
|
||||||
|
|
||||||
python -m chatmaild.fsreport /path/to/chatmail.ini
|
|
||||||
|
|
||||||
to show storage summaries for all "cur" folders
|
|
||||||
|
|
||||||
python -m chatmaild.fsreport /path/to/chatmail.ini --mdir cur
|
|
||||||
|
|
||||||
to show storage summaries only for first 1000 mailboxes
|
|
||||||
|
|
||||||
python -m chatmaild.fsreport /path/to/chatmail.ini --maxnum 1000
|
|
||||||
|
|
||||||
to write Prometheus textfile for node_exporter
|
|
||||||
|
|
||||||
python -m chatmaild.fsreport --textfile /var/lib/prometheus/node-exporter/
|
|
||||||
|
|
||||||
writes to /var/lib/prometheus/node-exporter/fsreport.prom
|
|
||||||
|
|
||||||
to also write legacy metrics.py style output (default: /var/www/html/metrics):
|
|
||||||
|
|
||||||
python -m chatmaild.fsreport --textfile /var/lib/prometheus/node-exporter/ --legacy-metrics
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os
|
|
||||||
import tempfile
|
|
||||||
from argparse import ArgumentParser
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
from chatmaild.config import read_config
|
|
||||||
from chatmaild.expire import iter_mailboxes
|
|
||||||
|
|
||||||
DAYSECONDS = 24 * 60 * 60
|
|
||||||
MONTHSECONDS = DAYSECONDS * 30
|
|
||||||
|
|
||||||
|
|
||||||
def HSize(size: int):
|
|
||||||
"""Format a size integer as a Human-readable string Kilobyte, Megabyte or Gigabyte"""
|
|
||||||
if size < 10000:
|
|
||||||
return f"{size / 1000:5.2f}K"
|
|
||||||
if size < 1000 * 1000:
|
|
||||||
return f"{size / 1000:5.0f}K"
|
|
||||||
if size < 1000 * 1000 * 1000:
|
|
||||||
return f"{int(size / 1000000):5.0f}M"
|
|
||||||
return f"{size / 1000000000:5.2f}G"
|
|
||||||
|
|
||||||
|
|
||||||
class Report:
|
|
||||||
def __init__(self, now, min_login_age, mdir):
|
|
||||||
self.size_extra = 0
|
|
||||||
self.size_messages = 0
|
|
||||||
self.now = now
|
|
||||||
self.min_login_age = min_login_age
|
|
||||||
self.mdir = mdir
|
|
||||||
|
|
||||||
self.num_ci_logins = self.num_all_logins = 0
|
|
||||||
self.login_buckets = {x: 0 for x in (1, 10, 30, 40, 80, 100, 150)}
|
|
||||||
|
|
||||||
KiB = 1024
|
|
||||||
MiB = 1024 * KiB
|
|
||||||
self.message_size_thresholds = (
|
|
||||||
0,
|
|
||||||
100 * KiB,
|
|
||||||
MiB // 2,
|
|
||||||
1 * MiB,
|
|
||||||
2 * MiB,
|
|
||||||
5 * MiB,
|
|
||||||
10 * MiB,
|
|
||||||
)
|
|
||||||
self.message_buckets = {x: 0 for x in self.message_size_thresholds}
|
|
||||||
self.message_count_buckets = {x: 0 for x in self.message_size_thresholds}
|
|
||||||
|
|
||||||
def process_mailbox_stat(self, mailbox):
|
|
||||||
# categorize login times
|
|
||||||
last_login = mailbox.last_login
|
|
||||||
if last_login:
|
|
||||||
self.num_all_logins += 1
|
|
||||||
if os.path.basename(mailbox.basedir)[:3] == "ci-":
|
|
||||||
self.num_ci_logins += 1
|
|
||||||
else:
|
|
||||||
for days in self.login_buckets:
|
|
||||||
if last_login >= self.now - days * DAYSECONDS:
|
|
||||||
self.login_buckets[days] += 1
|
|
||||||
|
|
||||||
cutoff_login_date = self.now - self.min_login_age * DAYSECONDS
|
|
||||||
if last_login and last_login <= cutoff_login_date:
|
|
||||||
# categorize message sizes
|
|
||||||
for size in self.message_buckets:
|
|
||||||
for msg in mailbox.messages:
|
|
||||||
if msg.size >= size:
|
|
||||||
if self.mdir and f"/{self.mdir}/" not in msg.path:
|
|
||||||
continue
|
|
||||||
self.message_buckets[size] += msg.size
|
|
||||||
self.message_count_buckets[size] += 1
|
|
||||||
|
|
||||||
self.size_messages += sum(entry.size for entry in mailbox.messages)
|
|
||||||
self.size_extra += sum(entry.size for entry in mailbox.extrafiles)
|
|
||||||
|
|
||||||
def dump_summary(self):
|
|
||||||
all_messages = self.size_messages
|
|
||||||
print()
|
|
||||||
print("## Mailbox storage use analysis")
|
|
||||||
print(f"Mailbox data total size: {HSize(self.size_extra + all_messages)}")
|
|
||||||
print(f"Messages total size : {HSize(all_messages)}")
|
|
||||||
try:
|
|
||||||
percent = self.size_extra / (self.size_extra + all_messages) * 100
|
|
||||||
except ZeroDivisionError:
|
|
||||||
percent = 100
|
|
||||||
print(f"Extra files : {HSize(self.size_extra)} ({percent:.2f}%)")
|
|
||||||
|
|
||||||
print()
|
|
||||||
if self.min_login_age:
|
|
||||||
print(f"### Message storage for {self.min_login_age} days old logins")
|
|
||||||
|
|
||||||
pref = f"[{self.mdir}] " if self.mdir else ""
|
|
||||||
for minsize, sumsize in self.message_buckets.items():
|
|
||||||
count = self.message_count_buckets[minsize]
|
|
||||||
percent = (sumsize / all_messages * 100) if all_messages else 0
|
|
||||||
print(
|
|
||||||
f"{pref}larger than {HSize(minsize)}: {HSize(sumsize)} ({percent:.2f}%), {count} msgs"
|
|
||||||
)
|
|
||||||
|
|
||||||
user_logins = self.num_all_logins - self.num_ci_logins
|
|
||||||
|
|
||||||
def p(num):
|
|
||||||
return f"({num / user_logins * 100:2.2f}%)" if user_logins else "100%"
|
|
||||||
|
|
||||||
print()
|
|
||||||
print(f"## Login stats, from date reference {datetime.fromtimestamp(self.now)}")
|
|
||||||
print(f"all: {HSize(self.num_all_logins)}")
|
|
||||||
print(f"non-ci: {HSize(user_logins)}")
|
|
||||||
print(f"ci: {HSize(self.num_ci_logins)}")
|
|
||||||
for days, active in self.login_buckets.items():
|
|
||||||
print(f"last {days:3} days: {HSize(active)} {p(active)}")
|
|
||||||
|
|
||||||
def _write_atomic(self, filepath, content):
|
|
||||||
"""Atomically write content to filepath via tmp+rename."""
|
|
||||||
dirpath = os.path.dirname(os.path.abspath(filepath))
|
|
||||||
fd, tmppath = tempfile.mkstemp(dir=dirpath, suffix=".tmp")
|
|
||||||
try:
|
|
||||||
with os.fdopen(fd, "w") as f:
|
|
||||||
f.write(content)
|
|
||||||
os.chmod(tmppath, 0o644)
|
|
||||||
os.rename(tmppath, filepath)
|
|
||||||
except BaseException:
|
|
||||||
try:
|
|
||||||
os.unlink(tmppath)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
raise
|
|
||||||
|
|
||||||
def dump_textfile(self, filepath):
|
|
||||||
"""Dump metrics in Prometheus exposition format."""
|
|
||||||
lines = []
|
|
||||||
|
|
||||||
lines.append("# HELP chatmail_storage_bytes Mailbox storage in bytes.")
|
|
||||||
lines.append("# TYPE chatmail_storage_bytes gauge")
|
|
||||||
lines.append(f'chatmail_storage_bytes{{kind="messages"}} {self.size_messages}')
|
|
||||||
lines.append(f'chatmail_storage_bytes{{kind="extra"}} {self.size_extra}')
|
|
||||||
total = self.size_extra + self.size_messages
|
|
||||||
lines.append(f'chatmail_storage_bytes{{kind="total"}} {total}')
|
|
||||||
|
|
||||||
lines.append("# HELP chatmail_messages_bytes Sum of msg bytes >= threshold.")
|
|
||||||
lines.append("# TYPE chatmail_messages_bytes gauge")
|
|
||||||
for minsize, sumsize in self.message_buckets.items():
|
|
||||||
lines.append(f'chatmail_messages_bytes{{min_size="{minsize}"}} {sumsize}')
|
|
||||||
|
|
||||||
lines.append("# HELP chatmail_messages_count Number of msgs >= size threshold.")
|
|
||||||
lines.append("# TYPE chatmail_messages_count gauge")
|
|
||||||
for minsize, count in self.message_count_buckets.items():
|
|
||||||
lines.append(f'chatmail_messages_count{{min_size="{minsize}"}} {count}')
|
|
||||||
|
|
||||||
lines.append("# HELP chatmail_accounts Number of accounts.")
|
|
||||||
lines.append("# TYPE chatmail_accounts gauge")
|
|
||||||
user_logins = self.num_all_logins - self.num_ci_logins
|
|
||||||
lines.append(f'chatmail_accounts{{kind="all"}} {self.num_all_logins}')
|
|
||||||
lines.append(f'chatmail_accounts{{kind="ci"}} {self.num_ci_logins}')
|
|
||||||
lines.append(f'chatmail_accounts{{kind="user"}} {user_logins}')
|
|
||||||
|
|
||||||
lines.append(
|
|
||||||
"# HELP chatmail_accounts_active Non-CI accounts active within N days."
|
|
||||||
)
|
|
||||||
lines.append("# TYPE chatmail_accounts_active gauge")
|
|
||||||
for days, active in self.login_buckets.items():
|
|
||||||
lines.append(f'chatmail_accounts_active{{days="{days}"}} {active}')
|
|
||||||
|
|
||||||
self._write_atomic(filepath, "\n".join(lines) + "\n")
|
|
||||||
|
|
||||||
def dump_compat_textfile(self, filepath):
|
|
||||||
"""Dump legacy metrics.py style metrics."""
|
|
||||||
user_logins = self.num_all_logins - self.num_ci_logins
|
|
||||||
lines = [
|
|
||||||
"# HELP total number of accounts",
|
|
||||||
"# TYPE accounts gauge",
|
|
||||||
f"accounts {self.num_all_logins}",
|
|
||||||
"# HELP number of CI accounts",
|
|
||||||
"# TYPE ci_accounts gauge",
|
|
||||||
f"ci_accounts {self.num_ci_logins}",
|
|
||||||
"# HELP number of non-CI accounts",
|
|
||||||
"# TYPE nonci_accounts gauge",
|
|
||||||
f"nonci_accounts {user_logins}",
|
|
||||||
]
|
|
||||||
self._write_atomic(filepath, "\n".join(lines) + "\n")
|
|
||||||
|
|
||||||
|
|
||||||
def main(args=None):
|
|
||||||
"""Report about filesystem storage usage of all mailboxes and messages"""
|
|
||||||
parser = ArgumentParser(description=main.__doc__)
|
|
||||||
ini = "/usr/local/lib/chatmaild/chatmail.ini"
|
|
||||||
parser.add_argument(
|
|
||||||
"chatmail_ini",
|
|
||||||
action="store",
|
|
||||||
nargs="?",
|
|
||||||
help=f"path pointing to chatmail.ini file, default: {ini}",
|
|
||||||
default=ini,
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--days",
|
|
||||||
default=0,
|
|
||||||
action="store",
|
|
||||||
help="assume date to be DAYS older than now",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--min-login-age",
|
|
||||||
default=0,
|
|
||||||
metavar="DAYS",
|
|
||||||
dest="min_login_age",
|
|
||||||
action="store",
|
|
||||||
help="only sum up message size if last login is at least DAYS days old",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--mdir",
|
|
||||||
metavar="{cur,new,tmp}",
|
|
||||||
action="store",
|
|
||||||
help="only consider messages in specified Maildir subdirectory for summary",
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
"--maxnum",
|
|
||||||
default=None,
|
|
||||||
action="store",
|
|
||||||
help="maximum number of mailboxes to iterate on",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--textfile",
|
|
||||||
metavar="PATH",
|
|
||||||
default=None,
|
|
||||||
help="write Prometheus textfile to PATH (directory or file); "
|
|
||||||
"if PATH is a directory, writes 'fsreport.prom' inside it",
|
|
||||||
)
|
|
||||||
parser.add_argument(
|
|
||||||
"--legacy-metrics",
|
|
||||||
metavar="FILENAME",
|
|
||||||
nargs="?",
|
|
||||||
const="/var/www/html/metrics",
|
|
||||||
default=None,
|
|
||||||
help="write legacy metrics.py textfile (default: /var/www/html/metrics)",
|
|
||||||
)
|
|
||||||
|
|
||||||
args = parser.parse_args(args)
|
|
||||||
|
|
||||||
config = read_config(args.chatmail_ini)
|
|
||||||
|
|
||||||
now = datetime.utcnow().timestamp()
|
|
||||||
if args.days:
|
|
||||||
now = now - 86400 * int(args.days)
|
|
||||||
|
|
||||||
maxnum = int(args.maxnum) if args.maxnum else None
|
|
||||||
rep = Report(now=now, min_login_age=int(args.min_login_age), mdir=args.mdir)
|
|
||||||
for mbox in iter_mailboxes(str(config.mailboxes_dir), maxnum=maxnum):
|
|
||||||
rep.process_mailbox_stat(mbox)
|
|
||||||
if args.textfile:
|
|
||||||
path = args.textfile
|
|
||||||
if os.path.isdir(path):
|
|
||||||
path = os.path.join(path, "fsreport.prom")
|
|
||||||
rep.dump_textfile(path)
|
|
||||||
if args.legacy_metrics:
|
|
||||||
rep.dump_compat_textfile(args.legacy_metrics)
|
|
||||||
if not args.textfile and not args.legacy_metrics:
|
|
||||||
rep.dump_summary()
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
main()
|
|
||||||
@@ -11,15 +11,11 @@ mail_domain = {mail_domain}
|
|||||||
# Restrictions on user addresses
|
# Restrictions on user addresses
|
||||||
#
|
#
|
||||||
|
|
||||||
# email sending rate per user and minute
|
# how many mails a user can send out per minute
|
||||||
max_user_send_per_minute = 60
|
max_user_send_per_minute = 60
|
||||||
|
|
||||||
# per-user max burst size for sending rate limiting (GCRA bucket capacity)
|
|
||||||
max_user_send_burst_size = 10
|
|
||||||
|
|
||||||
# maximum mailbox size of a chatmail address
|
# maximum mailbox size of a chatmail address
|
||||||
# Oldest messages will be removed automatically, so mailboxes never run full.
|
max_mailbox_size = 100M
|
||||||
max_mailbox_size = 500M
|
|
||||||
|
|
||||||
# maximum message size for an e-mail in bytes
|
# maximum message size for an e-mail in bytes
|
||||||
max_message_size = 31457280
|
max_message_size = 31457280
|
||||||
@@ -27,9 +23,6 @@ max_message_size = 31457280
|
|||||||
# days after which mails are unconditionally deleted
|
# days after which mails are unconditionally deleted
|
||||||
delete_mails_after = 20
|
delete_mails_after = 20
|
||||||
|
|
||||||
# days after which large messages (>200k) are unconditionally deleted
|
|
||||||
delete_large_after = 7
|
|
||||||
|
|
||||||
# days after which users without a successful login are deleted (database and mails)
|
# days after which users without a successful login are deleted (database and mails)
|
||||||
delete_inactive_users_after = 90
|
delete_inactive_users_after = 90
|
||||||
|
|
||||||
@@ -47,17 +40,7 @@ passthrough_senders =
|
|||||||
|
|
||||||
# list of e-mail recipients for which to accept outbound un-encrypted mails
|
# list of e-mail recipients for which to accept outbound un-encrypted mails
|
||||||
# (space-separated, item may start with "@" to whitelist whole recipient domains)
|
# (space-separated, item may start with "@" to whitelist whole recipient domains)
|
||||||
passthrough_recipients =
|
passthrough_recipients = xstore@testrun.org
|
||||||
|
|
||||||
# Use externally managed TLS certificates instead of built-in acmetool.
|
|
||||||
# Paths refer to files on the deployment server (not the build machine).
|
|
||||||
# Both files must already exist before running cmdeploy.
|
|
||||||
# Certificate renewal is your responsibility; changed files are
|
|
||||||
# picked up automatically by all relay services.
|
|
||||||
# tls_external_cert_and_key = /path/to/fullchain.pem /path/to/privkey.pem
|
|
||||||
|
|
||||||
# path to www directory - documented here: https://chatmail.at/doc/relay/getting_started.html#custom-web-pages
|
|
||||||
#www_folder = www
|
|
||||||
|
|
||||||
#
|
#
|
||||||
# Deployment Details
|
# Deployment Details
|
||||||
@@ -74,9 +57,6 @@ postfix_reinject_port_incoming = 10026
|
|||||||
# if set to "True" IPv6 is disabled
|
# if set to "True" IPv6 is disabled
|
||||||
disable_ipv6 = False
|
disable_ipv6 = False
|
||||||
|
|
||||||
# Your email adress, which will be used in acmetool to manage Let's Encrypt SSL certificates
|
|
||||||
acme_email =
|
|
||||||
|
|
||||||
# Defaults to https://iroh.{{mail_domain}} and running `iroh-relay` on the chatmail
|
# Defaults to https://iroh.{{mail_domain}} and running `iroh-relay` on the chatmail
|
||||||
# service.
|
# service.
|
||||||
# If you set it to anything else, the service will be disabled
|
# If you set it to anything else, the service will be disabled
|
||||||
@@ -110,12 +90,6 @@ acme_email =
|
|||||||
# so use this option with caution on production servers.
|
# so use this option with caution on production servers.
|
||||||
imap_rawlog = false
|
imap_rawlog = false
|
||||||
|
|
||||||
# set to true if you want to enable the IMAP COMPRESS Extension,
|
|
||||||
# which allows IMAP connections to be efficiently compressed.
|
|
||||||
# WARNING: Enabling this makes it impossible to hibernate IMAP
|
|
||||||
# processes which will result in much higher memory/RAM usage.
|
|
||||||
imap_compress = false
|
|
||||||
|
|
||||||
|
|
||||||
#
|
#
|
||||||
# Privacy Policy
|
# Privacy Policy
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
|
|
||||||
[privacy]
|
[privacy]
|
||||||
|
|
||||||
passthrough_recipients = privacy@testrun.org echo@{mail_domain}
|
passthrough_recipients = privacy@testrun.org xstore@testrun.org
|
||||||
|
|
||||||
privacy_postal =
|
privacy_postal =
|
||||||
Merlinux GmbH, Represented by the managing director H. Krekel,
|
Merlinux GmbH, Represented by the managing director H. Krekel,
|
||||||
|
|||||||
@@ -13,6 +13,8 @@ class LastLoginDictProxy(DictProxy):
|
|||||||
keyname = parts[1].split("/")
|
keyname = parts[1].split("/")
|
||||||
value = parts[2] if len(parts) > 2 else ""
|
value = parts[2] if len(parts) > 2 else ""
|
||||||
if keyname[0] == "shared" and keyname[1] == "last-login":
|
if keyname[0] == "shared" and keyname[1] == "last-login":
|
||||||
|
if addr.startswith("echo@"):
|
||||||
|
return True
|
||||||
addr = keyname[2]
|
addr = keyname[2]
|
||||||
timestamp = int(value)
|
timestamp = int(value)
|
||||||
user = self.config.get_user(addr)
|
user = self.config.get_user(addr)
|
||||||
|
|||||||
@@ -1,24 +1,14 @@
|
|||||||
import logging
|
import logging
|
||||||
import sys
|
import sys
|
||||||
import time
|
|
||||||
from contextlib import contextmanager
|
|
||||||
|
|
||||||
from .config import read_config
|
from .config import read_config
|
||||||
from .dictproxy import DictProxy
|
from .dictproxy import DictProxy
|
||||||
from .filedict import FileDict
|
from .filedict import FileDict
|
||||||
from .notifier import Notifier
|
from .notifier import Notifier
|
||||||
from .turnserver import turn_credentials
|
|
||||||
|
|
||||||
|
|
||||||
def _is_valid_token_timestamp(timestamp, now):
|
|
||||||
# Token if invalid after 90 days
|
|
||||||
# or if the timestamp is in the future.
|
|
||||||
return timestamp > now - 3600 * 24 * 90 and timestamp < now + 60
|
|
||||||
|
|
||||||
|
|
||||||
class Metadata:
|
class Metadata:
|
||||||
# each SETMETADATA on this key appends to dictionary
|
# each SETMETADATA on this key appends to a list of unique device tokens
|
||||||
# mapping of unique device tokens
|
|
||||||
# which only ever get removed if the upstream indicates the token is invalid
|
# which only ever get removed if the upstream indicates the token is invalid
|
||||||
DEVICETOKEN_KEY = "devicetoken"
|
DEVICETOKEN_KEY = "devicetoken"
|
||||||
|
|
||||||
@@ -28,60 +18,29 @@ class Metadata:
|
|||||||
def get_metadata_dict(self, addr):
|
def get_metadata_dict(self, addr):
|
||||||
return FileDict(self.vmail_dir / addr / "metadata.json")
|
return FileDict(self.vmail_dir / addr / "metadata.json")
|
||||||
|
|
||||||
@contextmanager
|
|
||||||
def _modify_tokens(self, addr):
|
|
||||||
with self.get_metadata_dict(addr).modify() as data:
|
|
||||||
tokens = data.setdefault(self.DEVICETOKEN_KEY, {})
|
|
||||||
now = int(time.time())
|
|
||||||
if isinstance(tokens, list):
|
|
||||||
data[self.DEVICETOKEN_KEY] = tokens = {t: now for t in tokens}
|
|
||||||
|
|
||||||
expired_tokens = [
|
|
||||||
token
|
|
||||||
for token, timestamp in tokens.items()
|
|
||||||
if not _is_valid_token_timestamp(tokens[token], now)
|
|
||||||
]
|
|
||||||
for expired_token in expired_tokens:
|
|
||||||
del tokens[expired_token]
|
|
||||||
|
|
||||||
yield tokens
|
|
||||||
|
|
||||||
def add_token_to_addr(self, addr, token):
|
def add_token_to_addr(self, addr, token):
|
||||||
with self._modify_tokens(addr) as tokens:
|
with self.get_metadata_dict(addr).modify() as data:
|
||||||
tokens[token] = int(time.time())
|
tokens = data.setdefault(self.DEVICETOKEN_KEY, [])
|
||||||
|
if token not in tokens:
|
||||||
|
tokens.append(token)
|
||||||
|
|
||||||
def remove_token_from_addr(self, addr, token):
|
def remove_token_from_addr(self, addr, token):
|
||||||
with self._modify_tokens(addr) as tokens:
|
with self.get_metadata_dict(addr).modify() as data:
|
||||||
|
tokens = data.get(self.DEVICETOKEN_KEY, [])
|
||||||
if token in tokens:
|
if token in tokens:
|
||||||
del tokens[token]
|
tokens.remove(token)
|
||||||
|
|
||||||
def get_tokens_for_addr(self, addr):
|
def get_tokens_for_addr(self, addr):
|
||||||
mdict = self.get_metadata_dict(addr).read()
|
mdict = self.get_metadata_dict(addr).read()
|
||||||
tokens = mdict.get(self.DEVICETOKEN_KEY, {})
|
return mdict.get(self.DEVICETOKEN_KEY, [])
|
||||||
|
|
||||||
now = int(time.time())
|
|
||||||
if isinstance(tokens, dict):
|
|
||||||
token_list = [
|
|
||||||
token
|
|
||||||
for token, timestamp in tokens.items()
|
|
||||||
if _is_valid_token_timestamp(timestamp, now)
|
|
||||||
]
|
|
||||||
if len(token_list) < len(tokens):
|
|
||||||
# Some tokens have expired, remove them.
|
|
||||||
with self._modify_tokens(addr) as _tokens:
|
|
||||||
pass
|
|
||||||
else:
|
|
||||||
token_list = []
|
|
||||||
return token_list
|
|
||||||
|
|
||||||
|
|
||||||
class MetadataDictProxy(DictProxy):
|
class MetadataDictProxy(DictProxy):
|
||||||
def __init__(self, notifier, metadata, iroh_relay=None, turn_hostname=None):
|
def __init__(self, notifier, metadata, iroh_relay=None):
|
||||||
super().__init__()
|
super().__init__()
|
||||||
self.notifier = notifier
|
self.notifier = notifier
|
||||||
self.metadata = metadata
|
self.metadata = metadata
|
||||||
self.iroh_relay = iroh_relay
|
self.iroh_relay = iroh_relay
|
||||||
self.turn_hostname = turn_hostname
|
|
||||||
|
|
||||||
def handle_lookup(self, parts):
|
def handle_lookup(self, parts):
|
||||||
# Lpriv/43f5f508a7ea0366dff30200c15250e3/devicetoken\tlkj123poi@c2.testrun.org
|
# Lpriv/43f5f508a7ea0366dff30200c15250e3/devicetoken\tlkj123poi@c2.testrun.org
|
||||||
@@ -100,15 +59,6 @@ class MetadataDictProxy(DictProxy):
|
|||||||
):
|
):
|
||||||
# Handle `GETMETADATA "" /shared/vendor/deltachat/irohrelay`
|
# Handle `GETMETADATA "" /shared/vendor/deltachat/irohrelay`
|
||||||
return f"O{self.iroh_relay}\n"
|
return f"O{self.iroh_relay}\n"
|
||||||
elif keyname == "vendor/vendor.dovecot/pvt/server/vendor/deltachat/turn":
|
|
||||||
try:
|
|
||||||
res = turn_credentials()
|
|
||||||
except Exception:
|
|
||||||
logging.exception("failed to get TURN credentials")
|
|
||||||
return "N\n"
|
|
||||||
port = 3478
|
|
||||||
return f"O{self.turn_hostname}:{port}:{res}\n"
|
|
||||||
|
|
||||||
logging.warning(f"lookup ignored: {parts!r}")
|
logging.warning(f"lookup ignored: {parts!r}")
|
||||||
return "N\n"
|
return "N\n"
|
||||||
|
|
||||||
@@ -132,7 +82,6 @@ def main():
|
|||||||
|
|
||||||
config = read_config(config_path)
|
config = read_config(config_path)
|
||||||
iroh_relay = config.iroh_relay
|
iroh_relay = config.iroh_relay
|
||||||
mail_domain = config.mail_domain
|
|
||||||
|
|
||||||
vmail_dir = config.mailboxes_dir
|
vmail_dir = config.mailboxes_dir
|
||||||
if not vmail_dir.exists():
|
if not vmail_dir.exists():
|
||||||
@@ -146,10 +95,7 @@ def main():
|
|||||||
notifier.start_notification_threads(metadata.remove_token_from_addr)
|
notifier.start_notification_threads(metadata.remove_token_from_addr)
|
||||||
|
|
||||||
dictproxy = MetadataDictProxy(
|
dictproxy = MetadataDictProxy(
|
||||||
notifier=notifier,
|
notifier=notifier, metadata=metadata, iroh_relay=iroh_relay
|
||||||
metadata=metadata,
|
|
||||||
iroh_relay=iroh_relay,
|
|
||||||
turn_hostname=mail_domain,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
dictproxy.serve_forever_from_socket(socket)
|
dictproxy.serve_forever_from_socket(socket)
|
||||||
|
|||||||
32
chatmaild/src/chatmaild/metrics.py
Normal file
32
chatmaild/src/chatmaild/metrics.py
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
|
||||||
|
def main(vmail_dir=None):
|
||||||
|
if vmail_dir is None:
|
||||||
|
vmail_dir = sys.argv[1]
|
||||||
|
|
||||||
|
accounts = 0
|
||||||
|
ci_accounts = 0
|
||||||
|
|
||||||
|
for path in Path(vmail_dir).iterdir():
|
||||||
|
if not path.joinpath("cur").is_dir():
|
||||||
|
continue
|
||||||
|
accounts += 1
|
||||||
|
if path.name[:3] in ("ci-", "ac_"):
|
||||||
|
ci_accounts += 1
|
||||||
|
|
||||||
|
print("# HELP total number of accounts")
|
||||||
|
print("# TYPE accounts gauge")
|
||||||
|
print(f"accounts {accounts}")
|
||||||
|
print("# HELP number of CI accounts")
|
||||||
|
print("# TYPE ci_accounts gauge")
|
||||||
|
print(f"ci_accounts {ci_accounts}")
|
||||||
|
print("# HELP number of non-CI accounts")
|
||||||
|
print("# TYPE nonci_accounts gauge")
|
||||||
|
print(f"nonci_accounts {accounts - ci_accounts}")
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
@@ -2,11 +2,10 @@
|
|||||||
|
|
||||||
"""CGI script for creating new accounts."""
|
"""CGI script for creating new accounts."""
|
||||||
|
|
||||||
import ipaddress
|
|
||||||
import json
|
import json
|
||||||
|
import random
|
||||||
import secrets
|
import secrets
|
||||||
import string
|
import string
|
||||||
from urllib.parse import quote
|
|
||||||
|
|
||||||
from chatmaild.config import Config, read_config
|
from chatmaild.config import Config, read_config
|
||||||
|
|
||||||
@@ -15,47 +14,22 @@ ALPHANUMERIC = string.ascii_lowercase + string.digits
|
|||||||
ALPHANUMERIC_PUNCT = string.ascii_letters + string.digits + string.punctuation
|
ALPHANUMERIC_PUNCT = string.ascii_letters + string.digits + string.punctuation
|
||||||
|
|
||||||
|
|
||||||
def wrap_ip(host):
|
|
||||||
if host.startswith("[") and host.endswith("]"):
|
|
||||||
return host
|
|
||||||
try:
|
|
||||||
ipaddress.ip_address(host)
|
|
||||||
return f"[{host}]"
|
|
||||||
except ValueError:
|
|
||||||
return host
|
|
||||||
|
|
||||||
|
|
||||||
def create_newemail_dict(config: Config):
|
def create_newemail_dict(config: Config):
|
||||||
user = "".join(
|
user = "".join(random.choices(ALPHANUMERIC, k=config.username_min_length))
|
||||||
secrets.choice(ALPHANUMERIC) for _ in range(config.username_max_length)
|
|
||||||
)
|
|
||||||
password = "".join(
|
password = "".join(
|
||||||
secrets.choice(ALPHANUMERIC_PUNCT)
|
secrets.choice(ALPHANUMERIC_PUNCT)
|
||||||
for _ in range(config.password_min_length + 3)
|
for _ in range(config.password_min_length + 3)
|
||||||
)
|
)
|
||||||
return dict(email=f"{user}@{wrap_ip(config.mail_domain)}", password=f"{password}")
|
return dict(email=f"{user}@{config.mail_domain}", password=f"{password}")
|
||||||
|
|
||||||
|
|
||||||
def create_dclogin_url(email, password):
|
|
||||||
"""Build a dclogin: URL with credentials and self-signed cert acceptance.
|
|
||||||
|
|
||||||
Uses ic=3 (AcceptInvalidCertificates) so chatmail clients
|
|
||||||
can connect to servers with self-signed TLS certificates.
|
|
||||||
"""
|
|
||||||
return f"dclogin:{quote(email, safe='@')}?p={quote(password, safe='')}&v=1&ic=3"
|
|
||||||
|
|
||||||
|
|
||||||
def print_new_account():
|
def print_new_account():
|
||||||
config = read_config(CONFIG_PATH)
|
config = read_config(CONFIG_PATH)
|
||||||
creds = create_newemail_dict(config)
|
creds = create_newemail_dict(config)
|
||||||
|
|
||||||
result = dict(email=creds["email"], password=creds["password"])
|
|
||||||
if config.tls_cert_mode == "self":
|
|
||||||
result["dclogin_url"] = create_dclogin_url(creds["email"], creds["password"])
|
|
||||||
|
|
||||||
print("Content-Type: application/json")
|
print("Content-Type: application/json")
|
||||||
print("")
|
print("")
|
||||||
print(json.dumps(result))
|
print(json.dumps(creds))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
|||||||
@@ -17,11 +17,11 @@ and which are scheduled for retry using exponential back-off timing.
|
|||||||
If a token notification would be scheduled more than DROP_DEADLINE seconds
|
If a token notification would be scheduled more than DROP_DEADLINE seconds
|
||||||
after its first attempt, it is dropped with a log error.
|
after its first attempt, it is dropped with a log error.
|
||||||
|
|
||||||
Note that tokens are opaque to the notification machinery here
|
Note that tokens are completely opaque to the notification machinery here
|
||||||
and are encrypted foreclosing all ability to distinguish
|
and will in the future be encrypted foreclosing all ability to distinguish
|
||||||
which device token ultimately goes to which phone-provider notification service,
|
which device token ultimately goes to which phone-provider notification service,
|
||||||
or to understand the relation of "device tokens" and chatmail addresses.
|
or to understand the relation of "device tokens" and chatmail addresses.
|
||||||
The meaning and format of tokens is basically a matter of chatmail Core and
|
The meaning and format of tokens is basically a matter of Delta-Chat Core and
|
||||||
the `notification.delta.chat` service.
|
the `notification.delta.chat` service.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@@ -95,12 +95,7 @@ class Notifier:
|
|||||||
logging.warning(f"removing spurious queue item: {queue_path!r}")
|
logging.warning(f"removing spurious queue item: {queue_path!r}")
|
||||||
queue_path.unlink()
|
queue_path.unlink()
|
||||||
continue
|
continue
|
||||||
try:
|
|
||||||
queue_item = PersistentQueueItem.read_from_path(queue_path)
|
queue_item = PersistentQueueItem.read_from_path(queue_path)
|
||||||
except ValueError:
|
|
||||||
logging.warning(f"removing spurious queue item: {queue_path!r}")
|
|
||||||
queue_path.unlink()
|
|
||||||
continue
|
|
||||||
self.queue_for_retry(queue_item)
|
self.queue_for_retry(queue_item)
|
||||||
|
|
||||||
def queue_for_retry(self, queue_item, retry_num=0):
|
def queue_for_retry(self, queue_item, retry_num=0):
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from chatmaild.config import parse_size_mb, read_config
|
from chatmaild.config import read_config
|
||||||
|
|
||||||
|
|
||||||
def test_read_config_basic(example_config):
|
def test_read_config_basic(example_config):
|
||||||
@@ -33,9 +33,8 @@ def test_read_config_testrun(make_config):
|
|||||||
assert config.filtermail_smtp_port == 10080
|
assert config.filtermail_smtp_port == 10080
|
||||||
assert config.postfix_reinject_port == 10025
|
assert config.postfix_reinject_port == 10025
|
||||||
assert config.max_user_send_per_minute == 60
|
assert config.max_user_send_per_minute == 60
|
||||||
assert config.max_mailbox_size == "500M"
|
assert config.max_mailbox_size == "100M"
|
||||||
assert config.delete_mails_after == "20"
|
assert config.delete_mails_after == "20"
|
||||||
assert config.delete_large_after == "7"
|
|
||||||
assert config.username_min_length == 9
|
assert config.username_min_length == 9
|
||||||
assert config.username_max_length == 9
|
assert config.username_max_length == 9
|
||||||
assert config.password_min_length == 9
|
assert config.password_min_length == 9
|
||||||
@@ -73,65 +72,3 @@ def test_config_userstate_paths(make_config, tmp_path):
|
|||||||
def test_config_max_message_size(make_config, tmp_path):
|
def test_config_max_message_size(make_config, tmp_path):
|
||||||
config = make_config("something.testrun.org", dict(max_message_size="10000"))
|
config = make_config("something.testrun.org", dict(max_message_size="10000"))
|
||||||
assert config.max_message_size == 10000
|
assert config.max_message_size == 10000
|
||||||
|
|
||||||
|
|
||||||
def test_config_tls_default_acme(make_config):
|
|
||||||
config = make_config("chat.example.org")
|
|
||||||
assert config.tls_cert_mode == "acme"
|
|
||||||
assert config.tls_cert_path == "/var/lib/acme/live/chat.example.org/fullchain"
|
|
||||||
assert config.tls_key_path == "/var/lib/acme/live/chat.example.org/privkey"
|
|
||||||
|
|
||||||
|
|
||||||
def test_config_tls_self(make_config):
|
|
||||||
config = make_config("_test.example.org")
|
|
||||||
assert config.tls_cert_mode == "self"
|
|
||||||
assert config.tls_cert_path == "/etc/ssl/certs/mailserver.pem"
|
|
||||||
assert config.tls_key_path == "/etc/ssl/private/mailserver.key"
|
|
||||||
|
|
||||||
|
|
||||||
def test_config_tls_external(make_config):
|
|
||||||
config = make_config(
|
|
||||||
"chat.example.org",
|
|
||||||
{
|
|
||||||
"tls_external_cert_and_key": "/custom/fullchain.pem /custom/privkey.pem",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
assert config.tls_cert_mode == "external"
|
|
||||||
assert config.tls_cert_path == "/custom/fullchain.pem"
|
|
||||||
assert config.tls_key_path == "/custom/privkey.pem"
|
|
||||||
|
|
||||||
|
|
||||||
def test_config_tls_external_overrides_underscore(make_config):
|
|
||||||
config = make_config(
|
|
||||||
"_test.example.org",
|
|
||||||
{
|
|
||||||
"tls_external_cert_and_key": "/certs/fullchain.pem /certs/privkey.pem",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
assert config.tls_cert_mode == "external"
|
|
||||||
assert config.tls_cert_path == "/certs/fullchain.pem"
|
|
||||||
assert config.tls_key_path == "/certs/privkey.pem"
|
|
||||||
|
|
||||||
|
|
||||||
def test_config_tls_external_bad_format(make_config):
|
|
||||||
with pytest.raises(ValueError, match="two space-separated"):
|
|
||||||
make_config(
|
|
||||||
"chat.example.org",
|
|
||||||
{
|
|
||||||
"tls_external_cert_and_key": "/only/one/path.pem",
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def test_parse_size_mb():
|
|
||||||
assert parse_size_mb("500M") == 500
|
|
||||||
assert parse_size_mb("2G") == 2048
|
|
||||||
assert parse_size_mb(" 1g ") == 1024
|
|
||||||
assert parse_size_mb("100MB") == 100
|
|
||||||
assert parse_size_mb("256") == 256
|
|
||||||
|
|
||||||
|
|
||||||
def test_max_mailbox_size_mb(make_config):
|
|
||||||
config = make_config("chat.example.org")
|
|
||||||
assert config.max_mailbox_size == "500M"
|
|
||||||
assert config.max_mailbox_size_mb == 500
|
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
import time
|
import time
|
||||||
|
|
||||||
|
from chatmaild.delete_inactive_users import delete_inactive_users
|
||||||
from chatmaild.doveauth import AuthDictProxy
|
from chatmaild.doveauth import AuthDictProxy
|
||||||
from chatmaild.expire import daily_expire_main as main_expire
|
|
||||||
|
|
||||||
|
|
||||||
def test_login_timestamps(example_config):
|
def test_login_timestamps(example_config):
|
||||||
@@ -45,12 +45,7 @@ def test_delete_inactive_users(example_config):
|
|||||||
for addr in to_remove:
|
for addr in to_remove:
|
||||||
assert example_config.get_user(addr).maildir.exists()
|
assert example_config.get_user(addr).maildir.exists()
|
||||||
|
|
||||||
main_expire(
|
delete_inactive_users(example_config)
|
||||||
args=[
|
|
||||||
"--remove",
|
|
||||||
str(example_config._inipath),
|
|
||||||
]
|
|
||||||
)
|
|
||||||
|
|
||||||
for p in example_config.mailboxes_dir.iterdir():
|
for p in example_config.mailboxes_dir.iterdir():
|
||||||
assert not p.name.startswith("old")
|
assert not p.name.startswith("old")
|
||||||
|
|||||||
@@ -120,60 +120,6 @@ def test_handle_dovecot_protocol_iterate(gencreds, example_config):
|
|||||||
assert not lines[2]
|
assert not lines[2]
|
||||||
|
|
||||||
|
|
||||||
def test_invalid_localpart_characters(make_config):
|
|
||||||
"""Test that is_allowed_to_create rejects localparts with invalid characters."""
|
|
||||||
config = make_config("chat.example.org", {"username_min_length": "3"})
|
|
||||||
password = "zequ0Aimuchoodaechik"
|
|
||||||
domain = config.mail_domain
|
|
||||||
|
|
||||||
# valid localparts
|
|
||||||
assert is_allowed_to_create(config, f"abc123@{domain}", password)
|
|
||||||
assert is_allowed_to_create(config, f"a.b-c_d@{domain}", password)
|
|
||||||
|
|
||||||
# uppercase rejected
|
|
||||||
assert not is_allowed_to_create(config, f"Abc123@{domain}", password)
|
|
||||||
assert not is_allowed_to_create(config, f"ABCDEFG@{domain}", password)
|
|
||||||
|
|
||||||
# spaces and special chars rejected
|
|
||||||
assert not is_allowed_to_create(config, f"a b cde@{domain}", password)
|
|
||||||
assert not is_allowed_to_create(config, f"abc+def@{domain}", password)
|
|
||||||
assert not is_allowed_to_create(config, f"abc!def@{domain}", password)
|
|
||||||
assert not is_allowed_to_create(config, f"ab@cdef@{domain}", password)
|
|
||||||
assert not is_allowed_to_create(config, f"abc/def@{domain}", password)
|
|
||||||
assert not is_allowed_to_create(config, f"abc\\def@{domain}", password)
|
|
||||||
|
|
||||||
|
|
||||||
def test_concurrent_creation_same_account(dictproxy):
|
|
||||||
"""Test that concurrent creation of the same account doesn't corrupt password."""
|
|
||||||
addr = "racetest1@chat.example.org"
|
|
||||||
password = "zequ0Aimuchoodaechik"
|
|
||||||
num_threads = 10
|
|
||||||
results = queue.Queue()
|
|
||||||
|
|
||||||
def create():
|
|
||||||
try:
|
|
||||||
res = dictproxy.lookup_passdb(addr, password)
|
|
||||||
results.put(("ok", res))
|
|
||||||
except Exception:
|
|
||||||
results.put(("err", traceback.format_exc()))
|
|
||||||
|
|
||||||
threads = [threading.Thread(target=create, daemon=True) for _ in range(num_threads)]
|
|
||||||
for t in threads:
|
|
||||||
t.start()
|
|
||||||
for t in threads:
|
|
||||||
t.join(timeout=10)
|
|
||||||
|
|
||||||
passwords_seen = set()
|
|
||||||
for _ in range(num_threads):
|
|
||||||
status, res = results.get()
|
|
||||||
if status == "err":
|
|
||||||
pytest.fail(f"concurrent creation failed\n{res}")
|
|
||||||
passwords_seen.add(res["password"])
|
|
||||||
|
|
||||||
# all threads must see the same password hash
|
|
||||||
assert len(passwords_seen) == 1
|
|
||||||
|
|
||||||
|
|
||||||
def test_50_concurrent_lookups_different_accounts(gencreds, dictproxy):
|
def test_50_concurrent_lookups_different_accounts(gencreds, dictproxy):
|
||||||
num_threads = 50
|
num_threads = 50
|
||||||
req_per_thread = 5
|
req_per_thread = 5
|
||||||
|
|||||||
@@ -1,254 +0,0 @@
|
|||||||
import itertools
|
|
||||||
import os
|
|
||||||
import random
|
|
||||||
import time
|
|
||||||
from datetime import datetime
|
|
||||||
from fnmatch import fnmatch
|
|
||||||
from pathlib import Path
|
|
||||||
|
|
||||||
import pytest
|
|
||||||
|
|
||||||
from chatmaild.expire import (
|
|
||||||
FileEntry,
|
|
||||||
MailboxStat,
|
|
||||||
expire_to_target,
|
|
||||||
get_file_entry,
|
|
||||||
iter_mailboxes,
|
|
||||||
os_listdir_if_exists,
|
|
||||||
parse_dovecot_filename,
|
|
||||||
quota_expire_main,
|
|
||||||
scan_mailbox_messages,
|
|
||||||
)
|
|
||||||
from chatmaild.expire import daily_expire_main as expiry_main
|
|
||||||
from chatmaild.fsreport import main as report_main
|
|
||||||
|
|
||||||
MB = 1024 * 1024
|
|
||||||
|
|
||||||
|
|
||||||
def fill_mbox(folderdir):
|
|
||||||
password = folderdir.joinpath("password")
|
|
||||||
password.write_text("xxx")
|
|
||||||
folderdir.joinpath("maildirsize").write_text("xxx")
|
|
||||||
|
|
||||||
garbagedir = folderdir.joinpath("garbagedir")
|
|
||||||
garbagedir.mkdir()
|
|
||||||
garbagedir.joinpath("bimbum").write_text("hello")
|
|
||||||
|
|
||||||
create_new_messages(folderdir, ["cur/msg1"], size=500)
|
|
||||||
create_new_messages(folderdir, ["new/msg2"], size=600)
|
|
||||||
|
|
||||||
|
|
||||||
def create_new_messages(basedir, relpaths, size=1000, days=0):
|
|
||||||
now = datetime.utcnow().timestamp()
|
|
||||||
|
|
||||||
for relpath in relpaths:
|
|
||||||
msg_path = Path(basedir).joinpath(relpath)
|
|
||||||
msg_path.parent.mkdir(parents=True, exist_ok=True)
|
|
||||||
msg_path.write_text("x" * size)
|
|
||||||
# accessed now, modified N days ago
|
|
||||||
os.utime(msg_path, (now, now - days * 86400))
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
|
||||||
def mbox1(example_config):
|
|
||||||
mboxdir = example_config.mailboxes_dir.joinpath("mailbox1@example.org")
|
|
||||||
mboxdir.mkdir()
|
|
||||||
fill_mbox(mboxdir)
|
|
||||||
return MailboxStat(mboxdir)
|
|
||||||
|
|
||||||
|
|
||||||
def test_deltachat_folder(example_config):
|
|
||||||
"""Test old setups that might have a .DeltaChat folder where messages also need to get removed."""
|
|
||||||
mboxdir = example_config.mailboxes_dir.joinpath("mailbox1@example.org")
|
|
||||||
mboxdir.mkdir()
|
|
||||||
mbox2dir = mboxdir.joinpath(".DeltaChat")
|
|
||||||
mbox2dir.mkdir()
|
|
||||||
fill_mbox(mbox2dir)
|
|
||||||
mb = MailboxStat(mboxdir)
|
|
||||||
assert len(mb.messages) == 2
|
|
||||||
|
|
||||||
|
|
||||||
def test_filentry_ordering(tmp_path):
|
|
||||||
l = [FileEntry(f"x{i}", size=i + 10, mtime=1000 - i) for i in range(10)]
|
|
||||||
sorted = list(l)
|
|
||||||
random.shuffle(l)
|
|
||||||
l.sort(key=lambda x: x.size)
|
|
||||||
assert l == sorted
|
|
||||||
|
|
||||||
|
|
||||||
def test_no_mailbxoes(tmp_path, capsys):
|
|
||||||
assert [] == list(iter_mailboxes(str(tmp_path.joinpath("notexists")), maxnum=10))
|
|
||||||
out, err = capsys.readouterr()
|
|
||||||
assert "no mailboxes" in err
|
|
||||||
|
|
||||||
|
|
||||||
def test_stats_mailbox(mbox1):
|
|
||||||
password = Path(mbox1.basedir).joinpath("password")
|
|
||||||
assert mbox1.last_login == password.stat().st_mtime
|
|
||||||
assert len(mbox1.messages) == 2
|
|
||||||
|
|
||||||
msgs = list(sorted(mbox1.messages, key=lambda x: x.size))
|
|
||||||
assert len(msgs) == 2
|
|
||||||
assert msgs[0].size == 500 # cur
|
|
||||||
assert msgs[1].size == 600 # new
|
|
||||||
|
|
||||||
create_new_messages(mbox1.basedir, ["large-extra"], size=1000)
|
|
||||||
create_new_messages(mbox1.basedir, ["index-something"], size=3)
|
|
||||||
mbox2 = MailboxStat(mbox1.basedir)
|
|
||||||
assert len(mbox2.extrafiles) == 5
|
|
||||||
assert mbox2.extrafiles[0].size == 1000
|
|
||||||
|
|
||||||
# cope well with mailbox dirs that have no password (for whatever reason)
|
|
||||||
Path(mbox1.basedir).joinpath("password").unlink()
|
|
||||||
mbox3 = MailboxStat(mbox1.basedir)
|
|
||||||
assert mbox3.last_login is None
|
|
||||||
|
|
||||||
|
|
||||||
def test_report_no_mailboxes(example_config):
|
|
||||||
args = (str(example_config._inipath),)
|
|
||||||
report_main(args)
|
|
||||||
|
|
||||||
|
|
||||||
def test_report(mbox1, example_config):
|
|
||||||
args = (str(example_config._inipath),)
|
|
||||||
report_main(args)
|
|
||||||
args = list(args) + "--days 1".split()
|
|
||||||
report_main(args)
|
|
||||||
args = list(args) + "--min-login-age 1".split()
|
|
||||||
report_main(args)
|
|
||||||
args = list(args) + "--mdir cur".split()
|
|
||||||
report_main(args)
|
|
||||||
|
|
||||||
|
|
||||||
def test_report_mdir_filters_by_path(mbox1, example_config):
|
|
||||||
"""Test that Report with mdir='cur' only counts messages in cur/ subdirectory."""
|
|
||||||
from chatmaild.fsreport import Report
|
|
||||||
|
|
||||||
now = datetime.utcnow().timestamp()
|
|
||||||
|
|
||||||
# Set password mtime to old enough so min_login_age check passes
|
|
||||||
password = Path(mbox1.basedir).joinpath("password")
|
|
||||||
old_time = now - 86400 * 10 # 10 days ago
|
|
||||||
os.utime(password, (old_time, old_time))
|
|
||||||
|
|
||||||
# Reload mailbox with updated mtime
|
|
||||||
from chatmaild.expire import MailboxStat
|
|
||||||
|
|
||||||
mbox = MailboxStat(mbox1.basedir)
|
|
||||||
|
|
||||||
# Report without mdir — should count all messages
|
|
||||||
rep_all = Report(now=now, min_login_age=1, mdir=None)
|
|
||||||
rep_all.process_mailbox_stat(mbox)
|
|
||||||
total_all = rep_all.message_buckets[0]
|
|
||||||
|
|
||||||
# Report with mdir='cur' — should only count cur/ messages
|
|
||||||
rep_cur = Report(now=now, min_login_age=1, mdir="cur")
|
|
||||||
rep_cur.process_mailbox_stat(mbox)
|
|
||||||
total_cur = rep_cur.message_buckets[0]
|
|
||||||
|
|
||||||
# Report with mdir='new' — should only count new/ messages
|
|
||||||
rep_new = Report(now=now, min_login_age=1, mdir="new")
|
|
||||||
rep_new.process_mailbox_stat(mbox)
|
|
||||||
total_new = rep_new.message_buckets[0]
|
|
||||||
|
|
||||||
# cur has 500-byte msg, new has 600-byte msg (from fill_mbox)
|
|
||||||
assert total_cur == 500
|
|
||||||
assert total_new == 600
|
|
||||||
assert total_all == 500 + 600
|
|
||||||
|
|
||||||
|
|
||||||
def test_expiry_cli_basic(example_config, mbox1):
|
|
||||||
args = (str(example_config._inipath),)
|
|
||||||
expiry_main(args)
|
|
||||||
|
|
||||||
|
|
||||||
def test_expiry_cli_old_files(capsys, example_config, mbox1):
|
|
||||||
relpaths_old = ["cur/msg_old1", "cur/msg_old1"]
|
|
||||||
cutoff_days = int(example_config.delete_mails_after) + 1
|
|
||||||
create_new_messages(mbox1.basedir, relpaths_old, size=1000, days=cutoff_days)
|
|
||||||
|
|
||||||
relpaths_large = ["cur/msg_old_large1", "new/msg_old_large2"]
|
|
||||||
cutoff_days = int(example_config.delete_large_after) + 1
|
|
||||||
create_new_messages(
|
|
||||||
mbox1.basedir, relpaths_large, size=1000 * 300, days=cutoff_days
|
|
||||||
)
|
|
||||||
|
|
||||||
create_new_messages(mbox1.basedir, ["cur/shouldstay"], size=1000 * 300, days=1)
|
|
||||||
|
|
||||||
args = str(example_config._inipath), "--remove", "-v"
|
|
||||||
expiry_main(args)
|
|
||||||
out, err = capsys.readouterr()
|
|
||||||
|
|
||||||
allpaths = relpaths_old + relpaths_large + ["maildirsize"]
|
|
||||||
for path in allpaths:
|
|
||||||
for line in err.split("\n"):
|
|
||||||
if fnmatch(line, f"removing*{path}"):
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
if path != "new/msg_old_large2":
|
|
||||||
pytest.fail(f"failed to remove {path}\n{err}")
|
|
||||||
|
|
||||||
assert "shouldstay" not in err
|
|
||||||
|
|
||||||
|
|
||||||
def test_get_file_entry(tmp_path):
|
|
||||||
assert get_file_entry(str(tmp_path.joinpath("123123"))) is None
|
|
||||||
p = tmp_path.joinpath("x")
|
|
||||||
p.write_text("hello")
|
|
||||||
entry = get_file_entry(str(p))
|
|
||||||
assert entry.size == 5
|
|
||||||
assert entry.mtime
|
|
||||||
|
|
||||||
|
|
||||||
def test_os_listdir_if_exists(tmp_path):
|
|
||||||
tmp_path.joinpath("x").write_text("hello")
|
|
||||||
assert len(os_listdir_if_exists(str(tmp_path))) == 1
|
|
||||||
assert len(os_listdir_if_exists(str(tmp_path.joinpath("123123")))) == 0
|
|
||||||
|
|
||||||
|
|
||||||
# --- quota expire tests ---
|
|
||||||
|
|
||||||
_msg_counter = itertools.count(1)
|
|
||||||
|
|
||||||
|
|
||||||
def _create_message(basedir, sub, size, days_old=0, disk_size=None):
|
|
||||||
seq = next(_msg_counter)
|
|
||||||
mtime = int(time.time() - days_old * 86400)
|
|
||||||
name = f"{mtime}.M1P1Q{seq}.hostname,S={size},W={size}:2,S"
|
|
||||||
path = basedir / sub / name
|
|
||||||
path.parent.mkdir(parents=True, exist_ok=True)
|
|
||||||
path.write_bytes(b"x" * (disk_size if disk_size is not None else size))
|
|
||||||
os.utime(path, (mtime, mtime))
|
|
||||||
return path
|
|
||||||
|
|
||||||
|
|
||||||
def test_parse_dovecot_filename():
|
|
||||||
e = parse_dovecot_filename("cur/1775324677.M448978P3029757.exam,S=3235,W=3305:2,S")
|
|
||||||
assert e.path == "cur/1775324677.M448978P3029757.exam,S=3235,W=3305:2,S"
|
|
||||||
assert e.mtime == 1775324677
|
|
||||||
assert e.quota_size == 3235
|
|
||||||
assert parse_dovecot_filename("cur/msg_without_structure") is None
|
|
||||||
|
|
||||||
|
|
||||||
def test_expire_to_target(tmp_path):
|
|
||||||
_create_message(tmp_path, "cur", MB, days_old=10, disk_size=100)
|
|
||||||
_create_message(tmp_path, "new", MB, days_old=5)
|
|
||||||
_create_message(tmp_path, "cur", MB, days_old=0) # undeletable (<1 hour)
|
|
||||||
assert len(scan_mailbox_messages(tmp_path)) == 3
|
|
||||||
# removes oldest first, uses S= size not disk size
|
|
||||||
removed = expire_to_target(tmp_path, MB)
|
|
||||||
assert removed == 2
|
|
||||||
msgs = scan_mailbox_messages(tmp_path)
|
|
||||||
assert len(msgs) == 1
|
|
||||||
# the surviving message is the fresh undeletable one
|
|
||||||
assert msgs[0].mtime > time.time() - 3600
|
|
||||||
|
|
||||||
|
|
||||||
def test_quota_expire_main(tmp_path, capsys):
|
|
||||||
mbox = tmp_path / "user@example.org"
|
|
||||||
_create_message(mbox, "cur", 2 * MB, days_old=5)
|
|
||||||
(mbox / "maildirsize").write_text("x")
|
|
||||||
quota_expire_main([str(1), str(mbox)])
|
|
||||||
_, err = capsys.readouterr()
|
|
||||||
assert "quota-expire: removed 1 message(s) from user@example.org" in err
|
|
||||||
assert not (mbox / "maildirsize").exists()
|
|
||||||
306
chatmaild/src/chatmaild/tests/test_filtermail.py
Normal file
306
chatmaild/src/chatmaild/tests/test_filtermail.py
Normal file
@@ -0,0 +1,306 @@
|
|||||||
|
import pytest
|
||||||
|
|
||||||
|
from chatmaild.filtermail import (
|
||||||
|
IncomingBeforeQueueHandler,
|
||||||
|
OutgoingBeforeQueueHandler,
|
||||||
|
SendRateLimiter,
|
||||||
|
check_armored_payload,
|
||||||
|
check_encrypted,
|
||||||
|
is_securejoin,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.fixture
|
||||||
|
def maildomain():
|
||||||
|
# let's not depend on a real chatmail instance for the offline tests below
|
||||||
|
return "chatmail.example.org"
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.fixture
|
||||||
|
def handler(make_config, maildomain):
|
||||||
|
config = make_config(maildomain)
|
||||||
|
return OutgoingBeforeQueueHandler(config)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.fixture
|
||||||
|
def inhandler(make_config, maildomain):
|
||||||
|
config = make_config(maildomain)
|
||||||
|
return IncomingBeforeQueueHandler(config)
|
||||||
|
|
||||||
|
|
||||||
|
def test_reject_forged_from(maildata, gencreds, handler):
|
||||||
|
class env:
|
||||||
|
mail_from = gencreds()[0]
|
||||||
|
rcpt_tos = [gencreds()[0]]
|
||||||
|
|
||||||
|
# test that the filter lets good mail through
|
||||||
|
to_addr = gencreds()[0]
|
||||||
|
env.content = maildata(
|
||||||
|
"encrypted.eml", from_addr=env.mail_from, to_addr=to_addr
|
||||||
|
).as_bytes()
|
||||||
|
|
||||||
|
assert not handler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
# test that the filter rejects forged mail
|
||||||
|
env.content = maildata(
|
||||||
|
"encrypted.eml", from_addr="forged@c3.testrun.org", to_addr=to_addr
|
||||||
|
).as_bytes()
|
||||||
|
error = handler.check_DATA(envelope=env)
|
||||||
|
assert "500" in error
|
||||||
|
|
||||||
|
|
||||||
|
def test_filtermail_no_encryption_detection(maildata):
|
||||||
|
msg = maildata(
|
||||||
|
"plain.eml", from_addr="some@example.org", to_addr="other@example.org"
|
||||||
|
)
|
||||||
|
assert not check_encrypted(msg)
|
||||||
|
|
||||||
|
# https://xkcd.com/1181/
|
||||||
|
msg = maildata(
|
||||||
|
"fake-encrypted.eml", from_addr="some@example.org", to_addr="other@example.org"
|
||||||
|
)
|
||||||
|
assert not check_encrypted(msg)
|
||||||
|
|
||||||
|
|
||||||
|
def test_filtermail_securejoin_detection(maildata):
|
||||||
|
msg = maildata(
|
||||||
|
"securejoin-vc.eml", from_addr="some@example.org", to_addr="other@example.org"
|
||||||
|
)
|
||||||
|
assert is_securejoin(msg)
|
||||||
|
|
||||||
|
msg = maildata(
|
||||||
|
"securejoin-vc-fake.eml",
|
||||||
|
from_addr="some@example.org",
|
||||||
|
to_addr="other@example.org",
|
||||||
|
)
|
||||||
|
assert not is_securejoin(msg)
|
||||||
|
|
||||||
|
|
||||||
|
def test_filtermail_encryption_detection(maildata):
|
||||||
|
msg = maildata(
|
||||||
|
"encrypted.eml",
|
||||||
|
from_addr="1@example.org",
|
||||||
|
to_addr="2@example.org",
|
||||||
|
subject="Subject does not matter, will be replaced anyway",
|
||||||
|
)
|
||||||
|
assert check_encrypted(msg)
|
||||||
|
|
||||||
|
|
||||||
|
def test_filtermail_no_literal_packets(maildata):
|
||||||
|
"""Test that literal OpenPGP packet is not considered an encrypted mail."""
|
||||||
|
msg = maildata("literal.eml", from_addr="1@example.org", to_addr="2@example.org")
|
||||||
|
assert not check_encrypted(msg)
|
||||||
|
|
||||||
|
|
||||||
|
def test_filtermail_unencrypted_mdn(maildata, gencreds):
|
||||||
|
"""Unencrypted MDNs should not pass."""
|
||||||
|
from_addr = gencreds()[0]
|
||||||
|
to_addr = gencreds()[0] + ".other"
|
||||||
|
msg = maildata("mdn.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
assert not check_encrypted(msg)
|
||||||
|
|
||||||
|
|
||||||
|
def test_send_rate_limiter():
|
||||||
|
limiter = SendRateLimiter()
|
||||||
|
for i in range(100):
|
||||||
|
if limiter.is_sending_allowed("some@example.org", 10):
|
||||||
|
if i <= 10:
|
||||||
|
continue
|
||||||
|
pytest.fail("limiter didn't work")
|
||||||
|
else:
|
||||||
|
assert i == 11
|
||||||
|
break
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_excempt_privacy(maildata, gencreds, handler):
|
||||||
|
from_addr = gencreds()[0]
|
||||||
|
to_addr = "privacy@testrun.org"
|
||||||
|
handler.config.passthrough_recipients = [to_addr]
|
||||||
|
false_to = "privacy@something.org"
|
||||||
|
|
||||||
|
msg = maildata("plain.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
# assert that None/no error is returned
|
||||||
|
assert not handler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
class env2:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr, false_to]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
assert "523" in handler.check_DATA(envelope=env2)
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_self_send_autocrypt_setup_message(maildata, gencreds, handler):
|
||||||
|
from_addr = gencreds()[0]
|
||||||
|
to_addr = from_addr
|
||||||
|
|
||||||
|
msg = maildata("asm.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
assert not handler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_send_fails(maildata, gencreds, handler):
|
||||||
|
from_addr = gencreds()[0]
|
||||||
|
to_addr = gencreds()[0]
|
||||||
|
|
||||||
|
msg = maildata("plain.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
res = handler.check_DATA(envelope=env)
|
||||||
|
assert "523 Encryption Needed" in res
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_incoming_fails(maildata, gencreds, inhandler):
|
||||||
|
from_addr = gencreds()[0]
|
||||||
|
to_addr, password = gencreds()
|
||||||
|
|
||||||
|
msg = maildata("plain.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
user = inhandler.config.get_user(to_addr)
|
||||||
|
user.set_password(password)
|
||||||
|
res = inhandler.check_DATA(envelope=env)
|
||||||
|
assert "523 Encryption Needed" in res
|
||||||
|
|
||||||
|
user.allow_incoming_cleartext()
|
||||||
|
assert not inhandler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_incoming_mailer_daemon(maildata, gencreds, inhandler):
|
||||||
|
from_addr = "mailer-daemon@example.org"
|
||||||
|
to_addr = gencreds()[0]
|
||||||
|
|
||||||
|
msg = maildata("mailer-daemon.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
assert not inhandler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_passthrough_domains(maildata, gencreds, handler):
|
||||||
|
from_addr = gencreds()[0]
|
||||||
|
to_addr = "privacy@x.y.z"
|
||||||
|
handler.config.passthrough_recipients = ["@x.y.z"]
|
||||||
|
false_to = "something@x.y"
|
||||||
|
|
||||||
|
msg = maildata("plain.eml", from_addr=from_addr, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
# assert that None/no error is returned
|
||||||
|
assert not handler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
class env2:
|
||||||
|
mail_from = from_addr
|
||||||
|
rcpt_tos = [to_addr, false_to]
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
assert "523" in handler.check_DATA(envelope=env2)
|
||||||
|
|
||||||
|
|
||||||
|
def test_cleartext_passthrough_senders(gencreds, handler, maildata):
|
||||||
|
acc1 = gencreds()[0]
|
||||||
|
to_addr = "recipient@something.org"
|
||||||
|
handler.config.passthrough_senders = [acc1]
|
||||||
|
|
||||||
|
msg = maildata("plain.eml", from_addr=acc1, to_addr=to_addr)
|
||||||
|
|
||||||
|
class env:
|
||||||
|
mail_from = acc1
|
||||||
|
rcpt_tos = to_addr
|
||||||
|
content = msg.as_bytes()
|
||||||
|
|
||||||
|
# assert that None/no error is returned
|
||||||
|
assert not handler.check_DATA(envelope=env)
|
||||||
|
|
||||||
|
|
||||||
|
def test_check_armored_payload():
|
||||||
|
payload = """-----BEGIN PGP MESSAGE-----\r
|
||||||
|
\r
|
||||||
|
wU4DSqFx0d1yqAoSAQdAYkX/ZN/Az4B0k7X47zKyWrXxlDEdS3WOy0Yf2+GJTFgg\r
|
||||||
|
Zk5ql0mLG8Ze+ZifCS0XMO4otlemSyJ0K1ZPdFMGzUDBTgNqzkFabxXoXRIBB0AM\r
|
||||||
|
755wlX41X6Ay3KhnwBq7yEqSykVH6F3x11iHPKraLCAGZoaS8bKKNy/zg5slda1X\r
|
||||||
|
pt14b4aC1VwtSnYhcRRELNLD/wE2TFif+g7poMmFY50VyMPLYjVP96Z5QCT4+z4H\r
|
||||||
|
Ikh/pRRN8S3JNMrRJHc6prooSJmLcx47Y5un7VFy390MsJ+LiUJuQMDdYWRAinfs\r
|
||||||
|
Ebm89Ezjm7F03qbFPXE0X4ZNzVXS/eKO0uhJQdiov/vmbn41rNtHmNpqjaO0vi5+\r
|
||||||
|
sS9tR7yDUrIXiCUCN78eBLVioxtktsPZm5cDORbQWzv+7nmCEz9/JowCUcBVdCGn\r
|
||||||
|
1ofOaH82JCAX/cRx08pLaDNj6iolVBsi56Dd+2bGxJOZOG2AMcEyz0pXY0dOAJCD\r
|
||||||
|
iUThcQeGIdRnU3j8UBcnIEsjLu2+C+rrwMZQESMWKnJ0rnqTk0pK5kXScr6F/L0L\r
|
||||||
|
UE49ccIexNm3xZvYr5drszr6wz3Tv5fdue87P4etBt90gF/Vzknck+g1LLlkzZkp\r
|
||||||
|
d8dI0k2tOSPjUbDPnSy1x+X73WGpPZmj0kWT+RGvq0nH6UkJj3AQTG2qf1T8jK+3\r
|
||||||
|
rTp3LR9vDkMwDjX4R8SA9c0wdnUzzr79OYQC9lTnzcx+fM6BBmgQ2GrS33jaFLp7\r
|
||||||
|
L6/DFpCl5zhnPjM/2dKvMkw/Kd6XS/vjwsO405FQdjSDiQEEAZA+ZvAfcjdccbbU\r
|
||||||
|
yCO+x0QNdeBsufDVnh3xvzuWy4CICdTQT4s1AWRPCzjOj+SGmx5WqCLWfsd8Ma0+\r
|
||||||
|
w/C7SfTYu1FDQILLM+llpq1M/9GPley4QZ8JQjo262AyPXsPF/OW48uuZz0Db1xT\r
|
||||||
|
Yh4iHBztj4VSdy7l2+IyaIf7cnL4EEBFxv/MwmVDXvDlxyvfAfIsd3D9SvJESzKZ\r
|
||||||
|
VWDYwaocgeCN+ojKu1p885lu1EfRbX3fr3YO02K5/c2JYDkc0Py0W3wUP/J1XUax\r
|
||||||
|
pbKpzwlkxEgtmzsGqsOfMJqBV3TNDrOA2uBsa+uBqP5MGYLZ49S/4v/bW9I01Cr1\r
|
||||||
|
D2ZkV510Y1Vgo66WlP8mRqOTyt/5WRhPD+MxXdk67BNN/PmO6tMlVoJDuk+XwWPR\r
|
||||||
|
t2TvNaND/yabT9eYI55Og4fzKD6RIjouUX8DvKLkm+7aXxVs2uuLQ3Jco3O82z55\r
|
||||||
|
dbShU1jYsrw9oouXUz06MHPbkdhNbF/2hfhZ2qA31sNeovJw65iUv7sDKX3LVWgJ\r
|
||||||
|
10jlywcDwqlU8CO7WC9lGixYTbnOkYZpXCGEl8e6Jbs79l42YFo4ogYpFK1NXFhV\r
|
||||||
|
kOXRmDf/wmfj+c/ld3L2PkvwlgofhCudOQknZbo3ub1gjiTn7L+lMGHIj/3suMIl\r
|
||||||
|
ID4EUxAXScIM1ZEz2fjtW5jATlqYcLjLTbf/olw6HFyPNH+9IssqXeZNKnGwPUB9\r
|
||||||
|
3lTXsg0tpzl+x7F/2WjEw1DSNhjC0KnHt1vEYNMkUGDGFdN9y3ERLqX/FIgiASUb\r
|
||||||
|
bTvAVupnAK3raBezGmhrs6LsQtLS9P0VvQiLU3uDhMqw8Z4SISLpcD+NnVBHzQqm\r
|
||||||
|
6W5Qn/8xsCL6av18yUVTi2G3igt3QCNoYx9evt2ZcIkNoyyagUVjfZe5GHXh8Dnz\r
|
||||||
|
GaBXW/hg3HlXLRGaQu4RYCzBMJILcO25OhZOg6jbkCLiEexQlm2e9krB5cXR49Al\r
|
||||||
|
UN4fiB0KR9JyG2ayUdNJVkXZSZLnHyRgiaadlpUo16LVvw==\r
|
||||||
|
=b5Kp\r
|
||||||
|
-----END PGP MESSAGE-----\r
|
||||||
|
\r
|
||||||
|
\r
|
||||||
|
"""
|
||||||
|
|
||||||
|
assert check_armored_payload(payload) == True
|
||||||
|
|
||||||
|
payload = payload.removesuffix("\r\n")
|
||||||
|
assert check_armored_payload(payload) == True
|
||||||
|
|
||||||
|
payload = payload.removesuffix("\r\n")
|
||||||
|
assert check_armored_payload(payload) == True
|
||||||
|
|
||||||
|
payload = payload.removesuffix("\r\n")
|
||||||
|
assert check_armored_payload(payload) == True
|
||||||
|
|
||||||
|
payload = """-----BEGIN PGP MESSAGE-----\r
|
||||||
|
\r
|
||||||
|
HELLOWORLD
|
||||||
|
-----END PGP MESSAGE-----\r
|
||||||
|
\r
|
||||||
|
"""
|
||||||
|
assert check_armored_payload(payload) == False
|
||||||
|
|
||||||
|
payload = """-----BEGIN PGP MESSAGE-----\r
|
||||||
|
\r
|
||||||
|
=njUN
|
||||||
|
-----END PGP MESSAGE-----\r
|
||||||
|
\r
|
||||||
|
"""
|
||||||
|
assert check_armored_payload(payload) == False
|
||||||
@@ -1,90 +0,0 @@
|
|||||||
import shutil
|
|
||||||
import smtplib
|
|
||||||
import subprocess
|
|
||||||
import sys
|
|
||||||
|
|
||||||
import pytest
|
|
||||||
|
|
||||||
pytestmark = pytest.mark.skipif(
|
|
||||||
shutil.which("filtermail") is None,
|
|
||||||
reason="filtermail binary not found",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
|
||||||
def smtpserver():
|
|
||||||
from pytest_localserver import smtp
|
|
||||||
|
|
||||||
server = smtp.Server("127.0.0.1")
|
|
||||||
server.start()
|
|
||||||
yield server
|
|
||||||
server.stop()
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
|
||||||
def make_popen(request):
|
|
||||||
def popen(cmdargs, stdout=subprocess.PIPE, stderr=subprocess.PIPE, **kw):
|
|
||||||
p = subprocess.Popen(
|
|
||||||
cmdargs,
|
|
||||||
stdout=subprocess.PIPE,
|
|
||||||
stderr=subprocess.PIPE,
|
|
||||||
)
|
|
||||||
|
|
||||||
def fin():
|
|
||||||
p.terminate()
|
|
||||||
out, err = p.communicate()
|
|
||||||
print(out.decode("ascii"))
|
|
||||||
print(err.decode("ascii"), file=sys.stderr)
|
|
||||||
|
|
||||||
request.addfinalizer(fin)
|
|
||||||
return p
|
|
||||||
|
|
||||||
return popen
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize("filtermail_mode", ["outgoing", "incoming"])
|
|
||||||
def test_one_mail(
|
|
||||||
make_config, make_popen, smtpserver, maildata, filtermail_mode, monkeypatch
|
|
||||||
):
|
|
||||||
monkeypatch.setenv("PYTHONUNBUFFERED", "1")
|
|
||||||
# DKIM is tested by cmdeploy tests.
|
|
||||||
monkeypatch.setenv("FILTERMAIL_SKIP_DKIM", "1")
|
|
||||||
smtp_inject_port = 20025
|
|
||||||
if filtermail_mode == "outgoing":
|
|
||||||
settings = dict(
|
|
||||||
postfix_reinject_port=smtpserver.port,
|
|
||||||
filtermail_smtp_port=smtp_inject_port,
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
settings = dict(
|
|
||||||
postfix_reinject_port_incoming=smtpserver.port,
|
|
||||||
filtermail_smtp_port_incoming=smtp_inject_port,
|
|
||||||
)
|
|
||||||
|
|
||||||
config = make_config("example.org", settings=settings)
|
|
||||||
path = str(config._inipath)
|
|
||||||
|
|
||||||
popen = make_popen(["filtermail", path, filtermail_mode])
|
|
||||||
line = popen.stderr.readline().strip()
|
|
||||||
|
|
||||||
# skip a warning that FILTERMAIL_SKIP_DKIM shouldn't be used in prod
|
|
||||||
if b"DKIM verification DISABLED!" in line:
|
|
||||||
line = popen.stderr.readline().strip()
|
|
||||||
if b"loop" not in line:
|
|
||||||
print(line.decode("ascii"), file=sys.stderr)
|
|
||||||
pytest.fail("starting filtermail failed")
|
|
||||||
|
|
||||||
addr = f"user1@{config.mail_domain}"
|
|
||||||
config.get_user(addr).set_password("l1k2j3l1k2j3l")
|
|
||||||
|
|
||||||
# send encrypted mail
|
|
||||||
data = str(maildata("encrypted.eml", from_addr=addr, to_addr=addr))
|
|
||||||
client = smtplib.SMTP("localhost", smtp_inject_port)
|
|
||||||
client.sendmail(addr, [addr], data)
|
|
||||||
assert len(smtpserver.outbox) == 1
|
|
||||||
|
|
||||||
# send un-encrypted mail that errors
|
|
||||||
data = str(maildata("fake-encrypted.eml", from_addr=addr, to_addr=addr))
|
|
||||||
with pytest.raises(smtplib.SMTPDataError) as e:
|
|
||||||
client.sendmail(addr, [addr], data)
|
|
||||||
assert e.value.smtp_code == 523
|
|
||||||
@@ -36,3 +36,29 @@ def test_handle_dovecot_request_last_login(testaddr, example_config):
|
|||||||
res = dictproxy.handle_dovecot_request(msg, dictproxy_transactions)
|
res = dictproxy.handle_dovecot_request(msg, dictproxy_transactions)
|
||||||
assert res == "O\n"
|
assert res == "O\n"
|
||||||
assert len(dictproxy_transactions) == 0
|
assert len(dictproxy_transactions) == 0
|
||||||
|
|
||||||
|
|
||||||
|
def test_handle_dovecot_request_last_login_echobot(example_config):
|
||||||
|
dictproxy = LastLoginDictProxy(config=example_config)
|
||||||
|
|
||||||
|
authproxy = AuthDictProxy(config=example_config)
|
||||||
|
testaddr = f"echo@{example_config.mail_domain}"
|
||||||
|
authproxy.lookup_passdb(testaddr, "ignore")
|
||||||
|
user = dictproxy.config.get_user(testaddr)
|
||||||
|
|
||||||
|
transactions = {}
|
||||||
|
|
||||||
|
# set last-login info for user
|
||||||
|
tx = "1111"
|
||||||
|
msg = f"B{tx}\t{testaddr}"
|
||||||
|
res = dictproxy.handle_dovecot_request(msg, transactions)
|
||||||
|
assert not res
|
||||||
|
assert transactions == {tx: dict(addr=testaddr, res="O\n")}
|
||||||
|
|
||||||
|
timestamp = int(time.time())
|
||||||
|
msg = f"S{tx}\tshared/last-login/{testaddr}\t{timestamp}"
|
||||||
|
res = dictproxy.handle_dovecot_request(msg, transactions)
|
||||||
|
assert not res
|
||||||
|
assert len(transactions) == 1
|
||||||
|
read_timestamp = user.get_last_login_timestamp()
|
||||||
|
assert read_timestamp is None
|
||||||
|
|||||||
@@ -242,22 +242,6 @@ def test_requeue_removes_tmp_files(notifier, metadata, testaddr, caplog):
|
|||||||
assert queue_item.addr == testaddr
|
assert queue_item.addr == testaddr
|
||||||
|
|
||||||
|
|
||||||
def test_requeue_removes_invalid_files(notifier, metadata, testaddr, caplog):
|
|
||||||
metadata.add_token_to_addr(testaddr, "01234")
|
|
||||||
notifier.new_message_for_addr(testaddr, metadata)
|
|
||||||
# empty/invalid files should be ignored
|
|
||||||
p = notifier.queue_dir.joinpath("1203981203")
|
|
||||||
p.touch()
|
|
||||||
notifier2 = notifier.__class__(notifier.queue_dir)
|
|
||||||
notifier2.requeue_persistent_queue_items()
|
|
||||||
assert "spurious" in caplog.records[0].msg
|
|
||||||
assert not p.exists()
|
|
||||||
assert notifier2.retry_queues[0].qsize() == 1
|
|
||||||
when, queue_item = notifier2.retry_queues[0].get()
|
|
||||||
assert when <= int(time.time())
|
|
||||||
assert queue_item.addr == testaddr
|
|
||||||
|
|
||||||
|
|
||||||
def test_start_and_stop_notification_threads(notifier, testaddr):
|
def test_start_and_stop_notification_threads(notifier, testaddr):
|
||||||
threads = notifier.start_notification_threads(None)
|
threads = notifier.start_notification_threads(None)
|
||||||
for retry_num, threadlist in threads.items():
|
for retry_num, threadlist in threads.items():
|
||||||
@@ -314,51 +298,6 @@ def test_persistent_queue_items(tmp_path, testaddr, token):
|
|||||||
assert not queue_item < item2 and not item2 < queue_item
|
assert not queue_item < item2 and not item2 < queue_item
|
||||||
|
|
||||||
|
|
||||||
def test_turn_credentials_exception_returns_N(notifier, metadata, monkeypatch):
|
|
||||||
"""Test that turn_credentials() failure returns N\\n instead of crashing."""
|
|
||||||
import chatmaild.metadata
|
|
||||||
|
|
||||||
dictproxy = MetadataDictProxy(
|
|
||||||
notifier=notifier,
|
|
||||||
metadata=metadata,
|
|
||||||
turn_hostname="turn.example.org",
|
|
||||||
)
|
|
||||||
|
|
||||||
def mock_turn_credentials():
|
|
||||||
raise ConnectionRefusedError("socket not available")
|
|
||||||
|
|
||||||
monkeypatch.setattr(chatmaild.metadata, "turn_credentials", mock_turn_credentials)
|
|
||||||
|
|
||||||
transactions = {}
|
|
||||||
res = dictproxy.handle_dovecot_request(
|
|
||||||
"Lshared/0123/vendor/vendor.dovecot/pvt/server/vendor/deltachat/turn"
|
|
||||||
"\tuser@example.org",
|
|
||||||
transactions,
|
|
||||||
)
|
|
||||||
assert res == "N\n"
|
|
||||||
|
|
||||||
|
|
||||||
def test_turn_credentials_success(notifier, metadata, monkeypatch):
|
|
||||||
"""Test that valid turn_credentials() returns TURN URI."""
|
|
||||||
import chatmaild.metadata
|
|
||||||
|
|
||||||
dictproxy = MetadataDictProxy(
|
|
||||||
notifier=notifier,
|
|
||||||
metadata=metadata,
|
|
||||||
turn_hostname="turn.example.org",
|
|
||||||
)
|
|
||||||
|
|
||||||
monkeypatch.setattr(chatmaild.metadata, "turn_credentials", lambda: "user:pass")
|
|
||||||
|
|
||||||
transactions = {}
|
|
||||||
res = dictproxy.handle_dovecot_request(
|
|
||||||
"Lshared/0123/vendor/vendor.dovecot/pvt/server/vendor/deltachat/turn"
|
|
||||||
"\tuser@example.org",
|
|
||||||
transactions,
|
|
||||||
)
|
|
||||||
assert res == "Oturn.example.org:3478:user:pass\n"
|
|
||||||
|
|
||||||
|
|
||||||
def test_iroh_relay(dictproxy):
|
def test_iroh_relay(dictproxy):
|
||||||
rfile = io.BytesIO(
|
rfile = io.BytesIO(
|
||||||
b"\n".join(
|
b"\n".join(
|
||||||
|
|||||||
24
chatmaild/src/chatmaild/tests/test_metrics.py
Normal file
24
chatmaild/src/chatmaild/tests/test_metrics.py
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
from chatmaild.metrics import main
|
||||||
|
|
||||||
|
|
||||||
|
def test_main(tmp_path, capsys):
|
||||||
|
paths = []
|
||||||
|
for x in ("ci-asllkj", "ac_12l3kj", "qweqwe", "ci-l1k2j31l2k3"):
|
||||||
|
p = tmp_path.joinpath(x)
|
||||||
|
p.mkdir()
|
||||||
|
p.joinpath("cur").mkdir()
|
||||||
|
paths.append(p)
|
||||||
|
|
||||||
|
tmp_path.joinpath("nomailbox").mkdir()
|
||||||
|
|
||||||
|
main(tmp_path)
|
||||||
|
out, _ = capsys.readouterr()
|
||||||
|
d = {}
|
||||||
|
for line in out.split("\n"):
|
||||||
|
if line.strip() and not line.startswith("#"):
|
||||||
|
name, num = line.split()
|
||||||
|
d[name] = int(num)
|
||||||
|
|
||||||
|
assert d["accounts"] == 4
|
||||||
|
assert d["ci_accounts"] == 3
|
||||||
|
assert d["nonci_accounts"] == 1
|
||||||
@@ -48,8 +48,6 @@ def test_migration(tmp_path, example_config, caplog):
|
|||||||
assert passdb_path.stat().st_size > 10000
|
assert passdb_path.stat().st_size > 10000
|
||||||
|
|
||||||
example_config.passdb_path = passdb_path
|
example_config.passdb_path = passdb_path
|
||||||
# ensure logging.info records are captured regardless of global configuration
|
|
||||||
caplog.set_level("INFO")
|
|
||||||
|
|
||||||
assert not caplog.records
|
assert not caplog.records
|
||||||
|
|
||||||
|
|||||||
@@ -1,11 +1,7 @@
|
|||||||
import json
|
import json
|
||||||
|
|
||||||
import chatmaild
|
import chatmaild
|
||||||
from chatmaild.newemail import (
|
from chatmaild.newemail import create_newemail_dict, print_new_account
|
||||||
create_dclogin_url,
|
|
||||||
create_newemail_dict,
|
|
||||||
print_new_account,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def test_create_newemail_dict(example_config):
|
def test_create_newemail_dict(example_config):
|
||||||
@@ -19,24 +15,6 @@ def test_create_newemail_dict(example_config):
|
|||||||
assert ac1["password"] != ac2["password"]
|
assert ac1["password"] != ac2["password"]
|
||||||
|
|
||||||
|
|
||||||
def test_create_newemail_dict_ip(make_config):
|
|
||||||
config = make_config("1.2.3.4")
|
|
||||||
ac = create_newemail_dict(config)
|
|
||||||
assert ac["email"].endswith("@[1.2.3.4]")
|
|
||||||
|
|
||||||
|
|
||||||
def test_create_dclogin_url():
|
|
||||||
url = create_dclogin_url("user@example.org", "p@ss w+rd")
|
|
||||||
assert url.startswith("dclogin:")
|
|
||||||
assert "v=1" in url
|
|
||||||
assert "ic=3" in url
|
|
||||||
|
|
||||||
assert "user@example.org" in url
|
|
||||||
# password special chars must be encoded
|
|
||||||
assert "p%40ss" in url
|
|
||||||
assert "w%2Brd" in url
|
|
||||||
|
|
||||||
|
|
||||||
def test_print_new_account(capsys, monkeypatch, maildomain, tmpdir, example_config):
|
def test_print_new_account(capsys, monkeypatch, maildomain, tmpdir, example_config):
|
||||||
monkeypatch.setattr(chatmaild.newemail, "CONFIG_PATH", str(example_config._inipath))
|
monkeypatch.setattr(chatmaild.newemail, "CONFIG_PATH", str(example_config._inipath))
|
||||||
print_new_account()
|
print_new_account()
|
||||||
@@ -47,20 +25,3 @@ def test_print_new_account(capsys, monkeypatch, maildomain, tmpdir, example_conf
|
|||||||
dic = json.loads(lines[2])
|
dic = json.loads(lines[2])
|
||||||
assert dic["email"].endswith(f"@{example_config.mail_domain}")
|
assert dic["email"].endswith(f"@{example_config.mail_domain}")
|
||||||
assert len(dic["password"]) >= 10
|
assert len(dic["password"]) >= 10
|
||||||
# default tls_cert=acme should not include dclogin_url
|
|
||||||
assert "dclogin_url" not in dic
|
|
||||||
|
|
||||||
|
|
||||||
def test_print_new_account_self_signed(capsys, monkeypatch, make_config):
|
|
||||||
config = make_config("_test.example.org")
|
|
||||||
monkeypatch.setattr(chatmaild.newemail, "CONFIG_PATH", str(config._inipath))
|
|
||||||
print_new_account()
|
|
||||||
out, err = capsys.readouterr()
|
|
||||||
lines = out.split("\n")
|
|
||||||
dic = json.loads(lines[2])
|
|
||||||
assert "dclogin_url" in dic
|
|
||||||
url = dic["dclogin_url"]
|
|
||||||
assert url.startswith("dclogin:")
|
|
||||||
assert "ic=3" in url
|
|
||||||
|
|
||||||
assert dic["email"].split("@")[0] in url
|
|
||||||
|
|||||||
@@ -1,73 +0,0 @@
|
|||||||
import socket
|
|
||||||
import threading
|
|
||||||
import time
|
|
||||||
from unittest.mock import patch
|
|
||||||
|
|
||||||
import pytest
|
|
||||||
|
|
||||||
from chatmaild.turnserver import turn_credentials
|
|
||||||
|
|
||||||
SOCKET_PATH = "/run/chatmail-turn/turn.socket"
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
|
||||||
def turn_socket(tmp_path):
|
|
||||||
"""Create a real Unix socket server at a temp path."""
|
|
||||||
sock_path = str(tmp_path / "turn.socket")
|
|
||||||
server = socket.socket(socket.AF_UNIX, socket.SOCK_STREAM)
|
|
||||||
server.bind(sock_path)
|
|
||||||
server.listen(1)
|
|
||||||
yield sock_path, server
|
|
||||||
server.close()
|
|
||||||
|
|
||||||
|
|
||||||
def _call_turn_credentials(sock_path):
|
|
||||||
"""Call turn_credentials but connect to sock_path instead of hardcoded path."""
|
|
||||||
original_connect = socket.socket.connect
|
|
||||||
|
|
||||||
def patched_connect(self, address):
|
|
||||||
if address == SOCKET_PATH:
|
|
||||||
address = sock_path
|
|
||||||
return original_connect(self, address)
|
|
||||||
|
|
||||||
with patch.object(socket.socket, "connect", patched_connect):
|
|
||||||
return turn_credentials()
|
|
||||||
|
|
||||||
|
|
||||||
def test_turn_credentials_timeout(turn_socket):
|
|
||||||
"""Server accepts but never responds — must raise socket.timeout."""
|
|
||||||
sock_path, server = turn_socket
|
|
||||||
|
|
||||||
def accept_and_hang():
|
|
||||||
conn, _ = server.accept()
|
|
||||||
time.sleep(30)
|
|
||||||
conn.close()
|
|
||||||
|
|
||||||
t = threading.Thread(target=accept_and_hang, daemon=True)
|
|
||||||
t.start()
|
|
||||||
|
|
||||||
with pytest.raises(socket.timeout):
|
|
||||||
_call_turn_credentials(sock_path)
|
|
||||||
|
|
||||||
|
|
||||||
def test_turn_credentials_connection_refused(tmp_path):
|
|
||||||
"""Socket file doesn't exist — must raise ConnectionRefusedError or FileNotFoundError."""
|
|
||||||
missing = str(tmp_path / "nonexistent.socket")
|
|
||||||
with pytest.raises((ConnectionRefusedError, FileNotFoundError)):
|
|
||||||
_call_turn_credentials(missing)
|
|
||||||
|
|
||||||
|
|
||||||
def test_turn_credentials_success(turn_socket):
|
|
||||||
"""Server responds with credentials — must return stripped string."""
|
|
||||||
sock_path, server = turn_socket
|
|
||||||
|
|
||||||
def respond():
|
|
||||||
conn, _ = server.accept()
|
|
||||||
conn.sendall(b"testuser:testpass\n")
|
|
||||||
conn.close()
|
|
||||||
|
|
||||||
t = threading.Thread(target=respond, daemon=True)
|
|
||||||
t.start()
|
|
||||||
|
|
||||||
result = _call_turn_credentials(sock_path)
|
|
||||||
assert result == "testuser:testpass"
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
#!/usr/bin/env python3
|
|
||||||
import socket
|
|
||||||
|
|
||||||
|
|
||||||
def turn_credentials() -> str:
|
|
||||||
with socket.socket(socket.AF_UNIX, socket.SOCK_STREAM) as client_socket:
|
|
||||||
client_socket.settimeout(5)
|
|
||||||
client_socket.connect("/run/chatmail-turn/turn.socket")
|
|
||||||
with client_socket.makefile("rb") as file:
|
|
||||||
return file.readline().decode("utf-8").strip()
|
|
||||||
@@ -19,7 +19,7 @@ class User:
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def can_track(self):
|
def can_track(self):
|
||||||
return "@" in self.addr
|
return "@" in self.addr and not self.addr.startswith("echo@")
|
||||||
|
|
||||||
def get_userdb_dict(self):
|
def get_userdb_dict(self):
|
||||||
"""Return a non-empty dovecot 'userdb' style dict
|
"""Return a non-empty dovecot 'userdb' style dict
|
||||||
@@ -55,6 +55,7 @@ class User:
|
|||||||
try:
|
try:
|
||||||
write_bytes_atomic(self.password_path, password)
|
write_bytes_atomic(self.password_path, password)
|
||||||
except PermissionError:
|
except PermissionError:
|
||||||
|
if not self.addr.startswith("echo@"):
|
||||||
logging.error(f"could not write password for: {self.addr}")
|
logging.error(f"could not write password for: {self.addr}")
|
||||||
raise
|
raise
|
||||||
self.enforce_E2EE_path.touch()
|
self.enforce_E2EE_path.touch()
|
||||||
|
|||||||
94
cliff.toml
94
cliff.toml
@@ -1,94 +0,0 @@
|
|||||||
# git-cliff ~ configuration file
|
|
||||||
# https://git-cliff.org/docs/configuration
|
|
||||||
|
|
||||||
|
|
||||||
[changelog]
|
|
||||||
# A Tera template to be rendered for each release in the changelog.
|
|
||||||
# See https://keats.github.io/tera/docs/#introduction
|
|
||||||
body = """
|
|
||||||
{% if version %}\
|
|
||||||
## [{{ version | trim_start_matches(pat="v") }}] - {{ timestamp | date(format="%Y-%m-%d") }}
|
|
||||||
{% else %}\
|
|
||||||
## [unreleased]
|
|
||||||
{% endif %}\
|
|
||||||
{% for group, commits in commits | group_by(attribute="group") %}
|
|
||||||
### {{ group | striptags | trim | upper_first }}
|
|
||||||
{% for commit in commits %}
|
|
||||||
- {% if commit.scope %}*({{ commit.scope }})* {% endif %}\
|
|
||||||
{% if commit.breaking %}[**breaking**] {% endif %}\
|
|
||||||
{{ commit.message | upper_first }}\
|
|
||||||
{% endfor %}
|
|
||||||
{% endfor %}
|
|
||||||
"""
|
|
||||||
# Remove leading and trailing whitespaces from the changelog's body.
|
|
||||||
trim = true
|
|
||||||
# Render body even when there are no releases to process.
|
|
||||||
render_always = true
|
|
||||||
# An array of regex based postprocessors to modify the changelog.
|
|
||||||
postprocessors = [
|
|
||||||
# Replace the placeholder <REPO> with a URL.
|
|
||||||
#{ pattern = '<REPO>', replace = "https://github.com/orhun/git-cliff" },
|
|
||||||
]
|
|
||||||
# render body even when there are no releases to process
|
|
||||||
# render_always = true
|
|
||||||
# output file path
|
|
||||||
# output = "test.md"
|
|
||||||
|
|
||||||
[git]
|
|
||||||
# Parse commits according to the conventional commits specification.
|
|
||||||
# See https://www.conventionalcommits.org
|
|
||||||
conventional_commits = true
|
|
||||||
# Exclude commits that do not match the conventional commits specification.
|
|
||||||
filter_unconventional = true
|
|
||||||
# Require all commits to be conventional.
|
|
||||||
# Takes precedence over filter_unconventional.
|
|
||||||
require_conventional = false
|
|
||||||
# Split commits on newlines, treating each line as an individual commit.
|
|
||||||
split_commits = false
|
|
||||||
# An array of regex based parsers to modify commit messages prior to further processing.
|
|
||||||
commit_preprocessors = [
|
|
||||||
# Replace issue numbers with link templates to be updated in `changelog.postprocessors`.
|
|
||||||
#{ pattern = '\((\w+\s)?#([0-9]+)\)', replace = "([#${2}](<REPO>/issues/${2}))"},
|
|
||||||
# Check spelling of the commit message using https://github.com/crate-ci/typos.
|
|
||||||
# If the spelling is incorrect, it will be fixed automatically.
|
|
||||||
#{ pattern = '.*', replace_command = 'typos --write-changes -' },
|
|
||||||
]
|
|
||||||
# Prevent commits that are breaking from being excluded by commit parsers.
|
|
||||||
protect_breaking_commits = false
|
|
||||||
# An array of regex based parsers for extracting data from the commit message.
|
|
||||||
# Assigns commits to groups.
|
|
||||||
# Optionally sets the commit's scope and can decide to exclude commits from further processing.
|
|
||||||
commit_parsers = [
|
|
||||||
{ message = "^feat", group = "Features" },
|
|
||||||
{ message = "^fix", group = "Bug Fixes" },
|
|
||||||
{ message = "^docs", group = "Documentation" },
|
|
||||||
{ message = "^perf", group = "Performance" },
|
|
||||||
{ message = "^refactor", group = "Refactor" },
|
|
||||||
{ message = "^style", group = "Styling" },
|
|
||||||
{ message = "^test", group = "Testing" },
|
|
||||||
{ message = "^chore\\(release\\): prepare for", skip = true },
|
|
||||||
{ message = "^chore\\(deps.*\\)", skip = true },
|
|
||||||
{ message = "^chore\\(pr\\)", skip = true },
|
|
||||||
{ message = "^chore\\(pull\\)", skip = true },
|
|
||||||
{ message = "^chore|^ci", group = "Miscellaneous Tasks" },
|
|
||||||
{ body = ".*security", group = "Security" },
|
|
||||||
{ message = "^revert", group = "Revert" },
|
|
||||||
{ message = ".*", group = "Other" },
|
|
||||||
]
|
|
||||||
# Exclude commits that are not matched by any commit parser.
|
|
||||||
filter_commits = false
|
|
||||||
# Fail on a commit that is not matched by any commit parser.
|
|
||||||
fail_on_unmatched_commit = false
|
|
||||||
# An array of link parsers for extracting external references, and turning them into URLs, using regex.
|
|
||||||
link_parsers = []
|
|
||||||
# Include only the tags that belong to the current branch.
|
|
||||||
use_branch_tags = false
|
|
||||||
# Order releases topologically instead of chronologically.
|
|
||||||
topo_order = false
|
|
||||||
# Order commits topologically instead of chronologically.
|
|
||||||
topo_order_commits = true
|
|
||||||
# Order of commits in each group/release within the changelog.
|
|
||||||
# Allowed values: newest, oldest
|
|
||||||
sort_commits = "oldest"
|
|
||||||
# Process submodules commits
|
|
||||||
recurse_submodules = false
|
|
||||||
@@ -10,6 +10,7 @@ dependencies = [
|
|||||||
"pillow",
|
"pillow",
|
||||||
"qrcode",
|
"qrcode",
|
||||||
"markdown",
|
"markdown",
|
||||||
|
"pytest",
|
||||||
"setuptools>=68",
|
"setuptools>=68",
|
||||||
"termcolor",
|
"termcolor",
|
||||||
"build",
|
"build",
|
||||||
@@ -19,8 +20,6 @@ dependencies = [
|
|||||||
"pytest-xdist",
|
"pytest-xdist",
|
||||||
"execnet",
|
"execnet",
|
||||||
"imap_tools",
|
"imap_tools",
|
||||||
"deltachat-rpc-client",
|
|
||||||
"deltachat-rpc-server",
|
|
||||||
]
|
]
|
||||||
|
|
||||||
[project.scripts]
|
[project.scripts]
|
||||||
@@ -42,6 +41,3 @@ lint.select = [
|
|||||||
"PLE", # Pylint Error
|
"PLE", # Pylint Error
|
||||||
"PLW", # Pylint Warning
|
"PLW", # Pylint Warning
|
||||||
]
|
]
|
||||||
lint.ignore = [
|
|
||||||
"PLC0415" # import-outside-top-level
|
|
||||||
]
|
|
||||||
|
|||||||
729
cmdeploy/src/cmdeploy/__init__.py
Normal file
729
cmdeploy/src/cmdeploy/__init__.py
Normal file
@@ -0,0 +1,729 @@
|
|||||||
|
"""
|
||||||
|
Chat Mail pyinfra deploy.
|
||||||
|
"""
|
||||||
|
|
||||||
|
import importlib.resources
|
||||||
|
import io
|
||||||
|
import shutil
|
||||||
|
import subprocess
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
from chatmaild.config import Config, read_config
|
||||||
|
from pyinfra import facts, host
|
||||||
|
from pyinfra.facts.files import File
|
||||||
|
from pyinfra.facts.systemd import SystemdEnabled
|
||||||
|
from pyinfra.operations import apt, files, pip, server, systemd
|
||||||
|
|
||||||
|
from .acmetool import deploy_acmetool
|
||||||
|
|
||||||
|
|
||||||
|
def _build_chatmaild(dist_dir) -> None:
|
||||||
|
dist_dir = Path(dist_dir).resolve()
|
||||||
|
if dist_dir.exists():
|
||||||
|
shutil.rmtree(dist_dir)
|
||||||
|
dist_dir.mkdir()
|
||||||
|
subprocess.check_output(
|
||||||
|
[sys.executable, "-m", "build", "-n"]
|
||||||
|
+ ["--sdist", "chatmaild", "--outdir", str(dist_dir)]
|
||||||
|
)
|
||||||
|
entries = list(dist_dir.iterdir())
|
||||||
|
assert len(entries) == 1
|
||||||
|
return entries[0]
|
||||||
|
|
||||||
|
|
||||||
|
def remove_legacy_artifacts():
|
||||||
|
# disable legacy doveauth-dictproxy.service
|
||||||
|
if host.get_fact(SystemdEnabled).get("doveauth-dictproxy.service"):
|
||||||
|
systemd.service(
|
||||||
|
name="Disable legacy doveauth-dictproxy.service",
|
||||||
|
service="doveauth-dictproxy.service",
|
||||||
|
running=False,
|
||||||
|
enabled=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def _install_remote_venv_with_chatmaild(config) -> None:
|
||||||
|
remove_legacy_artifacts()
|
||||||
|
dist_file = _build_chatmaild(dist_dir=Path("chatmaild/dist"))
|
||||||
|
remote_base_dir = "/usr/local/lib/chatmaild"
|
||||||
|
remote_dist_file = f"{remote_base_dir}/dist/{dist_file.name}"
|
||||||
|
remote_venv_dir = f"{remote_base_dir}/venv"
|
||||||
|
remote_chatmail_inipath = f"{remote_base_dir}/chatmail.ini"
|
||||||
|
root_owned = dict(user="root", group="root", mode="644")
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="apt install python3-virtualenv",
|
||||||
|
packages=["python3-virtualenv"],
|
||||||
|
)
|
||||||
|
|
||||||
|
files.put(
|
||||||
|
name="Upload chatmaild source package",
|
||||||
|
src=dist_file.open("rb"),
|
||||||
|
dest=remote_dist_file,
|
||||||
|
create_remote_dir=True,
|
||||||
|
**root_owned,
|
||||||
|
)
|
||||||
|
|
||||||
|
files.put(
|
||||||
|
name=f"Upload {remote_chatmail_inipath}",
|
||||||
|
src=config._getbytefile(),
|
||||||
|
dest=remote_chatmail_inipath,
|
||||||
|
**root_owned,
|
||||||
|
)
|
||||||
|
|
||||||
|
pip.virtualenv(
|
||||||
|
name=f"chatmaild virtualenv {remote_venv_dir}",
|
||||||
|
path=remote_venv_dir,
|
||||||
|
always_copy=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="install gcc and headers to build crypt_r source package",
|
||||||
|
packages=["gcc", "python3-dev"],
|
||||||
|
)
|
||||||
|
|
||||||
|
server.shell(
|
||||||
|
name=f"forced pip-install {dist_file.name}",
|
||||||
|
commands=[
|
||||||
|
f"{remote_venv_dir}/bin/pip install --force-reinstall {remote_dist_file}"
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
|
files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("metrics.cron.j2"),
|
||||||
|
dest="/etc/cron.d/chatmail-metrics",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config={
|
||||||
|
"mailboxes_dir": config.mailboxes_dir,
|
||||||
|
"execpath": f"{remote_venv_dir}/bin/chatmail-metrics",
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
# install systemd units
|
||||||
|
for fn in (
|
||||||
|
"doveauth",
|
||||||
|
"filtermail",
|
||||||
|
"filtermail-incoming",
|
||||||
|
"echobot",
|
||||||
|
"chatmail-metadata",
|
||||||
|
"lastlogin",
|
||||||
|
):
|
||||||
|
execpath = fn if fn != "filtermail-incoming" else "filtermail"
|
||||||
|
params = dict(
|
||||||
|
execpath=f"{remote_venv_dir}/bin/{execpath}",
|
||||||
|
config_path=remote_chatmail_inipath,
|
||||||
|
remote_venv_dir=remote_venv_dir,
|
||||||
|
mail_domain=config.mail_domain,
|
||||||
|
)
|
||||||
|
source_path = importlib.resources.files(__package__).joinpath(
|
||||||
|
"service", f"{fn}.service.f"
|
||||||
|
)
|
||||||
|
content = source_path.read_text().format(**params).encode()
|
||||||
|
|
||||||
|
files.put(
|
||||||
|
name=f"Upload {fn}.service",
|
||||||
|
src=io.BytesIO(content),
|
||||||
|
dest=f"/etc/systemd/system/{fn}.service",
|
||||||
|
**root_owned,
|
||||||
|
)
|
||||||
|
systemd.service(
|
||||||
|
name=f"Setup {fn} service",
|
||||||
|
service=f"{fn}.service",
|
||||||
|
running=True,
|
||||||
|
enabled=True,
|
||||||
|
restarted=True,
|
||||||
|
daemon_reload=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def _configure_opendkim(domain: str, dkim_selector: str = "dkim") -> bool:
|
||||||
|
"""Configures OpenDKIM"""
|
||||||
|
need_restart = False
|
||||||
|
|
||||||
|
main_config = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("opendkim/opendkim.conf"),
|
||||||
|
dest="/etc/opendkim.conf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config={"domain_name": domain, "opendkim_selector": dkim_selector},
|
||||||
|
)
|
||||||
|
need_restart |= main_config.changed
|
||||||
|
|
||||||
|
screen_script = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("opendkim/screen.lua"),
|
||||||
|
dest="/etc/opendkim/screen.lua",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= screen_script.changed
|
||||||
|
|
||||||
|
final_script = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("opendkim/final.lua"),
|
||||||
|
dest="/etc/opendkim/final.lua",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= final_script.changed
|
||||||
|
|
||||||
|
files.directory(
|
||||||
|
name="Add opendkim directory to /etc",
|
||||||
|
path="/etc/opendkim",
|
||||||
|
user="opendkim",
|
||||||
|
group="opendkim",
|
||||||
|
mode="750",
|
||||||
|
present=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
keytable = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("opendkim/KeyTable"),
|
||||||
|
dest="/etc/dkimkeys/KeyTable",
|
||||||
|
user="opendkim",
|
||||||
|
group="opendkim",
|
||||||
|
mode="644",
|
||||||
|
config={"domain_name": domain, "opendkim_selector": dkim_selector},
|
||||||
|
)
|
||||||
|
need_restart |= keytable.changed
|
||||||
|
|
||||||
|
signing_table = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("opendkim/SigningTable"),
|
||||||
|
dest="/etc/dkimkeys/SigningTable",
|
||||||
|
user="opendkim",
|
||||||
|
group="opendkim",
|
||||||
|
mode="644",
|
||||||
|
config={"domain_name": domain, "opendkim_selector": dkim_selector},
|
||||||
|
)
|
||||||
|
need_restart |= signing_table.changed
|
||||||
|
files.directory(
|
||||||
|
name="Add opendkim socket directory to /var/spool/postfix",
|
||||||
|
path="/var/spool/postfix/opendkim",
|
||||||
|
user="opendkim",
|
||||||
|
group="opendkim",
|
||||||
|
mode="750",
|
||||||
|
present=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="apt install opendkim opendkim-tools",
|
||||||
|
packages=["opendkim", "opendkim-tools"],
|
||||||
|
)
|
||||||
|
|
||||||
|
if not host.get_fact(File, f"/etc/dkimkeys/{dkim_selector}.private"):
|
||||||
|
server.shell(
|
||||||
|
name="Generate OpenDKIM domain keys",
|
||||||
|
commands=[
|
||||||
|
f"/usr/sbin/opendkim-genkey -D /etc/dkimkeys -d {domain} -s {dkim_selector}"
|
||||||
|
],
|
||||||
|
_use_su_login=True,
|
||||||
|
_su_user="opendkim",
|
||||||
|
)
|
||||||
|
|
||||||
|
service_file = files.put(
|
||||||
|
name="Configure opendkim to restart once a day",
|
||||||
|
src=importlib.resources.files(__package__).joinpath("opendkim/systemd.conf"),
|
||||||
|
dest="/etc/systemd/system/opendkim.service.d/10-prevent-memory-leak.conf",
|
||||||
|
)
|
||||||
|
need_restart |= service_file.changed
|
||||||
|
|
||||||
|
|
||||||
|
return need_restart
|
||||||
|
|
||||||
|
|
||||||
|
def _uninstall_mta_sts_daemon() -> None:
|
||||||
|
# Remove configuration.
|
||||||
|
files.file("/etc/mta-sts-daemon.yml", present=False)
|
||||||
|
|
||||||
|
files.directory("/usr/local/lib/postfix-mta-sts-resolver", present=False)
|
||||||
|
|
||||||
|
files.file("/etc/systemd/system/mta-sts-daemon.service", present=False)
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="Stop MTA-STS daemon",
|
||||||
|
service="mta-sts-daemon.service",
|
||||||
|
daemon_reload=True,
|
||||||
|
running=False,
|
||||||
|
enabled=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def _configure_postfix(config: Config, debug: bool = False) -> bool:
|
||||||
|
"""Configures Postfix SMTP server."""
|
||||||
|
need_restart = False
|
||||||
|
|
||||||
|
main_config = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("postfix/main.cf.j2"),
|
||||||
|
dest="/etc/postfix/main.cf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config=config,
|
||||||
|
disable_ipv6=config.disable_ipv6,
|
||||||
|
)
|
||||||
|
need_restart |= main_config.changed
|
||||||
|
|
||||||
|
master_config = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("postfix/master.cf.j2"),
|
||||||
|
dest="/etc/postfix/master.cf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
debug=debug,
|
||||||
|
config=config,
|
||||||
|
)
|
||||||
|
need_restart |= master_config.changed
|
||||||
|
|
||||||
|
header_cleanup = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath(
|
||||||
|
"postfix/submission_header_cleanup"
|
||||||
|
),
|
||||||
|
dest="/etc/postfix/submission_header_cleanup",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= header_cleanup.changed
|
||||||
|
|
||||||
|
# Login map that 1:1 maps email address to login.
|
||||||
|
login_map = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("postfix/login_map"),
|
||||||
|
dest="/etc/postfix/login_map",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= login_map.changed
|
||||||
|
|
||||||
|
return need_restart
|
||||||
|
|
||||||
|
|
||||||
|
def _configure_dovecot(config: Config, debug: bool = False) -> bool:
|
||||||
|
"""Configures Dovecot IMAP server."""
|
||||||
|
need_restart = False
|
||||||
|
|
||||||
|
main_config = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("dovecot/dovecot.conf.j2"),
|
||||||
|
dest="/etc/dovecot/dovecot.conf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config=config,
|
||||||
|
debug=debug,
|
||||||
|
disable_ipv6=config.disable_ipv6,
|
||||||
|
)
|
||||||
|
need_restart |= main_config.changed
|
||||||
|
auth_config = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("dovecot/auth.conf"),
|
||||||
|
dest="/etc/dovecot/auth.conf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= auth_config.changed
|
||||||
|
lua_push_notification_script = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath(
|
||||||
|
"dovecot/push_notification.lua"
|
||||||
|
),
|
||||||
|
dest="/etc/dovecot/push_notification.lua",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= lua_push_notification_script.changed
|
||||||
|
|
||||||
|
files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("dovecot/expunge.cron.j2"),
|
||||||
|
dest="/etc/cron.d/expunge",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config=config,
|
||||||
|
)
|
||||||
|
|
||||||
|
# as per https://doc.dovecot.org/configuration_manual/os/
|
||||||
|
# it is recommended to set the following inotify limits
|
||||||
|
for name in ("max_user_instances", "max_user_watches"):
|
||||||
|
key = f"fs.inotify.{name}"
|
||||||
|
server.sysctl(
|
||||||
|
name=f"Change {key}",
|
||||||
|
key=key,
|
||||||
|
value=65535,
|
||||||
|
persist=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
return need_restart
|
||||||
|
|
||||||
|
|
||||||
|
def _configure_nginx(config: Config, debug: bool = False) -> bool:
|
||||||
|
"""Configures nginx HTTP server."""
|
||||||
|
need_restart = False
|
||||||
|
|
||||||
|
main_config = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("nginx/nginx.conf.j2"),
|
||||||
|
dest="/etc/nginx/nginx.conf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config={"domain_name": config.mail_domain},
|
||||||
|
disable_ipv6=config.disable_ipv6,
|
||||||
|
)
|
||||||
|
need_restart |= main_config.changed
|
||||||
|
|
||||||
|
autoconfig = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("nginx/autoconfig.xml.j2"),
|
||||||
|
dest="/var/www/html/.well-known/autoconfig/mail/config-v1.1.xml",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config={"domain_name": config.mail_domain},
|
||||||
|
)
|
||||||
|
need_restart |= autoconfig.changed
|
||||||
|
|
||||||
|
mta_sts_config = files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("nginx/mta-sts.txt.j2"),
|
||||||
|
dest="/var/www/html/.well-known/mta-sts.txt",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
config={"domain_name": config.mail_domain},
|
||||||
|
)
|
||||||
|
need_restart |= mta_sts_config.changed
|
||||||
|
|
||||||
|
# install CGI newemail script
|
||||||
|
#
|
||||||
|
cgi_dir = "/usr/lib/cgi-bin"
|
||||||
|
files.directory(
|
||||||
|
name=f"Ensure {cgi_dir} exists",
|
||||||
|
path=cgi_dir,
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
)
|
||||||
|
|
||||||
|
files.put(
|
||||||
|
name="Upload cgi newemail.py script",
|
||||||
|
src=importlib.resources.files("chatmaild").joinpath("newemail.py").open("rb"),
|
||||||
|
dest=f"{cgi_dir}/newemail.py",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="755",
|
||||||
|
)
|
||||||
|
|
||||||
|
return need_restart
|
||||||
|
|
||||||
|
|
||||||
|
def _remove_rspamd() -> None:
|
||||||
|
"""Remove rspamd"""
|
||||||
|
apt.packages(name="Remove rspamd", packages="rspamd", present=False)
|
||||||
|
|
||||||
|
|
||||||
|
def check_config(config):
|
||||||
|
mail_domain = config.mail_domain
|
||||||
|
if mail_domain != "testrun.org" and not mail_domain.endswith(".testrun.org"):
|
||||||
|
blocked_words = "merlinux schmieder testrun.org".split()
|
||||||
|
for key in config.__dict__:
|
||||||
|
value = config.__dict__[key]
|
||||||
|
if key.startswith("privacy") and any(
|
||||||
|
x in str(value) for x in blocked_words
|
||||||
|
):
|
||||||
|
raise ValueError(
|
||||||
|
f"please set your own privacy contacts/addresses in {config._inipath}"
|
||||||
|
)
|
||||||
|
return config
|
||||||
|
|
||||||
|
|
||||||
|
def deploy_mtail(config):
|
||||||
|
apt.packages(
|
||||||
|
name="Install mtail",
|
||||||
|
packages=["mtail"],
|
||||||
|
)
|
||||||
|
|
||||||
|
# Using our own systemd unit instead of `/usr/lib/systemd/system/mtail.service`.
|
||||||
|
# This allows to read from journalctl instead of log files.
|
||||||
|
files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("mtail/mtail.service.j2"),
|
||||||
|
dest="/etc/systemd/system/mtail.service",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
address=config.mtail_address or "127.0.0.1",
|
||||||
|
port=3903,
|
||||||
|
)
|
||||||
|
|
||||||
|
mtail_conf = files.put(
|
||||||
|
name="Mtail configuration",
|
||||||
|
src=importlib.resources.files(__package__).joinpath(
|
||||||
|
"mtail/delivered_mail.mtail"
|
||||||
|
),
|
||||||
|
dest="/etc/mtail/delivered_mail.mtail",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="Start and enable mtail",
|
||||||
|
service="mtail.service",
|
||||||
|
running=bool(config.mtail_address),
|
||||||
|
enabled=bool(config.mtail_address),
|
||||||
|
restarted=mtail_conf.changed,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def deploy_iroh_relay(config) -> None:
|
||||||
|
(url, sha256sum) = {
|
||||||
|
"x86_64": (
|
||||||
|
"https://github.com/n0-computer/iroh/releases/download/v0.28.1/iroh-relay-v0.28.1-x86_64-unknown-linux-musl.tar.gz",
|
||||||
|
"2ffacf7c0622c26b67a5895ee8e07388769599f60e5f52a3bd40a3258db89b2c",
|
||||||
|
),
|
||||||
|
"aarch64": (
|
||||||
|
"https://github.com/n0-computer/iroh/releases/download/v0.28.1/iroh-relay-v0.28.1-aarch64-unknown-linux-musl.tar.gz",
|
||||||
|
"b915037bcc1ff1110cc9fcb5de4a17c00ff576fd2f568cd339b3b2d54c420dc4",
|
||||||
|
),
|
||||||
|
}[host.get_fact(facts.server.Arch)]
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Install curl",
|
||||||
|
packages=["curl"],
|
||||||
|
)
|
||||||
|
|
||||||
|
server.shell(
|
||||||
|
name="Download iroh-relay",
|
||||||
|
commands=[
|
||||||
|
f"(echo '{sha256sum} /usr/local/bin/iroh-relay' | sha256sum -c) || (curl -L {url} | gunzip | tar -x -f - ./iroh-relay -O >/usr/local/bin/iroh-relay.new && mv /usr/local/bin/iroh-relay.new /usr/local/bin/iroh-relay)",
|
||||||
|
"chmod 755 /usr/local/bin/iroh-relay",
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
|
need_restart = False
|
||||||
|
|
||||||
|
systemd_unit = files.put(
|
||||||
|
name="Upload iroh-relay systemd unit",
|
||||||
|
src=importlib.resources.files(__package__).joinpath("iroh-relay.service"),
|
||||||
|
dest="/etc/systemd/system/iroh-relay.service",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= systemd_unit.changed
|
||||||
|
|
||||||
|
iroh_config = files.put(
|
||||||
|
name="Upload iroh-relay config",
|
||||||
|
src=importlib.resources.files(__package__).joinpath("iroh-relay.toml"),
|
||||||
|
dest="/etc/iroh-relay.toml",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
need_restart |= iroh_config.changed
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="Start and enable iroh-relay",
|
||||||
|
service="iroh-relay.service",
|
||||||
|
running=True,
|
||||||
|
enabled=config.enable_iroh_relay,
|
||||||
|
restarted=need_restart,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def deploy_chatmail(config_path: Path, disable_mail: bool) -> None:
|
||||||
|
"""Deploy a chat-mail instance.
|
||||||
|
|
||||||
|
:param config_path: path to chatmail.ini
|
||||||
|
:param disable_mail: whether to disable postfix & dovecot
|
||||||
|
"""
|
||||||
|
config = read_config(config_path)
|
||||||
|
check_config(config)
|
||||||
|
mail_domain = config.mail_domain
|
||||||
|
|
||||||
|
from .www import build_webpages
|
||||||
|
|
||||||
|
server.group(name="Create vmail group", group="vmail", system=True)
|
||||||
|
server.user(name="Create vmail user", user="vmail", group="vmail", system=True)
|
||||||
|
server.group(name="Create opendkim group", group="opendkim", system=True)
|
||||||
|
server.user(
|
||||||
|
name="Create opendkim user",
|
||||||
|
user="opendkim",
|
||||||
|
groups=["opendkim"],
|
||||||
|
system=True,
|
||||||
|
)
|
||||||
|
server.user(
|
||||||
|
name="Add postfix user to opendkim group for socket access",
|
||||||
|
user="postfix",
|
||||||
|
groups=["opendkim"],
|
||||||
|
system=True,
|
||||||
|
)
|
||||||
|
server.user(name="Create echobot user", user="echobot", system=True)
|
||||||
|
server.user(name="Create iroh user", user="iroh", system=True)
|
||||||
|
|
||||||
|
# Add our OBS repository for dovecot_no_delay
|
||||||
|
files.put(
|
||||||
|
name="Add Deltachat OBS GPG key to apt keyring",
|
||||||
|
src=importlib.resources.files(__package__).joinpath("obs-home-deltachat.gpg"),
|
||||||
|
dest="/etc/apt/keyrings/obs-home-deltachat.gpg",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
|
||||||
|
files.line(
|
||||||
|
name="Add DeltaChat OBS home repository to sources.list",
|
||||||
|
path="/etc/apt/sources.list",
|
||||||
|
line="deb [signed-by=/etc/apt/keyrings/obs-home-deltachat.gpg] https://download.opensuse.org/repositories/home:/deltachat/Debian_12/ ./",
|
||||||
|
escape_regex_characters=True,
|
||||||
|
ensure_newline=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.update(name="apt update", cache_time=24 * 3600)
|
||||||
|
apt.upgrade(name="upgrade apt packages", auto_remove=True)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Install rsync",
|
||||||
|
packages=["rsync"],
|
||||||
|
)
|
||||||
|
|
||||||
|
# Run local DNS resolver `unbound`.
|
||||||
|
# `resolvconf` takes care of setting up /etc/resolv.conf
|
||||||
|
# to use 127.0.0.1 as the resolver.
|
||||||
|
apt.packages(
|
||||||
|
name="Install unbound",
|
||||||
|
packages=["unbound", "unbound-anchor", "dnsutils"],
|
||||||
|
)
|
||||||
|
server.shell(
|
||||||
|
name="Generate root keys for validating DNSSEC",
|
||||||
|
commands=[
|
||||||
|
"unbound-anchor -a /var/lib/unbound/root.key || true",
|
||||||
|
"systemctl reset-failed unbound.service",
|
||||||
|
],
|
||||||
|
)
|
||||||
|
systemd.service(
|
||||||
|
name="Start and enable unbound",
|
||||||
|
service="unbound.service",
|
||||||
|
running=True,
|
||||||
|
enabled=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
deploy_iroh_relay(config)
|
||||||
|
|
||||||
|
# Deploy acmetool to have TLS certificates.
|
||||||
|
tls_domains = [mail_domain, f"mta-sts.{mail_domain}", f"www.{mail_domain}"]
|
||||||
|
deploy_acmetool(
|
||||||
|
domains=tls_domains,
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
# required for setfacl for echobot
|
||||||
|
name="Install acl",
|
||||||
|
packages="acl",
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Install Postfix",
|
||||||
|
packages="postfix",
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Install Dovecot",
|
||||||
|
packages=["dovecot-imapd", "dovecot-lmtpd"],
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Install nginx",
|
||||||
|
packages=["nginx", "libnginx-mod-stream"],
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Install fcgiwrap",
|
||||||
|
packages=["fcgiwrap"],
|
||||||
|
)
|
||||||
|
|
||||||
|
www_path = importlib.resources.files(__package__).joinpath("../../../www").resolve()
|
||||||
|
|
||||||
|
build_dir = www_path.joinpath("build")
|
||||||
|
src_dir = www_path.joinpath("src")
|
||||||
|
build_webpages(src_dir, build_dir, config)
|
||||||
|
files.rsync(f"{build_dir}/", "/var/www/html", flags=["-avz"])
|
||||||
|
|
||||||
|
_install_remote_venv_with_chatmaild(config)
|
||||||
|
debug = False
|
||||||
|
dovecot_need_restart = _configure_dovecot(config, debug=debug)
|
||||||
|
postfix_need_restart = _configure_postfix(config, debug=debug)
|
||||||
|
nginx_need_restart = _configure_nginx(config)
|
||||||
|
_uninstall_mta_sts_daemon()
|
||||||
|
|
||||||
|
_remove_rspamd()
|
||||||
|
opendkim_need_restart = _configure_opendkim(mail_domain, "opendkim")
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="Start and enable OpenDKIM",
|
||||||
|
service="opendkim.service",
|
||||||
|
running=True,
|
||||||
|
enabled=True,
|
||||||
|
daemon_reload=opendkim_need_restart,
|
||||||
|
restarted=opendkim_need_restart,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Dovecot should be started before Postfix
|
||||||
|
# because it creates authentication socket
|
||||||
|
# required by Postfix.
|
||||||
|
systemd.service(
|
||||||
|
name="disable dovecot for now" if disable_mail else "Start and enable Dovecot",
|
||||||
|
service="dovecot.service",
|
||||||
|
running=False if disable_mail else True,
|
||||||
|
enabled=False if disable_mail else True,
|
||||||
|
restarted=dovecot_need_restart if not disable_mail else False,
|
||||||
|
)
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="disable postfix for now" if disable_mail else "Start and enable Postfix",
|
||||||
|
service="postfix.service",
|
||||||
|
running=False if disable_mail else True,
|
||||||
|
enabled=False if disable_mail else True,
|
||||||
|
restarted=postfix_need_restart if not disable_mail else False,
|
||||||
|
)
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="Start and enable nginx",
|
||||||
|
service="nginx.service",
|
||||||
|
running=True,
|
||||||
|
enabled=True,
|
||||||
|
restarted=nginx_need_restart,
|
||||||
|
)
|
||||||
|
|
||||||
|
# This file is used by auth proxy.
|
||||||
|
# https://wiki.debian.org/EtcMailName
|
||||||
|
server.shell(
|
||||||
|
name="Setup /etc/mailname",
|
||||||
|
commands=[f"echo {mail_domain} >/etc/mailname; chmod 644 /etc/mailname"],
|
||||||
|
)
|
||||||
|
|
||||||
|
journald_conf = files.put(
|
||||||
|
name="Configure journald",
|
||||||
|
src=importlib.resources.files(__package__).joinpath("journald.conf"),
|
||||||
|
dest="/etc/systemd/journald.conf",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
systemd.service(
|
||||||
|
name="Start and enable journald",
|
||||||
|
service="systemd-journald.service",
|
||||||
|
running=True,
|
||||||
|
enabled=True,
|
||||||
|
restarted=journald_conf.changed,
|
||||||
|
)
|
||||||
|
files.directory(
|
||||||
|
name="Ensure old logs on disk are deleted",
|
||||||
|
path="/var/log/journal/",
|
||||||
|
present=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
apt.packages(
|
||||||
|
name="Ensure cron is installed",
|
||||||
|
packages=["cron"],
|
||||||
|
)
|
||||||
|
|
||||||
|
deploy_mtail(config)
|
||||||
@@ -1,56 +1,75 @@
|
|||||||
from pyinfra.operations import apt, server
|
import importlib.resources
|
||||||
|
|
||||||
from ..basedeploy import Deployer
|
from pyinfra import host
|
||||||
|
from pyinfra.facts.systemd import SystemdStatus
|
||||||
|
from pyinfra.operations import apt, files, server, systemd
|
||||||
|
|
||||||
|
|
||||||
class AcmetoolDeployer(Deployer):
|
def deploy_acmetool(email="", domains=[]):
|
||||||
def __init__(self, email, domains):
|
"""Deploy acmetool."""
|
||||||
self.domains = domains
|
|
||||||
self.email = email
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
apt.packages(
|
apt.packages(
|
||||||
name="Install acmetool",
|
name="Install acmetool",
|
||||||
packages=["acmetool"],
|
packages=["acmetool"],
|
||||||
)
|
)
|
||||||
|
|
||||||
self.remove_file("/etc/cron.d/acmetool")
|
files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("acmetool.cron").open("rb"),
|
||||||
self.put_executable("acmetool/acmetool.hook", "/etc/acme/hooks/nginx")
|
dest="/etc/cron.d/acmetool",
|
||||||
self.remove_file("/usr/lib/acme/hooks/nginx")
|
user="root",
|
||||||
|
group="root",
|
||||||
def configure(self):
|
mode="644",
|
||||||
self.put_template(
|
|
||||||
"acmetool/response-file.yaml.j2",
|
|
||||||
"/var/lib/acme/conf/responses",
|
|
||||||
email=self.email,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
self.put_template(
|
files.put(
|
||||||
"acmetool/target.yaml.j2",
|
src=importlib.resources.files(__package__).joinpath("acmetool.hook").open("rb"),
|
||||||
"/var/lib/acme/conf/target",
|
dest="/usr/lib/acme/hooks/nginx",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="744",
|
||||||
|
)
|
||||||
|
|
||||||
|
files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("response-file.yaml.j2"),
|
||||||
|
dest="/var/lib/acme/conf/responses",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
email=email,
|
||||||
|
)
|
||||||
|
|
||||||
|
files.template(
|
||||||
|
src=importlib.resources.files(__package__).joinpath("target.yaml.j2"),
|
||||||
|
dest="/var/lib/acme/conf/target",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
|
||||||
|
service_file = files.put(
|
||||||
|
src=importlib.resources.files(__package__).joinpath(
|
||||||
|
"acmetool-redirector.service"
|
||||||
|
),
|
||||||
|
dest="/etc/systemd/system/acmetool-redirector.service",
|
||||||
|
user="root",
|
||||||
|
group="root",
|
||||||
|
mode="644",
|
||||||
|
)
|
||||||
|
if host.get_fact(SystemdStatus).get("nginx.service"):
|
||||||
|
systemd.service(
|
||||||
|
name="Stop nginx service to free port 80",
|
||||||
|
service="nginx",
|
||||||
|
running=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
systemd.service(
|
||||||
|
name="Setup acmetool-redirector service",
|
||||||
|
service="acmetool-redirector.service",
|
||||||
|
running=True,
|
||||||
|
enabled=True,
|
||||||
|
restarted=service_file.changed,
|
||||||
)
|
)
|
||||||
|
|
||||||
server.shell(
|
server.shell(
|
||||||
name=f"Remove old acmetool desired files for {self.domains[0]}",
|
name=f"Request certificate for: {', '.join(domains)}",
|
||||||
commands=[f"rm -f /var/lib/acme/desired/{self.domains[0]}-*"],
|
commands=[f"acmetool want --xlog.severity=debug {' '.join(domains)}"],
|
||||||
)
|
|
||||||
self.put_template(
|
|
||||||
"acmetool/desired.yaml.j2",
|
|
||||||
f"/var/lib/acme/desired/{self.domains[0]}",
|
|
||||||
domains=self.domains,
|
|
||||||
)
|
|
||||||
|
|
||||||
self.ensure_systemd_unit("acmetool/acmetool-redirector.service")
|
|
||||||
self.ensure_systemd_unit("acmetool/acmetool-reconcile.service")
|
|
||||||
self.ensure_systemd_unit("acmetool/acmetool-reconcile.timer")
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service("acmetool-redirector.service")
|
|
||||||
self.ensure_service("acmetool-reconcile.service", running=False, enabled=False)
|
|
||||||
self.ensure_service("acmetool-reconcile.timer")
|
|
||||||
|
|
||||||
server.shell(
|
|
||||||
name=f"Reconcile certificates for: {', '.join(self.domains)}",
|
|
||||||
commands=["acmetool --batch --xlog.severity=debug reconcile"],
|
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -1,8 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=Renew TLS certificates with acmetool
|
|
||||||
After=network.target
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=oneshot
|
|
||||||
ExecStart=/usr/bin/acmetool --batch reconcile
|
|
||||||
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=Renew TLS certificates with acmetool
|
|
||||||
|
|
||||||
[Timer]
|
|
||||||
OnCalendar=*-*-* 16:20:00
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=timers.target
|
|
||||||
@@ -3,7 +3,7 @@ Description=acmetool HTTP redirector
|
|||||||
|
|
||||||
[Service]
|
[Service]
|
||||||
Type=notify
|
Type=notify
|
||||||
ExecStart=/usr/bin/acmetool redirector --service.uid=daemon --bind=127.0.0.1:402
|
ExecStart=/usr/bin/acmetool redirector --service.uid=daemon
|
||||||
Restart=always
|
Restart=always
|
||||||
RestartSec=30
|
RestartSec=30
|
||||||
|
|
||||||
|
|||||||
4
cmdeploy/src/cmdeploy/acmetool/acmetool.cron
Normal file
4
cmdeploy/src/cmdeploy/acmetool/acmetool.cron
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
SHELL=/bin/sh
|
||||||
|
PATH=/bin:/sbin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/local/sbin
|
||||||
|
MAILTO=root
|
||||||
|
20 16 * * * root /usr/bin/acmetool --batch reconcile && systemctl reload dovecot && systemctl reload postfix && systemctl reload nginx
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
satisfy:
|
|
||||||
names:
|
|
||||||
{%- for domain in domains %}
|
|
||||||
- {{ domain }}
|
|
||||||
{%- endfor %}
|
|
||||||
|
|
||||||
@@ -1,2 +1,2 @@
|
|||||||
"acme-enter-email": "{{ email }}"
|
"acme-enter-email": "{{ email }}"
|
||||||
"acme-agreement:https://letsencrypt.org/documents/LE-SA-v1.6-August-18-2025.pdf": true
|
"acme-agreement:https://letsencrypt.org/documents/LE-SA-v1.5-February-24-2025.pdf": true
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
request:
|
request:
|
||||||
provider: https://acme-v02.api.letsencrypt.org/directory
|
provider: https://acme-v02.api.letsencrypt.org/directory
|
||||||
key:
|
key:
|
||||||
type: ecdsa
|
type: rsa
|
||||||
ecdsa-curve: nistp256
|
|
||||||
challenge:
|
challenge:
|
||||||
webroot-paths:
|
webroot-paths:
|
||||||
- /var/www/html/.well-known/acme-challenge
|
- /var/www/html/.well-known/acme-challenge
|
||||||
|
|||||||
@@ -1,255 +0,0 @@
|
|||||||
import importlib.resources
|
|
||||||
import io
|
|
||||||
import os
|
|
||||||
from contextlib import contextmanager
|
|
||||||
|
|
||||||
from pyinfra import host
|
|
||||||
from pyinfra.facts.files import Sha256File
|
|
||||||
from pyinfra.facts.server import Command
|
|
||||||
from pyinfra.operations import files, server, systemd
|
|
||||||
|
|
||||||
|
|
||||||
def has_systemd():
|
|
||||||
"""Returns False during Docker image builds or any other non-systemd environment."""
|
|
||||||
return os.path.isdir("/run/systemd/system")
|
|
||||||
|
|
||||||
|
|
||||||
def is_in_container() -> bool:
|
|
||||||
"""Return True if running inside a container (Docker, LXC, etc.)."""
|
|
||||||
return (
|
|
||||||
host.get_fact(
|
|
||||||
Command,
|
|
||||||
"systemd-detect-virt --container --quiet 2>/dev/null && echo yes || true",
|
|
||||||
)
|
|
||||||
== "yes"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@contextmanager
|
|
||||||
def blocked_service_startup():
|
|
||||||
"""Prevent services from auto-starting during package installation.
|
|
||||||
|
|
||||||
Installs a ``/usr/sbin/policy-rc.d`` that exits 101, blocking any
|
|
||||||
service from being started by the package manager. This avoids bind
|
|
||||||
conflicts and CPU/RAM spikes during initial setup. The file is removed
|
|
||||||
when the context exits.
|
|
||||||
"""
|
|
||||||
# For documentation about policy-rc.d, see:
|
|
||||||
# https://people.debian.org/~hmh/invokerc.d-policyrc.d-specification.txt
|
|
||||||
files.put(
|
|
||||||
src=get_resource("policy-rc.d"),
|
|
||||||
dest="/usr/sbin/policy-rc.d",
|
|
||||||
user="root",
|
|
||||||
group="root",
|
|
||||||
mode="755",
|
|
||||||
)
|
|
||||||
yield
|
|
||||||
files.file("/usr/sbin/policy-rc.d", present=False)
|
|
||||||
|
|
||||||
|
|
||||||
def get_resource(arg, pkg=__package__):
|
|
||||||
return importlib.resources.files(pkg).joinpath(arg)
|
|
||||||
|
|
||||||
|
|
||||||
def configure_remote_units(deployer, mail_domain, units) -> None:
|
|
||||||
remote_base_dir = "/usr/local/lib/chatmaild"
|
|
||||||
remote_venv_dir = f"{remote_base_dir}/venv"
|
|
||||||
remote_chatmail_inipath = f"{remote_base_dir}/chatmail.ini"
|
|
||||||
|
|
||||||
# install systemd units
|
|
||||||
for fn in units:
|
|
||||||
params = dict(
|
|
||||||
execpath=f"{remote_venv_dir}/bin/{fn}",
|
|
||||||
config_path=remote_chatmail_inipath,
|
|
||||||
remote_venv_dir=remote_venv_dir,
|
|
||||||
mail_domain=mail_domain,
|
|
||||||
)
|
|
||||||
|
|
||||||
basename = fn if "." in fn else f"{fn}.service"
|
|
||||||
|
|
||||||
source_path = get_resource(f"service/{basename}.f")
|
|
||||||
content = source_path.read_text().format(**params).encode()
|
|
||||||
|
|
||||||
deployer.put_file(
|
|
||||||
src=io.BytesIO(content),
|
|
||||||
dest=f"/etc/systemd/system/{basename}",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def activate_remote_units(deployer, units) -> None:
|
|
||||||
# activate systemd units
|
|
||||||
for fn in units:
|
|
||||||
basename = fn if "." in fn else f"{fn}.service"
|
|
||||||
|
|
||||||
if fn == "chatmail-expire" or fn == "chatmail-fsreport":
|
|
||||||
# don't auto-start but let the corresponding timer trigger execution
|
|
||||||
enabled = False
|
|
||||||
else:
|
|
||||||
enabled = True
|
|
||||||
|
|
||||||
deployer.ensure_service(basename, running=enabled, enabled=enabled)
|
|
||||||
|
|
||||||
|
|
||||||
class Deployment:
|
|
||||||
def install(self, deployer):
|
|
||||||
# optional 'required_users' contains a list of (user, group, secondary-group-list) tuples.
|
|
||||||
# If the group is None, no group is created corresponding to that user.
|
|
||||||
# If the secondary group list is not None, all listed groups are created as well.
|
|
||||||
required_users = getattr(deployer, "required_users", [])
|
|
||||||
for user, group, groups in required_users:
|
|
||||||
if group is not None:
|
|
||||||
server.group(
|
|
||||||
name="Create {} group".format(group), group=group, system=True
|
|
||||||
)
|
|
||||||
if groups is not None:
|
|
||||||
for group2 in groups:
|
|
||||||
server.group(
|
|
||||||
name="Create {} group".format(group2), group=group2, system=True
|
|
||||||
)
|
|
||||||
server.user(
|
|
||||||
name="Create {} user".format(user),
|
|
||||||
user=user,
|
|
||||||
group=group,
|
|
||||||
groups=groups,
|
|
||||||
system=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.install()
|
|
||||||
|
|
||||||
def configure(self, deployer):
|
|
||||||
deployer.configure()
|
|
||||||
|
|
||||||
def activate(self, deployer):
|
|
||||||
deployer.activate()
|
|
||||||
|
|
||||||
def perform_stages(self, deployers):
|
|
||||||
default_stages = "install,configure,activate"
|
|
||||||
stages = os.getenv("CMDEPLOY_STAGES", default_stages).split(",")
|
|
||||||
|
|
||||||
for stage in stages:
|
|
||||||
for deployer in deployers:
|
|
||||||
getattr(self, stage)(deployer)
|
|
||||||
|
|
||||||
|
|
||||||
class Deployer:
|
|
||||||
need_restart = False
|
|
||||||
daemon_reload = False
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
pass
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
pass
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
pass
|
|
||||||
|
|
||||||
def ensure_service(self, service, running=True, enabled=True):
|
|
||||||
if running:
|
|
||||||
verb = "Start and enable"
|
|
||||||
else:
|
|
||||||
verb = "Stop"
|
|
||||||
systemd.service(
|
|
||||||
name=f"{verb} {service}",
|
|
||||||
service=service,
|
|
||||||
running=running,
|
|
||||||
enabled=enabled,
|
|
||||||
restarted=self.need_restart if running else False,
|
|
||||||
daemon_reload=self.daemon_reload,
|
|
||||||
)
|
|
||||||
self.daemon_reload = False
|
|
||||||
|
|
||||||
def ensure_systemd_unit(self, src, **kwargs):
|
|
||||||
dest_name = src.split("/")[-1].replace(".j2", "")
|
|
||||||
dest = f"/etc/systemd/system/{dest_name}"
|
|
||||||
if src.endswith(".j2"):
|
|
||||||
return self.put_template(src, dest, **kwargs)
|
|
||||||
return self.put_file(src, dest)
|
|
||||||
|
|
||||||
def put_file(self, src, dest, mode="644"):
|
|
||||||
if isinstance(src, str):
|
|
||||||
src = get_resource(src)
|
|
||||||
res = files.put(
|
|
||||||
name=f"Upload {dest}",
|
|
||||||
src=src,
|
|
||||||
dest=dest,
|
|
||||||
user="root",
|
|
||||||
group="root",
|
|
||||||
mode=mode,
|
|
||||||
)
|
|
||||||
|
|
||||||
return self._update_restart_signals(dest, res)
|
|
||||||
|
|
||||||
def put_executable(self, src, dest):
|
|
||||||
return self.put_file(src, dest, mode="755")
|
|
||||||
|
|
||||||
def put_template(self, src, dest, owner="root", **kwargs):
|
|
||||||
if isinstance(src, str):
|
|
||||||
src = get_resource(src)
|
|
||||||
res = files.template(
|
|
||||||
name=f"Upload {dest}",
|
|
||||||
src=src,
|
|
||||||
dest=dest,
|
|
||||||
user=owner,
|
|
||||||
group=owner,
|
|
||||||
mode="644",
|
|
||||||
**kwargs,
|
|
||||||
)
|
|
||||||
|
|
||||||
return self._update_restart_signals(dest, res)
|
|
||||||
|
|
||||||
def remove_file(self, dest):
|
|
||||||
res = files.file(name=f"Remove {dest}", path=dest, present=False)
|
|
||||||
return self._update_restart_signals(dest, res)
|
|
||||||
|
|
||||||
def ensure_line(self, path, line, **kwargs):
|
|
||||||
name = kwargs.pop("name", f"Ensure line in {path}")
|
|
||||||
res = files.line(name=name, path=path, line=line, **kwargs)
|
|
||||||
return self._update_restart_signals(path, res)
|
|
||||||
|
|
||||||
def ensure_directory(self, path, owner="root", mode="755", **kwargs):
|
|
||||||
name = kwargs.pop("name", f"Ensure directory {path}")
|
|
||||||
res = files.directory(
|
|
||||||
name=name,
|
|
||||||
path=path,
|
|
||||||
user=owner,
|
|
||||||
group=owner,
|
|
||||||
mode=mode,
|
|
||||||
present=True,
|
|
||||||
**kwargs,
|
|
||||||
)
|
|
||||||
return self._update_restart_signals(path, res)
|
|
||||||
|
|
||||||
def remove_directory(self, path, **kwargs):
|
|
||||||
name = kwargs.pop("name", f"Remove directory {path}")
|
|
||||||
res = files.directory(name=name, path=path, present=False, **kwargs)
|
|
||||||
return self._update_restart_signals(path, res)
|
|
||||||
|
|
||||||
def download_executable(self, url, dest, sha256sum, extract=None):
|
|
||||||
existing = host.get_fact(Sha256File, dest)
|
|
||||||
if existing == sha256sum:
|
|
||||||
return
|
|
||||||
|
|
||||||
tmp = f"{dest}.new"
|
|
||||||
if extract:
|
|
||||||
dl_cmd = f"curl -fSL {url} | {extract} >{tmp}"
|
|
||||||
else:
|
|
||||||
dl_cmd = f"curl -fSL {url} -o {tmp}"
|
|
||||||
|
|
||||||
server.shell(
|
|
||||||
name=f"Download {dest}",
|
|
||||||
commands=[
|
|
||||||
f"({dl_cmd}"
|
|
||||||
f" && echo '{sha256sum} {tmp}' | sha256sum -c"
|
|
||||||
f" && mv {tmp} {dest})",
|
|
||||||
f"chmod 755 {dest}",
|
|
||||||
],
|
|
||||||
)
|
|
||||||
self.need_restart = True
|
|
||||||
|
|
||||||
def _update_restart_signals(self, path, res):
|
|
||||||
if res.changed:
|
|
||||||
self.need_restart = True
|
|
||||||
if str(path).startswith("/etc/systemd/system/"):
|
|
||||||
self.daemon_reload = True
|
|
||||||
return res
|
|
||||||
30
cmdeploy/src/cmdeploy/chatmail.zone.j2
Normal file
30
cmdeploy/src/cmdeploy/chatmail.zone.j2
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
;
|
||||||
|
; Required DNS entries for chatmail servers
|
||||||
|
;
|
||||||
|
{% if A %}
|
||||||
|
{{ mail_domain }}. A {{ A }}
|
||||||
|
{% endif %}
|
||||||
|
{% if AAAA %}
|
||||||
|
{{ mail_domain }}. AAAA {{ AAAA }}
|
||||||
|
{% endif %}
|
||||||
|
{{ mail_domain }}. MX 10 {{ mail_domain }}.
|
||||||
|
_mta-sts.{{ mail_domain }}. TXT "v=STSv1; id={{ sts_id }}"
|
||||||
|
mta-sts.{{ mail_domain }}. CNAME {{ mail_domain }}.
|
||||||
|
www.{{ mail_domain }}. CNAME {{ mail_domain }}.
|
||||||
|
{{ dkim_entry }}
|
||||||
|
|
||||||
|
;
|
||||||
|
; Recommended DNS entries for interoperability and security-hardening
|
||||||
|
;
|
||||||
|
{{ mail_domain }}. TXT "v=spf1 a ~all"
|
||||||
|
_dmarc.{{ mail_domain }}. TXT "v=DMARC1;p=reject;adkim=s;aspf=s"
|
||||||
|
|
||||||
|
{% if acme_account_url %}
|
||||||
|
{{ mail_domain }}. CAA 0 issue "letsencrypt.org;accounturi={{ acme_account_url }}"
|
||||||
|
{% endif %}
|
||||||
|
_adsp._domainkey.{{ mail_domain }}. TXT "dkim=discardable"
|
||||||
|
|
||||||
|
_submission._tcp.{{ mail_domain }}. SRV 0 1 587 {{ mail_domain }}.
|
||||||
|
_submissions._tcp.{{ mail_domain }}. SRV 0 1 465 {{ mail_domain }}.
|
||||||
|
_imap._tcp.{{ mail_domain }}. SRV 0 1 143 {{ mail_domain }}.
|
||||||
|
_imaps._tcp.{{ mail_domain }}. SRV 0 1 993 {{ mail_domain }}.
|
||||||
@@ -5,6 +5,7 @@ along with command line option and subcommand parsing.
|
|||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
import importlib.resources
|
import importlib.resources
|
||||||
|
import importlib.util
|
||||||
import os
|
import os
|
||||||
import pathlib
|
import pathlib
|
||||||
import shutil
|
import shutil
|
||||||
@@ -18,7 +19,7 @@ from packaging import version
|
|||||||
from termcolor import colored
|
from termcolor import colored
|
||||||
|
|
||||||
from . import dns, remote
|
from . import dns, remote
|
||||||
from .sshexec import LocalExec, SSHExec
|
from .sshexec import SSHExec
|
||||||
|
|
||||||
#
|
#
|
||||||
# cmdeploy sub commands and options
|
# cmdeploy sub commands and options
|
||||||
@@ -31,30 +32,17 @@ def init_cmd_options(parser):
|
|||||||
action="store",
|
action="store",
|
||||||
help="fully qualified DNS domain name for your chatmail instance",
|
help="fully qualified DNS domain name for your chatmail instance",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
|
||||||
"--force",
|
|
||||||
dest="recreate_ini",
|
|
||||||
action="store_true",
|
|
||||||
help="force reacreate ini file",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def init_cmd(args, out):
|
def init_cmd(args, out):
|
||||||
"""Initialize chatmail config file."""
|
"""Initialize chatmail config file."""
|
||||||
mail_domain = args.chatmail_domain
|
mail_domain = args.chatmail_domain
|
||||||
inipath = args.inipath
|
|
||||||
if args.inipath.exists():
|
if args.inipath.exists():
|
||||||
if not args.recreate_ini:
|
print(f"Path exists, not modifying: {args.inipath}")
|
||||||
print(f"[WARNING] Path exists, not modifying: {inipath}")
|
|
||||||
return 1
|
return 1
|
||||||
else:
|
else:
|
||||||
print(
|
write_initial_config(args.inipath, mail_domain, overrides={})
|
||||||
f"[WARNING] Force argument was provided, deleting config file: {inipath}"
|
out.green(f"created config file for {mail_domain} in {args.inipath}")
|
||||||
)
|
|
||||||
inipath.unlink()
|
|
||||||
|
|
||||||
write_initial_config(inipath, mail_domain, overrides={})
|
|
||||||
out.green(f"created config file for {mail_domain} in {inipath}")
|
|
||||||
|
|
||||||
|
|
||||||
def run_cmd_options(parser):
|
def run_cmd_options(parser):
|
||||||
@@ -71,60 +59,43 @@ def run_cmd_options(parser):
|
|||||||
help="install/upgrade the server, but disable postfix & dovecot for now",
|
help="install/upgrade the server, but disable postfix & dovecot for now",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--website-only",
|
"--ssh-host",
|
||||||
action="store_true",
|
dest="ssh_host",
|
||||||
help="only update/deploy the website, skipping full server upgrade/deployment, useful when you only changed/updated the web pages and don't need to re-run a full server upgrade",
|
help="specify an SSH host to deploy to; uses mail_domain from chatmail.ini by default",
|
||||||
)
|
)
|
||||||
parser.add_argument(
|
|
||||||
"--skip-dns-check",
|
|
||||||
dest="dns_check_disabled",
|
|
||||||
action="store_true",
|
|
||||||
help="disable checks nslookup for dns",
|
|
||||||
)
|
|
||||||
add_ssh_host_option(parser)
|
|
||||||
|
|
||||||
|
|
||||||
def run_cmd(args, out):
|
def run_cmd(args, out):
|
||||||
"""Deploy chatmail services on the remote server."""
|
"""Deploy chatmail services on the remote server."""
|
||||||
|
|
||||||
ssh_host = args.ssh_host if args.ssh_host else args.config.mail_domain
|
sshexec = args.get_sshexec()
|
||||||
sshexec = get_sshexec(ssh_host)
|
|
||||||
require_iroh = args.config.enable_iroh_relay
|
require_iroh = args.config.enable_iroh_relay
|
||||||
strict_tls = args.config.tls_cert_mode == "acme"
|
|
||||||
if not args.dns_check_disabled:
|
|
||||||
remote_data = dns.get_initial_remote_data(sshexec, args.config.mail_domain)
|
remote_data = dns.get_initial_remote_data(sshexec, args.config.mail_domain)
|
||||||
if not dns.check_initial_remote_data(remote_data, strict_tls=strict_tls, print=out.red):
|
if not dns.check_initial_remote_data(remote_data, print=out.red):
|
||||||
return 1
|
return 1
|
||||||
|
|
||||||
env = os.environ.copy()
|
env = os.environ.copy()
|
||||||
env["CHATMAIL_INI"] = args.inipath
|
env["CHATMAIL_INI"] = args.inipath
|
||||||
env["CHATMAIL_WEBSITE_ONLY"] = "True" if args.website_only else ""
|
|
||||||
env["CHATMAIL_DISABLE_MAIL"] = "True" if args.disable_mail else ""
|
env["CHATMAIL_DISABLE_MAIL"] = "True" if args.disable_mail else ""
|
||||||
env["CHATMAIL_REQUIRE_IROH"] = "True" if require_iroh else ""
|
env["CHATMAIL_REQUIRE_IROH"] = "True" if require_iroh else ""
|
||||||
deploy_path = importlib.resources.files(__package__).joinpath("run.py").resolve()
|
deploy_path = importlib.resources.files(__package__).joinpath("deploy.py").resolve()
|
||||||
pyinf = "pyinfra --dry" if args.dry_run else "pyinfra"
|
pyinf = "pyinfra --dry" if args.dry_run else "pyinfra"
|
||||||
|
ssh_host = args.config.mail_domain if not args.ssh_host else args.ssh_host
|
||||||
cmd = f"{pyinf} --ssh-user root {ssh_host} {deploy_path} -y"
|
cmd = f"{pyinf} --ssh-user root {ssh_host} {deploy_path} -y"
|
||||||
if ssh_host == "localhost":
|
|
||||||
cmd = f"{pyinf} @local {deploy_path} -y"
|
|
||||||
|
|
||||||
if version.parse(pyinfra.__version__) < version.parse("3"):
|
if version.parse(pyinfra.__version__) < version.parse("3"):
|
||||||
out.red("Please re-run scripts/initenv.sh to update pyinfra to version 3.")
|
out.red("Please re-run scripts/initenv.sh to update pyinfra to version 3.")
|
||||||
return 1
|
return 1
|
||||||
|
|
||||||
try:
|
retcode = out.check_call(cmd, env=env)
|
||||||
out.check_call(cmd, env=env)
|
if retcode == 0:
|
||||||
if args.website_only:
|
out.green("Deploy completed, call `cmdeploy dns` next.")
|
||||||
out.green("Website deployment completed.")
|
elif not remote_data["acme_account_url"]:
|
||||||
elif not args.dns_check_disabled and strict_tls and not remote_data["acme_account_url"]:
|
|
||||||
out.red("Deploy completed but letsencrypt not configured")
|
out.red("Deploy completed but letsencrypt not configured")
|
||||||
out.red("Run 'cmdeploy run' again")
|
out.red("Run 'cmdeploy run' again")
|
||||||
|
retcode = 0
|
||||||
else:
|
else:
|
||||||
out.green("Deploy completed, call `cmdeploy dns` next.")
|
|
||||||
return 0
|
|
||||||
except subprocess.CalledProcessError:
|
|
||||||
out.red("Deploy failed")
|
out.red("Deploy failed")
|
||||||
return 1
|
return retcode
|
||||||
|
|
||||||
|
|
||||||
def dns_cmd_options(parser):
|
def dns_cmd_options(parser):
|
||||||
@@ -135,20 +106,16 @@ def dns_cmd_options(parser):
|
|||||||
default=None,
|
default=None,
|
||||||
help="write out a zonefile",
|
help="write out a zonefile",
|
||||||
)
|
)
|
||||||
add_ssh_host_option(parser)
|
|
||||||
|
|
||||||
|
|
||||||
def dns_cmd(args, out):
|
def dns_cmd(args, out):
|
||||||
"""Check DNS entries and optionally generate dns zone file."""
|
"""Check DNS entries and optionally generate dns zone file."""
|
||||||
ssh_host = args.ssh_host if args.ssh_host else args.config.mail_domain
|
sshexec = args.get_sshexec()
|
||||||
sshexec = get_sshexec(ssh_host, verbose=args.verbose)
|
|
||||||
tls_cert_mode = args.config.tls_cert_mode
|
|
||||||
strict_tls = tls_cert_mode == "acme"
|
|
||||||
remote_data = dns.get_initial_remote_data(sshexec, args.config.mail_domain)
|
remote_data = dns.get_initial_remote_data(sshexec, args.config.mail_domain)
|
||||||
if not dns.check_initial_remote_data(remote_data, strict_tls=strict_tls):
|
if not remote_data:
|
||||||
return 1
|
return 1
|
||||||
|
|
||||||
if strict_tls and not remote_data["acme_account_url"]:
|
if not remote_data["acme_account_url"]:
|
||||||
out.red("could not get letsencrypt account url, please run 'cmdeploy run'")
|
out.red("could not get letsencrypt account url, please run 'cmdeploy run'")
|
||||||
return 1
|
return 1
|
||||||
|
|
||||||
@@ -156,7 +123,6 @@ def dns_cmd(args, out):
|
|||||||
out.red("could not determine dkim_entry, please run 'cmdeploy run'")
|
out.red("could not determine dkim_entry, please run 'cmdeploy run'")
|
||||||
return 1
|
return 1
|
||||||
|
|
||||||
remote_data["strict_tls"] = strict_tls
|
|
||||||
zonefile = dns.get_filled_zone_file(remote_data)
|
zonefile = dns.get_filled_zone_file(remote_data)
|
||||||
|
|
||||||
if args.zonefile:
|
if args.zonefile:
|
||||||
@@ -170,15 +136,10 @@ def dns_cmd(args, out):
|
|||||||
return retcode
|
return retcode
|
||||||
|
|
||||||
|
|
||||||
def status_cmd_options(parser):
|
|
||||||
add_ssh_host_option(parser)
|
|
||||||
|
|
||||||
|
|
||||||
def status_cmd(args, out):
|
def status_cmd(args, out):
|
||||||
"""Display status for online chatmail instance."""
|
"""Display status for online chatmail instance."""
|
||||||
|
|
||||||
ssh_host = args.ssh_host if args.ssh_host else args.config.mail_domain
|
sshexec = args.get_sshexec()
|
||||||
sshexec = get_sshexec(ssh_host, verbose=args.verbose)
|
|
||||||
|
|
||||||
out.green(f"chatmail domain: {args.config.mail_domain}")
|
out.green(f"chatmail domain: {args.config.mail_domain}")
|
||||||
if args.config.privacy_mail:
|
if args.config.privacy_mail:
|
||||||
@@ -191,16 +152,23 @@ def status_cmd(args, out):
|
|||||||
|
|
||||||
|
|
||||||
def test_cmd_options(parser):
|
def test_cmd_options(parser):
|
||||||
add_ssh_host_option(parser)
|
parser.add_argument(
|
||||||
|
"--slow",
|
||||||
|
dest="slow",
|
||||||
|
action="store_true",
|
||||||
|
help="also run slow tests",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def test_cmd(args, out):
|
def test_cmd(args, out):
|
||||||
"""Run local and online tests for chatmail deployment."""
|
"""Run local and online tests for chatmail deployment.
|
||||||
|
|
||||||
env = os.environ.copy()
|
This will automatically pip-install 'deltachat' if it's not available.
|
||||||
env["CHATMAIL_INI"] = str(args.inipath.absolute())
|
"""
|
||||||
if args.ssh_host:
|
|
||||||
env["CHATMAIL_SSH"] = args.ssh_host
|
x = importlib.util.find_spec("deltachat")
|
||||||
|
if x is None:
|
||||||
|
out.check_call(f"{sys.executable} -m pip install deltachat")
|
||||||
|
|
||||||
pytest_path = shutil.which("pytest")
|
pytest_path = shutil.which("pytest")
|
||||||
pytest_args = [
|
pytest_args = [
|
||||||
@@ -212,7 +180,9 @@ def test_cmd(args, out):
|
|||||||
"-v",
|
"-v",
|
||||||
"--durations=5",
|
"--durations=5",
|
||||||
]
|
]
|
||||||
ret = out.run_ret(pytest_args, env=env)
|
if args.slow:
|
||||||
|
pytest_args.append("--slow")
|
||||||
|
ret = out.run_ret(pytest_args)
|
||||||
return ret
|
return ret
|
||||||
|
|
||||||
|
|
||||||
@@ -228,12 +198,7 @@ def fmt_cmd_options(parser):
|
|||||||
def fmt_cmd(args, out):
|
def fmt_cmd(args, out):
|
||||||
"""Run formattting fixes on all chatmail source code."""
|
"""Run formattting fixes on all chatmail source code."""
|
||||||
|
|
||||||
chatmaild_dir = importlib.resources.files("chatmaild").resolve()
|
sources = [str(importlib.resources.files(x)) for x in ("chatmaild", "cmdeploy")]
|
||||||
cmdeploy_dir = chatmaild_dir.joinpath(
|
|
||||||
"..", "..", "..", "cmdeploy", "src", "cmdeploy"
|
|
||||||
).resolve()
|
|
||||||
sources = [str(chatmaild_dir), str(cmdeploy_dir)]
|
|
||||||
|
|
||||||
format_args = [shutil.which("ruff"), "format"]
|
format_args = [shutil.which("ruff"), "format"]
|
||||||
check_args = [shutil.which("ruff"), "check"]
|
check_args = [shutil.which("ruff"), "check"]
|
||||||
|
|
||||||
@@ -299,21 +264,12 @@ class Out:
|
|||||||
return proc.returncode
|
return proc.returncode
|
||||||
|
|
||||||
|
|
||||||
def add_ssh_host_option(parser):
|
|
||||||
parser.add_argument(
|
|
||||||
"--ssh-host",
|
|
||||||
dest="ssh_host",
|
|
||||||
help="Run commands on 'localhost' or on a specific SSH host "
|
|
||||||
"instead of chatmail.ini's mail_domain.",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def add_config_option(parser):
|
def add_config_option(parser):
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"--config",
|
"--config",
|
||||||
dest="inipath",
|
dest="inipath",
|
||||||
action="store",
|
action="store",
|
||||||
default=Path(os.environ.get("CHATMAIL_INI", "chatmail.ini")),
|
default=Path("chatmail.ini"),
|
||||||
type=Path,
|
type=Path,
|
||||||
help="path to the chatmail.ini file",
|
help="path to the chatmail.ini file",
|
||||||
)
|
)
|
||||||
@@ -363,14 +319,6 @@ def get_parser():
|
|||||||
return parser
|
return parser
|
||||||
|
|
||||||
|
|
||||||
def get_sshexec(ssh_host: str, verbose=True):
|
|
||||||
if ssh_host in ["localhost", "@local"]:
|
|
||||||
return LocalExec(verbose)
|
|
||||||
if verbose:
|
|
||||||
print(f"[ssh] login to {ssh_host}")
|
|
||||||
return SSHExec(ssh_host, verbose=verbose)
|
|
||||||
|
|
||||||
|
|
||||||
def main(args=None):
|
def main(args=None):
|
||||||
"""Provide main entry point for 'cmdeploy' CLI invocation."""
|
"""Provide main entry point for 'cmdeploy' CLI invocation."""
|
||||||
parser = get_parser()
|
parser = get_parser()
|
||||||
@@ -378,6 +326,12 @@ def main(args=None):
|
|||||||
if not hasattr(args, "func"):
|
if not hasattr(args, "func"):
|
||||||
return parser.parse_args(["-h"])
|
return parser.parse_args(["-h"])
|
||||||
|
|
||||||
|
def get_sshexec():
|
||||||
|
print(f"[ssh] login to {args.config.mail_domain}")
|
||||||
|
return SSHExec(args.config.mail_domain, verbose=args.verbose)
|
||||||
|
|
||||||
|
args.get_sshexec = get_sshexec
|
||||||
|
|
||||||
out = Out()
|
out = Out()
|
||||||
kwargs = {}
|
kwargs = {}
|
||||||
if args.func.__name__ not in ("init_cmd", "fmt_cmd"):
|
if args.func.__name__ not in ("init_cmd", "fmt_cmd"):
|
||||||
|
|||||||
@@ -3,9 +3,7 @@ import os
|
|||||||
|
|
||||||
import pyinfra
|
import pyinfra
|
||||||
|
|
||||||
# pyinfra runs this module as a python file and not as a module so
|
from cmdeploy import deploy_chatmail
|
||||||
# import paths must be absolute
|
|
||||||
from cmdeploy.deployers import deploy_chatmail
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
@@ -14,9 +12,8 @@ def main():
|
|||||||
importlib.resources.files("cmdeploy").joinpath("../../../chatmail.ini"),
|
importlib.resources.files("cmdeploy").joinpath("../../../chatmail.ini"),
|
||||||
)
|
)
|
||||||
disable_mail = bool(os.environ.get("CHATMAIL_DISABLE_MAIL"))
|
disable_mail = bool(os.environ.get("CHATMAIL_DISABLE_MAIL"))
|
||||||
website_only = bool(os.environ.get("CHATMAIL_WEBSITE_ONLY"))
|
|
||||||
|
|
||||||
deploy_chatmail(config_path, disable_mail, website_only)
|
deploy_chatmail(config_path, disable_mail)
|
||||||
|
|
||||||
|
|
||||||
if pyinfra.is_cli:
|
if pyinfra.is_cli:
|
||||||
@@ -1,555 +0,0 @@
|
|||||||
"""
|
|
||||||
Chat Mail pyinfra deploy.
|
|
||||||
"""
|
|
||||||
|
|
||||||
import shutil
|
|
||||||
import subprocess
|
|
||||||
import sys
|
|
||||||
from io import BytesIO, StringIO
|
|
||||||
from pathlib import Path
|
|
||||||
|
|
||||||
from chatmaild.config import read_config
|
|
||||||
from pyinfra import facts, host, logger
|
|
||||||
from pyinfra.api import FactBase
|
|
||||||
from pyinfra.facts import hardware
|
|
||||||
from pyinfra.facts.systemd import SystemdEnabled
|
|
||||||
from pyinfra.operations import apt, files, pip, server, systemd
|
|
||||||
|
|
||||||
from cmdeploy.cmdeploy import Out
|
|
||||||
|
|
||||||
from .acmetool import AcmetoolDeployer
|
|
||||||
from .basedeploy import (
|
|
||||||
Deployer,
|
|
||||||
Deployment,
|
|
||||||
activate_remote_units,
|
|
||||||
blocked_service_startup,
|
|
||||||
configure_remote_units,
|
|
||||||
has_systemd,
|
|
||||||
is_in_container,
|
|
||||||
)
|
|
||||||
from .dovecot.deployer import DovecotDeployer
|
|
||||||
from .external.deployer import ExternalTlsDeployer
|
|
||||||
from .filtermail.deployer import FiltermailDeployer
|
|
||||||
from .mtail.deployer import MtailDeployer
|
|
||||||
from .nginx.deployer import NginxDeployer
|
|
||||||
from .opendkim.deployer import OpendkimDeployer
|
|
||||||
from .postfix.deployer import PostfixDeployer
|
|
||||||
from .selfsigned.deployer import SelfSignedTlsDeployer
|
|
||||||
from .www import build_webpages, find_merge_conflict, get_paths
|
|
||||||
|
|
||||||
|
|
||||||
class Port(FactBase):
|
|
||||||
"""
|
|
||||||
Returns the process occupying a port.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def command(self, port: int) -> str:
|
|
||||||
return (
|
|
||||||
"ss -lptn 'src :%d' | awk 'NR>1 {print $6,$7}' | sed 's/users:((\"//;s/\".*//'"
|
|
||||||
% (port,)
|
|
||||||
)
|
|
||||||
|
|
||||||
def process(self, output: [str]) -> str:
|
|
||||||
return output[0]
|
|
||||||
|
|
||||||
|
|
||||||
def _build_chatmaild(dist_dir) -> None:
|
|
||||||
dist_dir = Path(dist_dir).resolve()
|
|
||||||
if dist_dir.exists():
|
|
||||||
shutil.rmtree(dist_dir)
|
|
||||||
dist_dir.mkdir()
|
|
||||||
subprocess.check_output(
|
|
||||||
[sys.executable, "-m", "build", "-n"]
|
|
||||||
+ ["--sdist", "chatmaild", "--outdir", str(dist_dir)]
|
|
||||||
)
|
|
||||||
entries = list(dist_dir.iterdir())
|
|
||||||
assert len(entries) == 1
|
|
||||||
return entries[0]
|
|
||||||
|
|
||||||
|
|
||||||
def remove_legacy_artifacts():
|
|
||||||
if not has_systemd():
|
|
||||||
return
|
|
||||||
# disable legacy doveauth-dictproxy.service
|
|
||||||
if host.get_fact(SystemdEnabled).get("doveauth-dictproxy.service"):
|
|
||||||
systemd.service(
|
|
||||||
name="Disable legacy doveauth-dictproxy.service",
|
|
||||||
service="doveauth-dictproxy.service",
|
|
||||||
running=False,
|
|
||||||
enabled=False,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def _install_remote_venv_with_chatmaild(deployer) -> None:
|
|
||||||
remove_legacy_artifacts()
|
|
||||||
dist_file = _build_chatmaild(dist_dir=Path("chatmaild/dist"))
|
|
||||||
remote_base_dir = "/usr/local/lib/chatmaild"
|
|
||||||
remote_dist_file = f"{remote_base_dir}/dist/{dist_file.name}"
|
|
||||||
remote_venv_dir = f"{remote_base_dir}/venv"
|
|
||||||
|
|
||||||
apt.packages(
|
|
||||||
name="apt install python3-virtualenv",
|
|
||||||
packages=["python3-virtualenv"],
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.ensure_directory(f"{remote_base_dir}/dist")
|
|
||||||
deployer.put_file(
|
|
||||||
src=dist_file.open("rb"),
|
|
||||||
dest=remote_dist_file,
|
|
||||||
)
|
|
||||||
|
|
||||||
pip.virtualenv(
|
|
||||||
name=f"chatmaild virtualenv {remote_venv_dir}",
|
|
||||||
path=remote_venv_dir,
|
|
||||||
always_copy=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
apt.packages(
|
|
||||||
name="install gcc and headers to build crypt_r source package",
|
|
||||||
packages=["gcc", "python3-dev"],
|
|
||||||
)
|
|
||||||
|
|
||||||
server.shell(
|
|
||||||
name=f"forced pip-install {dist_file.name}",
|
|
||||||
commands=[
|
|
||||||
f"{remote_venv_dir}/bin/pip install --force-reinstall {remote_dist_file}"
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def _configure_remote_venv_with_chatmaild(deployer, config) -> None:
|
|
||||||
remote_base_dir = "/usr/local/lib/chatmaild"
|
|
||||||
remote_chatmail_inipath = f"{remote_base_dir}/chatmail.ini"
|
|
||||||
|
|
||||||
deployer.put_file(
|
|
||||||
src=config._getbytefile(),
|
|
||||||
dest=remote_chatmail_inipath,
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.remove_file("/etc/cron.d/chatmail-metrics")
|
|
||||||
deployer.remove_file("/var/www/html/metrics")
|
|
||||||
|
|
||||||
|
|
||||||
class UnboundDeployer(Deployer):
|
|
||||||
def __init__(self, config):
|
|
||||||
self.config = config
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
# On an IPv4-only system, if unbound is started but not configured,
|
|
||||||
# it causes subsequent steps to fail to resolve hosts.
|
|
||||||
with blocked_service_startup():
|
|
||||||
apt.packages(
|
|
||||||
name="Install unbound",
|
|
||||||
packages=["unbound", "unbound-anchor", "dnsutils"],
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
# Remove dynamic resolver managers that compete for /etc/resolv.conf.
|
|
||||||
apt.packages(
|
|
||||||
name="Purge resolvconf",
|
|
||||||
packages=["resolvconf"],
|
|
||||||
present=False,
|
|
||||||
extra_uninstall_args="--purge",
|
|
||||||
)
|
|
||||||
# systemd-resolved can't be purged due to dependencies; stop and mask.
|
|
||||||
server.shell(
|
|
||||||
name="Stop and mask systemd-resolved",
|
|
||||||
commands=[
|
|
||||||
"systemctl stop systemd-resolved.service || true",
|
|
||||||
"systemctl mask systemd-resolved.service",
|
|
||||||
],
|
|
||||||
)
|
|
||||||
# Configure unbound resolver with Quad9 fallback and a trailing newline
|
|
||||||
# (SolusVM bug).
|
|
||||||
self.put_file(
|
|
||||||
src=BytesIO(b"nameserver 127.0.0.1\nnameserver 9.9.9.9\n"),
|
|
||||||
dest="/etc/resolv.conf",
|
|
||||||
)
|
|
||||||
server.shell(
|
|
||||||
name="Generate root keys for validating DNSSEC",
|
|
||||||
commands=[
|
|
||||||
"unbound-anchor -a /var/lib/unbound/root.key || true",
|
|
||||||
],
|
|
||||||
)
|
|
||||||
if self.config.disable_ipv6:
|
|
||||||
self.ensure_directory(
|
|
||||||
path="/etc/unbound/unbound.conf.d",
|
|
||||||
)
|
|
||||||
self.put_template(
|
|
||||||
"unbound/unbound.conf.j2",
|
|
||||||
"/etc/unbound/unbound.conf.d/chatmail.conf",
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
self.remove_file("/etc/unbound/unbound.conf.d/chatmail.conf")
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
server.shell(
|
|
||||||
name="Generate root keys for validating DNSSEC",
|
|
||||||
commands=[
|
|
||||||
"systemctl reset-failed unbound.service",
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
self.ensure_service("unbound.service")
|
|
||||||
|
|
||||||
self.ensure_service(
|
|
||||||
"unbound-resolvconf.service",
|
|
||||||
running=False,
|
|
||||||
enabled=False,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class MtastsDeployer(Deployer):
|
|
||||||
def configure(self):
|
|
||||||
# Remove configuration.
|
|
||||||
self.remove_file("/etc/mta-sts-daemon.yml")
|
|
||||||
self.remove_directory("/usr/local/lib/postfix-mta-sts-resolver")
|
|
||||||
self.remove_file("/etc/systemd/system/mta-sts-daemon.service")
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service(
|
|
||||||
"mta-sts-daemon.service",
|
|
||||||
running=False,
|
|
||||||
enabled=False,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class WebsiteDeployer(Deployer):
|
|
||||||
def __init__(self, config):
|
|
||||||
self.config = config
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
self.ensure_directory("/var/www")
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
www_path, src_dir, build_dir = get_paths(self.config)
|
|
||||||
# if www_folder was set to a non-existing folder, skip upload
|
|
||||||
if not www_path.is_dir():
|
|
||||||
logger.warning("Building web pages is disabled in chatmail.ini, skipping")
|
|
||||||
elif (path := find_merge_conflict(src_dir)) is not None:
|
|
||||||
logger.warning(
|
|
||||||
f"Merge conflict found in {path}, skipping website deployment. Fix merge conflict if you want to upload your web page."
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
# if www_folder is a hugo page, build it
|
|
||||||
if build_dir:
|
|
||||||
www_path = build_webpages(src_dir, build_dir, self.config)
|
|
||||||
if www_path is None:
|
|
||||||
logger.warning("Web page build failed, skipping website deployment")
|
|
||||||
return
|
|
||||||
# if it is not a hugo page, upload it as is
|
|
||||||
files.rsync(
|
|
||||||
f"{www_path}/", "/var/www/html", flags=["-avz", "--chown=www-data"]
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class LegacyRemoveDeployer(Deployer):
|
|
||||||
def install(self):
|
|
||||||
apt.packages(name="Remove rspamd", packages="rspamd", present=False)
|
|
||||||
|
|
||||||
# remove historic expunge script
|
|
||||||
# which is now implemented through a systemd timer (chatmail-expire)
|
|
||||||
self.remove_file("/etc/cron.d/expunge")
|
|
||||||
|
|
||||||
# Remove OBS repository key that is no longer used.
|
|
||||||
self.remove_file("/etc/apt/keyrings/obs-home-deltachat.gpg")
|
|
||||||
self.ensure_line(
|
|
||||||
path="/etc/apt/sources.list",
|
|
||||||
line="deb [signed-by=/etc/apt/keyrings/obs-home-deltachat.gpg] https://download.opensuse.org/repositories/home:/deltachat/Debian_12/ ./",
|
|
||||||
escape_regex_characters=True,
|
|
||||||
present=False,
|
|
||||||
)
|
|
||||||
|
|
||||||
# prior relay versions used filelogging
|
|
||||||
self.remove_directory("/var/log/journal/")
|
|
||||||
# remove echobot if it is still running
|
|
||||||
if has_systemd() and host.get_fact(SystemdEnabled).get("echobot.service"):
|
|
||||||
systemd.service(
|
|
||||||
name="Disable echobot.service",
|
|
||||||
service="echobot.service",
|
|
||||||
running=False,
|
|
||||||
enabled=False,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def check_config(config):
|
|
||||||
mail_domain = config.mail_domain
|
|
||||||
if mail_domain != "testrun.org" and not mail_domain.endswith(".testrun.org"):
|
|
||||||
blocked_words = "merlinux schmieder testrun.org".split()
|
|
||||||
for key in config.__dict__:
|
|
||||||
value = config.__dict__[key]
|
|
||||||
if key.startswith("privacy") and any(
|
|
||||||
x in str(value) for x in blocked_words
|
|
||||||
):
|
|
||||||
raise ValueError(
|
|
||||||
f"please set your own privacy contacts/addresses in {config._inipath}"
|
|
||||||
)
|
|
||||||
return config
|
|
||||||
|
|
||||||
|
|
||||||
class TurnDeployer(Deployer):
|
|
||||||
def __init__(self, mail_domain):
|
|
||||||
self.mail_domain = mail_domain
|
|
||||||
self.units = ["turnserver"]
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
(url, sha256sum) = {
|
|
||||||
"x86_64": (
|
|
||||||
"https://github.com/chatmail/chatmail-turn/releases/download/v0.4/chatmail-turn-x86_64-linux",
|
|
||||||
"1ec1f5c50122165e858a5a91bcba9037a28aa8cb8b64b8db570aa457c6141a8a",
|
|
||||||
),
|
|
||||||
"aarch64": (
|
|
||||||
"https://github.com/chatmail/chatmail-turn/releases/download/v0.4/chatmail-turn-aarch64-linux",
|
|
||||||
"0fb3e792419494e21ecad536464929dba706bb2c88884ed8f1788141d26fc756",
|
|
||||||
),
|
|
||||||
}[host.get_fact(facts.server.Arch)]
|
|
||||||
self.download_executable(url, "/usr/local/bin/chatmail-turn", sha256sum)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
configure_remote_units(self, self.mail_domain, self.units)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
activate_remote_units(self, self.units)
|
|
||||||
|
|
||||||
|
|
||||||
class IrohDeployer(Deployer):
|
|
||||||
def __init__(self, enable_iroh_relay):
|
|
||||||
self.enable_iroh_relay = enable_iroh_relay
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
(url, sha256sum) = {
|
|
||||||
"x86_64": (
|
|
||||||
"https://github.com/n0-computer/iroh/releases/download/v0.35.0/iroh-relay-v0.35.0-x86_64-unknown-linux-musl.tar.gz",
|
|
||||||
"45c81199dbd70f8c4c30fef7f3b9727ca6e3cea8f2831333eeaf8aa71bf0fac1",
|
|
||||||
),
|
|
||||||
"aarch64": (
|
|
||||||
"https://github.com/n0-computer/iroh/releases/download/v0.35.0/iroh-relay-v0.35.0-aarch64-unknown-linux-musl.tar.gz",
|
|
||||||
"f8ef27631fac213b3ef668d02acd5b3e215292746a3fc71d90c63115446008b1",
|
|
||||||
),
|
|
||||||
}[host.get_fact(facts.server.Arch)]
|
|
||||||
self.download_executable(
|
|
||||||
url,
|
|
||||||
"/usr/local/bin/iroh-relay",
|
|
||||||
sha256sum,
|
|
||||||
extract="gunzip | tar -xf - ./iroh-relay -O",
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
self.ensure_systemd_unit("iroh-relay.service")
|
|
||||||
self.put_file("iroh-relay.toml", "/etc/iroh-relay.toml")
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service(
|
|
||||||
"iroh-relay.service",
|
|
||||||
enabled=self.enable_iroh_relay,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class JournaldDeployer(Deployer):
|
|
||||||
def configure(self):
|
|
||||||
self.put_file("journald.conf", "/etc/systemd/journald.conf")
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service("systemd-journald.service")
|
|
||||||
|
|
||||||
|
|
||||||
class ChatmailVenvDeployer(Deployer):
|
|
||||||
def __init__(self, config):
|
|
||||||
self.config = config
|
|
||||||
self.units = (
|
|
||||||
"chatmail-metadata",
|
|
||||||
"lastlogin",
|
|
||||||
"chatmail-expire",
|
|
||||||
"chatmail-expire.timer",
|
|
||||||
"chatmail-fsreport",
|
|
||||||
"chatmail-fsreport.timer",
|
|
||||||
)
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
_install_remote_venv_with_chatmaild(self)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
_configure_remote_venv_with_chatmaild(self, self.config)
|
|
||||||
configure_remote_units(self, self.config.mail_domain, self.units)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
activate_remote_units(self, self.units)
|
|
||||||
|
|
||||||
|
|
||||||
class ChatmailDeployer(Deployer):
|
|
||||||
required_users = [
|
|
||||||
("vmail", "vmail", None),
|
|
||||||
("iroh", None, None),
|
|
||||||
]
|
|
||||||
|
|
||||||
def __init__(self, config):
|
|
||||||
self.config = config
|
|
||||||
self.mail_domain = config.mail_domain
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
self.put_file(
|
|
||||||
src=BytesIO(b'APT::Install-Recommends "false";\n'),
|
|
||||||
dest="/etc/apt/apt.conf.d/00InstallRecommends",
|
|
||||||
)
|
|
||||||
apt.update(name="apt update", cache_time=24 * 3600)
|
|
||||||
apt.upgrade(name="upgrade apt packages", auto_remove=True)
|
|
||||||
|
|
||||||
apt.packages(
|
|
||||||
name="Install curl",
|
|
||||||
packages=["curl"],
|
|
||||||
)
|
|
||||||
|
|
||||||
apt.packages(
|
|
||||||
name="Install rsync",
|
|
||||||
packages=["rsync"],
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
# metadata crashes if the mailboxes dir does not exist
|
|
||||||
self.ensure_directory(
|
|
||||||
str(self.config.mailboxes_dir),
|
|
||||||
owner="vmail",
|
|
||||||
mode="700",
|
|
||||||
)
|
|
||||||
|
|
||||||
# This file is used by auth proxy.
|
|
||||||
# https://wiki.debian.org/EtcMailName
|
|
||||||
server.shell(
|
|
||||||
name="Setup /etc/mailname",
|
|
||||||
commands=[
|
|
||||||
f"echo {self.mail_domain} >/etc/mailname; chmod 644 /etc/mailname"
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class FcgiwrapDeployer(Deployer):
|
|
||||||
def install(self):
|
|
||||||
apt.packages(
|
|
||||||
name="Install fcgiwrap",
|
|
||||||
packages=["fcgiwrap"],
|
|
||||||
)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service("fcgiwrap.service")
|
|
||||||
|
|
||||||
|
|
||||||
class GithashDeployer(Deployer):
|
|
||||||
def activate(self):
|
|
||||||
try:
|
|
||||||
git_hash = subprocess.check_output(["git", "rev-parse", "HEAD"]).decode()
|
|
||||||
except Exception:
|
|
||||||
git_hash = "unknown\n"
|
|
||||||
try:
|
|
||||||
git_diff = subprocess.check_output(["git", "diff"]).decode()
|
|
||||||
except Exception:
|
|
||||||
git_diff = ""
|
|
||||||
self.put_file(src=StringIO(git_hash + git_diff), dest="/etc/chatmail-version")
|
|
||||||
|
|
||||||
|
|
||||||
def get_tls_deployer(config, mail_domain):
|
|
||||||
"""Select the appropriate TLS deployer based on config."""
|
|
||||||
tls_domains = [mail_domain, f"mta-sts.{mail_domain}", f"www.{mail_domain}"]
|
|
||||||
|
|
||||||
if config.tls_cert_mode == "acme":
|
|
||||||
return AcmetoolDeployer(config.acme_email, tls_domains)
|
|
||||||
elif config.tls_cert_mode == "self":
|
|
||||||
return SelfSignedTlsDeployer(mail_domain)
|
|
||||||
elif config.tls_cert_mode == "external":
|
|
||||||
return ExternalTlsDeployer(config.tls_cert_path, config.tls_key_path)
|
|
||||||
else:
|
|
||||||
raise ValueError(f"Unknown tls_cert_mode: {config.tls_cert_mode}")
|
|
||||||
|
|
||||||
|
|
||||||
def deploy_chatmail(config_path: Path, disable_mail: bool, website_only: bool) -> None:
|
|
||||||
"""Deploy a chat-mail instance.
|
|
||||||
|
|
||||||
:param config_path: path to chatmail.ini
|
|
||||||
:param disable_mail: whether to disable postfix & dovecot
|
|
||||||
:param website_only: if True, only deploy the website
|
|
||||||
"""
|
|
||||||
config = read_config(config_path)
|
|
||||||
check_config(config)
|
|
||||||
mail_domain = config.mail_domain
|
|
||||||
|
|
||||||
if website_only:
|
|
||||||
Deployment().perform_stages([WebsiteDeployer(config)])
|
|
||||||
return
|
|
||||||
|
|
||||||
# Check if mtail_address interface is available (if configured)
|
|
||||||
if config.mtail_address and config.mtail_address not in (
|
|
||||||
"127.0.0.1",
|
|
||||||
"::1",
|
|
||||||
"localhost",
|
|
||||||
):
|
|
||||||
ipv4_addrs = host.get_fact(hardware.Ipv4Addrs)
|
|
||||||
all_addresses = [addr for addrs in ipv4_addrs.values() for addr in addrs]
|
|
||||||
if config.mtail_address not in all_addresses:
|
|
||||||
Out().red(
|
|
||||||
f"Deploy failed: mtail_address {config.mtail_address} is not available (VPN up?).\n"
|
|
||||||
)
|
|
||||||
exit(1)
|
|
||||||
|
|
||||||
if not is_in_container():
|
|
||||||
port_services = [
|
|
||||||
(["master", "smtpd"], 25),
|
|
||||||
("unbound", 53),
|
|
||||||
]
|
|
||||||
if config.tls_cert_mode == "acme":
|
|
||||||
port_services.append(("acmetool", 402))
|
|
||||||
port_services += [
|
|
||||||
(["imap-login", "dovecot"], 143),
|
|
||||||
# acmetool previously listened on port 80,
|
|
||||||
# so don't complain during upgrade that moved it to port 402
|
|
||||||
# and gave the port to nginx.
|
|
||||||
(["acmetool", "nginx"], 80),
|
|
||||||
("nginx", 443),
|
|
||||||
(["master", "smtpd"], 465),
|
|
||||||
(["master", "smtpd"], 587),
|
|
||||||
(["imap-login", "dovecot"], 993),
|
|
||||||
("iroh-relay", 3340),
|
|
||||||
("mtail", 3903),
|
|
||||||
("stats", 3904),
|
|
||||||
("nginx", 8443),
|
|
||||||
(["master", "smtpd"], config.postfix_reinject_port),
|
|
||||||
(["master", "smtpd"], config.postfix_reinject_port_incoming),
|
|
||||||
("filtermail", config.filtermail_smtp_port),
|
|
||||||
("filtermail", config.filtermail_smtp_port_incoming),
|
|
||||||
]
|
|
||||||
for service, port in port_services:
|
|
||||||
print(f"Checking if port {port} is available for {service}...")
|
|
||||||
running_service = host.get_fact(Port, port=port)
|
|
||||||
services = [service] if isinstance(service, str) else service
|
|
||||||
if running_service:
|
|
||||||
if running_service not in services:
|
|
||||||
Out().red(
|
|
||||||
f"Deploy failed: port {port} is occupied by: {running_service}"
|
|
||||||
)
|
|
||||||
exit(1)
|
|
||||||
|
|
||||||
tls_deployer = get_tls_deployer(config, mail_domain)
|
|
||||||
|
|
||||||
all_deployers = [
|
|
||||||
ChatmailDeployer(config),
|
|
||||||
LegacyRemoveDeployer(),
|
|
||||||
FiltermailDeployer(),
|
|
||||||
JournaldDeployer(),
|
|
||||||
UnboundDeployer(config),
|
|
||||||
TurnDeployer(mail_domain),
|
|
||||||
IrohDeployer(config.enable_iroh_relay),
|
|
||||||
tls_deployer,
|
|
||||||
WebsiteDeployer(config),
|
|
||||||
ChatmailVenvDeployer(config),
|
|
||||||
MtastsDeployer(),
|
|
||||||
OpendkimDeployer(mail_domain),
|
|
||||||
# Dovecot should be started before Postfix
|
|
||||||
# because it creates authentication socket
|
|
||||||
# required by Postfix.
|
|
||||||
DovecotDeployer(config, disable_mail),
|
|
||||||
PostfixDeployer(config, disable_mail),
|
|
||||||
FcgiwrapDeployer(),
|
|
||||||
NginxDeployer(config),
|
|
||||||
MtailDeployer(config.mtail_address),
|
|
||||||
GithashDeployer(),
|
|
||||||
]
|
|
||||||
|
|
||||||
Deployment().perform_stages(all_deployers)
|
|
||||||
@@ -1,36 +1,25 @@
|
|||||||
import datetime
|
import datetime
|
||||||
|
import importlib
|
||||||
|
|
||||||
|
from jinja2 import Template
|
||||||
|
|
||||||
from . import remote
|
from . import remote
|
||||||
|
|
||||||
|
|
||||||
def parse_zone_records(text):
|
|
||||||
"""Yield ``(name, ttl, rtype, rdata)`` from standard BIND-format text."""
|
|
||||||
for raw_line in text.splitlines():
|
|
||||||
line = raw_line.strip()
|
|
||||||
if not line or line.startswith(";"):
|
|
||||||
continue
|
|
||||||
try:
|
|
||||||
name, ttl, _in, rtype, rdata = line.split(None, 4)
|
|
||||||
except ValueError:
|
|
||||||
raise ValueError(f"Bad zone record line: {line!r}") from None
|
|
||||||
name = name.rstrip(".")
|
|
||||||
yield name, ttl, rtype.upper(), rdata
|
|
||||||
|
|
||||||
|
|
||||||
def get_initial_remote_data(sshexec, mail_domain):
|
def get_initial_remote_data(sshexec, mail_domain):
|
||||||
return sshexec.logged(
|
return sshexec.logged(
|
||||||
call=remote.rdns.perform_initial_checks, kwargs=dict(mail_domain=mail_domain)
|
call=remote.rdns.perform_initial_checks, kwargs=dict(mail_domain=mail_domain)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def check_initial_remote_data(remote_data, *, strict_tls=True, print=print):
|
def check_initial_remote_data(remote_data, *, print=print):
|
||||||
mail_domain = remote_data["mail_domain"]
|
mail_domain = remote_data["mail_domain"]
|
||||||
if not remote_data["A"] and not remote_data["AAAA"]:
|
if not remote_data["A"] and not remote_data["AAAA"]:
|
||||||
print(f"Missing A and/or AAAA DNS records for {mail_domain}!")
|
print(f"Missing A and/or AAAA DNS records for {mail_domain}!")
|
||||||
elif strict_tls and remote_data["MTA_STS"] != f"{mail_domain}.":
|
elif remote_data["MTA_STS"] != f"{mail_domain}.":
|
||||||
print("Missing MTA-STS CNAME record:")
|
print("Missing MTA-STS CNAME record:")
|
||||||
print(f"mta-sts.{mail_domain}. CNAME {mail_domain}.")
|
print(f"mta-sts.{mail_domain}. CNAME {mail_domain}.")
|
||||||
elif strict_tls and remote_data["WWW"] != f"{mail_domain}.":
|
elif remote_data["WWW"] != f"{mail_domain}.":
|
||||||
print("Missing www CNAME record:")
|
print("Missing www CNAME record:")
|
||||||
print(f"www.{mail_domain}. CNAME {mail_domain}.")
|
print(f"www.{mail_domain}. CNAME {mail_domain}.")
|
||||||
else:
|
else:
|
||||||
@@ -42,39 +31,13 @@ def get_filled_zone_file(remote_data):
|
|||||||
if not sts_id:
|
if not sts_id:
|
||||||
remote_data["sts_id"] = datetime.datetime.now().strftime("%Y%m%d%H%M")
|
remote_data["sts_id"] = datetime.datetime.now().strftime("%Y%m%d%H%M")
|
||||||
|
|
||||||
d = remote_data["mail_domain"]
|
template = importlib.resources.files(__package__).joinpath("chatmail.zone.j2")
|
||||||
|
content = template.read_text()
|
||||||
def append_record(name, rtype, rdata, ttl=3600):
|
zonefile = Template(content).render(**remote_data)
|
||||||
lines.append(f"{name:<40} {ttl:<6} IN {rtype:<5} {rdata}")
|
lines = [x.strip() for x in zonefile.split("\n") if x.strip()]
|
||||||
|
|
||||||
lines = ["; Required DNS entries"]
|
|
||||||
if remote_data.get("A"):
|
|
||||||
append_record(f"{d}.", "A", remote_data["A"])
|
|
||||||
if remote_data.get("AAAA"):
|
|
||||||
append_record(f"{d}.", "AAAA", remote_data["AAAA"])
|
|
||||||
append_record(f"{d}.", "MX", f"10 {d}.")
|
|
||||||
if remote_data.get("strict_tls"):
|
|
||||||
append_record(f"_mta-sts.{d}.", "TXT", f'"v=STSv1; id={remote_data["sts_id"]}"')
|
|
||||||
append_record(f"mta-sts.{d}.", "CNAME", f"{d}.")
|
|
||||||
append_record(f"www.{d}.", "CNAME", f"{d}.")
|
|
||||||
lines.append(remote_data["dkim_entry"])
|
|
||||||
lines.append("")
|
lines.append("")
|
||||||
lines.append("; Recommended DNS entries")
|
zonefile = "\n".join(lines)
|
||||||
append_record(f"{d}.", "TXT", '"v=spf1 a ~all"')
|
return zonefile
|
||||||
append_record(f"_dmarc.{d}.", "TXT", '"v=DMARC1;p=reject;adkim=s;aspf=s"')
|
|
||||||
if remote_data.get("acme_account_url"):
|
|
||||||
append_record(
|
|
||||||
f"{d}.",
|
|
||||||
"CAA",
|
|
||||||
f'0 issue "letsencrypt.org;accounturi={remote_data["acme_account_url"]}"',
|
|
||||||
)
|
|
||||||
append_record(f"_adsp._domainkey.{d}.", "TXT", '"dkim=discardable"')
|
|
||||||
append_record(f"_submission._tcp.{d}.", "SRV", f"0 1 587 {d}.")
|
|
||||||
append_record(f"_submissions._tcp.{d}.", "SRV", f"0 1 465 {d}.")
|
|
||||||
append_record(f"_imap._tcp.{d}.", "SRV", f"0 1 143 {d}.")
|
|
||||||
append_record(f"_imaps._tcp.{d}.", "SRV", f"0 1 993 {d}.")
|
|
||||||
lines.append("")
|
|
||||||
return "\n".join(lines)
|
|
||||||
|
|
||||||
|
|
||||||
def check_full_zone(sshexec, remote_data, out, zonefile) -> int:
|
def check_full_zone(sshexec, remote_data, out, zonefile) -> int:
|
||||||
@@ -83,7 +46,7 @@ def check_full_zone(sshexec, remote_data, out, zonefile) -> int:
|
|||||||
|
|
||||||
required_diff, recommended_diff = sshexec.logged(
|
required_diff, recommended_diff = sshexec.logged(
|
||||||
remote.rdns.check_zonefile,
|
remote.rdns.check_zonefile,
|
||||||
kwargs=dict(zonefile=zonefile, verbose=False),
|
kwargs=dict(zonefile=zonefile, mail_domain=remote_data["mail_domain"]),
|
||||||
)
|
)
|
||||||
|
|
||||||
returncode = 0
|
returncode = 0
|
||||||
@@ -95,8 +58,7 @@ def check_full_zone(sshexec, remote_data, out, zonefile) -> int:
|
|||||||
returncode = 1
|
returncode = 1
|
||||||
if remote_data.get("dkim_entry") in required_diff:
|
if remote_data.get("dkim_entry") in required_diff:
|
||||||
out(
|
out(
|
||||||
"If the DKIM entry above does not work with your DNS provider,"
|
"If the DKIM entry above does not work with your DNS provider, you can try this one:\n"
|
||||||
" you can try this one:\n"
|
|
||||||
)
|
)
|
||||||
out(remote_data.get("web_dkim_entry") + "\n")
|
out(remote_data.get("web_dkim_entry") + "\n")
|
||||||
if recommended_diff:
|
if recommended_diff:
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ iterate_prefix = userdb/
|
|||||||
|
|
||||||
default_pass_scheme = plain
|
default_pass_scheme = plain
|
||||||
# %E escapes characters " (double quote), ' (single quote) and \ (backslash) with \ (backslash).
|
# %E escapes characters " (double quote), ' (single quote) and \ (backslash) with \ (backslash).
|
||||||
# See <https://doc.dovecot.org/2.3/configuration_manual/config_file/config_variables/#modifiers>
|
# See <https://doc.dovecot.org/configuration_manual/config_file/config_variables/#modifiers>
|
||||||
# for documentation.
|
# for documentation.
|
||||||
#
|
#
|
||||||
# We escape user-provided input and use double quote as a separator.
|
# We escape user-provided input and use double quote as a separator.
|
||||||
|
|||||||
@@ -1,189 +0,0 @@
|
|||||||
import io
|
|
||||||
import urllib.request
|
|
||||||
|
|
||||||
from chatmaild.config import Config
|
|
||||||
from pyinfra import host
|
|
||||||
from pyinfra.facts.deb import DebPackages
|
|
||||||
from pyinfra.facts.server import Arch, Command, Sysctl
|
|
||||||
from pyinfra.operations import apt, files, server
|
|
||||||
|
|
||||||
from cmdeploy.basedeploy import (
|
|
||||||
Deployer,
|
|
||||||
activate_remote_units,
|
|
||||||
blocked_service_startup,
|
|
||||||
configure_remote_units,
|
|
||||||
is_in_container,
|
|
||||||
)
|
|
||||||
|
|
||||||
DOVECOT_ARCHIVE_VERSION = "2.3.21+dfsg1-3"
|
|
||||||
DOVECOT_PACKAGE_VERSION = f"1:{DOVECOT_ARCHIVE_VERSION}"
|
|
||||||
|
|
||||||
DOVECOT_SHA256 = {
|
|
||||||
("core", "amd64"): "dd060706f52a306fa863d874717210b9fe10536c824afe1790eec247ded5b27d",
|
|
||||||
("core", "arm64"): "e7548e8a82929722e973629ecc40fcfa886894cef3db88f23535149e7f730dc9",
|
|
||||||
("imapd", "amd64"): "8d8dc6fc00bbb6cdb25d345844f41ce2f1c53f764b79a838eb2a03103eebfa86",
|
|
||||||
("imapd", "arm64"): "178fa877ddd5df9930e8308b518f4b07df10e759050725f8217a0c1fb3fd707f",
|
|
||||||
("lmtpd", "amd64"): "2f69ba5e35363de50962d42cccbfe4ed8495265044e244007d7ccddad77513ab",
|
|
||||||
("lmtpd", "arm64"): "89f52fb36524f5877a177dff4a713ba771fd3f91f22ed0af7238d495e143b38f",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class DovecotDeployer(Deployer):
|
|
||||||
daemon_reload = False
|
|
||||||
|
|
||||||
def __init__(self, config, disable_mail):
|
|
||||||
self.config = config
|
|
||||||
self.disable_mail = disable_mail
|
|
||||||
self.units = ["doveauth"]
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
arch = host.get_fact(Arch)
|
|
||||||
with blocked_service_startup():
|
|
||||||
debs = []
|
|
||||||
for pkg in ("core", "imapd", "lmtpd"):
|
|
||||||
deb, changed = _download_dovecot_package(pkg, arch)
|
|
||||||
self.need_restart |= changed
|
|
||||||
if deb:
|
|
||||||
debs.append(deb)
|
|
||||||
if debs:
|
|
||||||
deb_list = " ".join(debs)
|
|
||||||
# First dpkg may fail on missing dependencies (stderr suppressed);
|
|
||||||
# apt-get --fix-broken pulls them in, then dpkg retries cleanly.
|
|
||||||
server.shell(
|
|
||||||
name="Install dovecot packages",
|
|
||||||
commands=[
|
|
||||||
f"dpkg --force-confdef --force-confold -i {deb_list} 2> /dev/null || true",
|
|
||||||
"DEBIAN_FRONTEND=noninteractive apt-get -y --fix-broken install",
|
|
||||||
f"dpkg --force-confdef --force-confold -i {deb_list}",
|
|
||||||
],
|
|
||||||
)
|
|
||||||
self.need_restart = True
|
|
||||||
self.put_file(
|
|
||||||
src=io.StringIO(
|
|
||||||
"Package: dovecot-*\n"
|
|
||||||
"Pin: version *\n"
|
|
||||||
"Pin-Priority: -1\n"
|
|
||||||
),
|
|
||||||
dest="/etc/apt/preferences.d/pin-dovecot",
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
configure_remote_units(self, self.config.mail_domain, self.units)
|
|
||||||
_configure_dovecot(self, self.config)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
activate_remote_units(self, self.units)
|
|
||||||
|
|
||||||
# Detect stale binary: package installed but service still runs old (deleted) binary.
|
|
||||||
if not self.disable_mail and not self.need_restart:
|
|
||||||
stale = host.get_fact(
|
|
||||||
Command,
|
|
||||||
'pid=$(systemctl show -p MainPID --value dovecot.service 2>/dev/null);'
|
|
||||||
' [ "${pid:-0}" != "0" ] && readlink "/proc/$pid/exe" 2>/dev/null | grep -q "(deleted)"'
|
|
||||||
" && echo STALE || true",
|
|
||||||
)
|
|
||||||
if stale == "STALE":
|
|
||||||
self.need_restart = True
|
|
||||||
|
|
||||||
active = not self.disable_mail
|
|
||||||
self.ensure_service(
|
|
||||||
"dovecot.service",
|
|
||||||
running=active,
|
|
||||||
enabled=active,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def _pick_url(primary, fallback):
|
|
||||||
try:
|
|
||||||
req = urllib.request.Request(primary, method="HEAD")
|
|
||||||
urllib.request.urlopen(req, timeout=10)
|
|
||||||
return primary
|
|
||||||
except Exception:
|
|
||||||
return fallback
|
|
||||||
|
|
||||||
|
|
||||||
def _download_dovecot_package(package: str, arch: str) -> tuple[str | None, bool]:
|
|
||||||
"""Download a dovecot .deb if needed, return (path, changed)."""
|
|
||||||
arch = "amd64" if arch == "x86_64" else arch
|
|
||||||
arch = "arm64" if arch == "aarch64" else arch
|
|
||||||
|
|
||||||
pkg_name = f"dovecot-{package}"
|
|
||||||
sha256 = DOVECOT_SHA256.get((package, arch))
|
|
||||||
if sha256 is None:
|
|
||||||
op = apt.packages(packages=[pkg_name])
|
|
||||||
return None, bool(getattr(op, "changed", False))
|
|
||||||
|
|
||||||
installed_versions = host.get_fact(DebPackages).get(pkg_name, [])
|
|
||||||
if DOVECOT_PACKAGE_VERSION in installed_versions:
|
|
||||||
return None, False
|
|
||||||
|
|
||||||
url_version = DOVECOT_ARCHIVE_VERSION.replace("+", "%2B")
|
|
||||||
deb_base = f"{pkg_name}_{url_version}_{arch}.deb"
|
|
||||||
primary_url = f"https://download.delta.chat/dovecot/{deb_base}"
|
|
||||||
fallback_url = f"https://github.com/chatmail/dovecot/releases/download/upstream%2F{url_version}/{deb_base}"
|
|
||||||
url = _pick_url(primary_url, fallback_url)
|
|
||||||
deb_filename = f"/root/{deb_base}"
|
|
||||||
|
|
||||||
files.download(
|
|
||||||
name=f"Download {pkg_name}",
|
|
||||||
src=url,
|
|
||||||
dest=deb_filename,
|
|
||||||
sha256sum=sha256,
|
|
||||||
cache_time=60 * 60 * 24 * 365 * 10, # never redownload the package
|
|
||||||
)
|
|
||||||
|
|
||||||
return deb_filename, True
|
|
||||||
|
|
||||||
def _configure_dovecot(deployer, config: Config, debug: bool = False):
|
|
||||||
"""Configures Dovecot IMAP server."""
|
|
||||||
deployer.put_template(
|
|
||||||
"dovecot/dovecot.conf.j2",
|
|
||||||
"/etc/dovecot/dovecot.conf",
|
|
||||||
config=config,
|
|
||||||
debug=debug,
|
|
||||||
disable_ipv6=config.disable_ipv6,
|
|
||||||
)
|
|
||||||
deployer.put_file("dovecot/auth.conf", "/etc/dovecot/auth.conf")
|
|
||||||
deployer.put_file(
|
|
||||||
"dovecot/push_notification.lua", "/etc/dovecot/push_notification.lua"
|
|
||||||
)
|
|
||||||
|
|
||||||
# as per https://doc.dovecot.org/2.3/configuration_manual/os/
|
|
||||||
# it is recommended to set the following inotify limits
|
|
||||||
can_modify = not is_in_container()
|
|
||||||
for name in ("max_user_instances", "max_user_watches"):
|
|
||||||
key = f"fs.inotify.{name}"
|
|
||||||
value = host.get_fact(Sysctl).get(key, 0)
|
|
||||||
if value > 65534:
|
|
||||||
continue
|
|
||||||
if not can_modify:
|
|
||||||
print(
|
|
||||||
"\n!!!! refusing to attempt sysctl setting in containers\n"
|
|
||||||
f"!!!! dovecot: sysctl {key!r}={value}, should be >65534 for production setups\n"
|
|
||||||
"!!!!"
|
|
||||||
)
|
|
||||||
continue
|
|
||||||
server.sysctl(
|
|
||||||
name=f"Change {key}",
|
|
||||||
key=key,
|
|
||||||
value=65535,
|
|
||||||
persist=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.ensure_line(
|
|
||||||
name="Set TZ environment variable",
|
|
||||||
path="/etc/environment",
|
|
||||||
line="TZ=:/etc/localtime",
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.put_file(
|
|
||||||
"service/10_restart_on_failure.conf",
|
|
||||||
"/etc/systemd/system/dovecot.service.d/10_restart.conf",
|
|
||||||
)
|
|
||||||
|
|
||||||
# Validate dovecot configuration before restart
|
|
||||||
if deployer.need_restart:
|
|
||||||
server.shell(
|
|
||||||
name="Validate dovecot configuration",
|
|
||||||
commands=["doveconf -n >/dev/null"],
|
|
||||||
)
|
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
## Dovecot configuration file
|
## Dovecot configuration file
|
||||||
|
|
||||||
{% if disable_ipv6 %}
|
{% if disable_ipv6 %}
|
||||||
listen = 0.0.0.0
|
listen = *
|
||||||
{% endif %}
|
{% endif %}
|
||||||
|
|
||||||
protocols = imap lmtp
|
protocols = imap lmtp
|
||||||
@@ -26,7 +26,7 @@ default_client_limit = 20000
|
|||||||
# Increase number of logged in IMAP connections.
|
# Increase number of logged in IMAP connections.
|
||||||
# Each connection is handled by a separate `imap` process.
|
# Each connection is handled by a separate `imap` process.
|
||||||
# `imap` process should have `client_limit=1` as described in
|
# `imap` process should have `client_limit=1` as described in
|
||||||
# <https://doc.dovecot.org/2.3/configuration_manual/service_configuration/#service-limits>
|
# <https://doc.dovecot.org/configuration_manual/service_configuration/#service-limits>
|
||||||
# so each logged in IMAP session will need its own `imap` process.
|
# so each logged in IMAP session will need its own `imap` process.
|
||||||
#
|
#
|
||||||
# If this limit is reached,
|
# If this limit is reached,
|
||||||
@@ -44,11 +44,11 @@ mail_server_comment = Chatmail server
|
|||||||
|
|
||||||
# `zlib` enables compressing messages stored in the maildir.
|
# `zlib` enables compressing messages stored in the maildir.
|
||||||
# See
|
# See
|
||||||
# <https://doc.dovecot.org/2.3/configuration_manual/zlib_plugin/>
|
# <https://doc.dovecot.org/configuration_manual/zlib_plugin/>
|
||||||
# for documentation.
|
# for documentation.
|
||||||
#
|
#
|
||||||
# quota plugin documentation:
|
# quota plugin documentation:
|
||||||
# <https://doc.dovecot.org/2.3/configuration_manual/quota_plugin/>
|
# <https://doc.dovecot.org/configuration_manual/quota_plugin/>
|
||||||
mail_plugins = zlib quota
|
mail_plugins = zlib quota
|
||||||
|
|
||||||
imap_capability = +XDELTAPUSH XCHATMAIL
|
imap_capability = +XDELTAPUSH XCHATMAIL
|
||||||
@@ -70,12 +70,6 @@ userdb {
|
|||||||
# Mailboxes are stored in the "mail" directory of the vmail user home.
|
# Mailboxes are stored in the "mail" directory of the vmail user home.
|
||||||
mail_location = maildir:{{ config.mailboxes_dir }}/%u
|
mail_location = maildir:{{ config.mailboxes_dir }}/%u
|
||||||
|
|
||||||
# index/cache files are not very useful for chatmail relay operations
|
|
||||||
# but it's not clear how to disable them completely.
|
|
||||||
# According to https://doc.dovecot.org/2.3/settings/advanced/#core_setting-mail_cache_max_size
|
|
||||||
# if the cache file becomes larger than the specified size, it is truncated by dovecot
|
|
||||||
mail_cache_max_size = 500K
|
|
||||||
|
|
||||||
namespace inbox {
|
namespace inbox {
|
||||||
inbox = yes
|
inbox = yes
|
||||||
|
|
||||||
@@ -113,7 +107,7 @@ mail_attribute_dict = proxy:/run/chatmail-metadata/metadata.socket:metadata
|
|||||||
# `imap_zlib` enables IMAP COMPRESS (RFC 4978).
|
# `imap_zlib` enables IMAP COMPRESS (RFC 4978).
|
||||||
# <https://datatracker.ietf.org/doc/html/rfc4978.html>
|
# <https://datatracker.ietf.org/doc/html/rfc4978.html>
|
||||||
protocol imap {
|
protocol imap {
|
||||||
mail_plugins = $mail_plugins imap_quota last_login {% if config.imap_compress %}imap_zlib{% endif %}
|
mail_plugins = $mail_plugins imap_zlib imap_quota last_login
|
||||||
imap_metadata = yes
|
imap_metadata = yes
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -125,19 +119,14 @@ plugin {
|
|||||||
|
|
||||||
protocol lmtp {
|
protocol lmtp {
|
||||||
# notify plugin is a dependency of push_notification plugin:
|
# notify plugin is a dependency of push_notification plugin:
|
||||||
# <https://doc.dovecot.org/2.3/settings/plugin/notify-plugin/>
|
# <https://doc.dovecot.org/settings/plugin/notify-plugin/>
|
||||||
#
|
#
|
||||||
# push_notification plugin documentation:
|
# push_notification plugin documentation:
|
||||||
# <https://doc.dovecot.org/2.3/configuration_manual/push_notification/>
|
# <https://doc.dovecot.org/configuration_manual/push_notification/>
|
||||||
#
|
#
|
||||||
# mail_lua and push_notification_lua are needed for Lua push notification handler.
|
# mail_lua and push_notification_lua are needed for Lua push notification handler.
|
||||||
# <https://doc.dovecot.org/2.3/configuration_manual/push_notification/#configuration>
|
# <https://doc.dovecot.org/configuration_manual/push_notification/#configuration>
|
||||||
mail_plugins = $mail_plugins mail_lua notify push_notification push_notification_lua
|
mail_plugins = $mail_plugins mail_lua notify push_notification push_notification_lua
|
||||||
|
|
||||||
# Disable fsync for LMTP. May lose delivered message,
|
|
||||||
# but unlikely to cause problems with multiple relays.
|
|
||||||
# https://doc.dovecot.org/2.3/admin_manual/mailbox_formats/#fsyncing
|
|
||||||
mail_fsync = never
|
|
||||||
}
|
}
|
||||||
|
|
||||||
plugin {
|
plugin {
|
||||||
@@ -149,31 +138,17 @@ plugin {
|
|||||||
}
|
}
|
||||||
|
|
||||||
plugin {
|
plugin {
|
||||||
|
# for now we define static quota-rules for all users
|
||||||
quota = maildir:User quota
|
quota = maildir:User quota
|
||||||
|
quota_rule = *:storage={{ config.max_mailbox_size }}
|
||||||
quota_max_mail_size={{ config.max_message_size }}
|
quota_max_mail_size={{ config.max_message_size }}
|
||||||
quota_grace = 0
|
quota_grace = 0
|
||||||
|
# quota_over_flag_value = TRUE
|
||||||
quota_rule = *:storage={{ config.max_mailbox_size_mb }}M
|
|
||||||
|
|
||||||
# Trigger at 75%% of quota, expire oldest messages down to 70%%.
|
|
||||||
# The percentages are chosen to prevent current Delta Chat users
|
|
||||||
# from seeing "quota warnings" which trigger at 80% and 95%.
|
|
||||||
|
|
||||||
quota_warning = storage=75%% quota-warning {{ config.max_mailbox_size_mb * 70 // 100 }} {{ config.mailboxes_dir }}/%u
|
|
||||||
}
|
|
||||||
|
|
||||||
service quota-warning {
|
|
||||||
executable = script /usr/local/lib/chatmaild/venv/bin/chatmail-quota-expire
|
|
||||||
user = vmail
|
|
||||||
unix_listener quota-warning {
|
|
||||||
user = vmail
|
|
||||||
mode = 0600
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
# push_notification configuration
|
# push_notification configuration
|
||||||
plugin {
|
plugin {
|
||||||
# <https://doc.dovecot.org/2.3/configuration_manual/push_notification/#lua-lua>
|
# <https://doc.dovecot.org/configuration_manual/push_notification/#lua-lua>
|
||||||
push_notification_driver = lua:file=/etc/dovecot/push_notification.lua
|
push_notification_driver = lua:file=/etc/dovecot/push_notification.lua
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -187,8 +162,6 @@ service lmtp {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
lmtp_add_received_header = no
|
|
||||||
|
|
||||||
service auth {
|
service auth {
|
||||||
unix_listener /var/spool/postfix/private/auth {
|
unix_listener /var/spool/postfix/private/auth {
|
||||||
mode = 0660
|
mode = 0660
|
||||||
@@ -204,34 +177,20 @@ service auth-worker {
|
|||||||
}
|
}
|
||||||
|
|
||||||
service imap-login {
|
service imap-login {
|
||||||
# High-performance mode as described in
|
# High-security mode.
|
||||||
# <https://doc.dovecot.org/2.3/admin_manual/login_processes/#high-performance-mode>
|
# Each process serves a single connection and exits afterwards.
|
||||||
#
|
# This is the default, but we set it explicitly to be sure.
|
||||||
# So-called high-security mode described in
|
# See <https://doc.dovecot.org/admin_manual/login_processes/#high-security-mode> for details.
|
||||||
# <https://doc.dovecot.org/2.3/admin_manual/login_processes/#high-security-mode>
|
service_count = 1
|
||||||
# and enabled by default with `service_count = 1` starts one process per connection
|
|
||||||
# and has problems logging in thousands of users after Dovecot restart.
|
|
||||||
service_count = 0
|
|
||||||
|
|
||||||
# Increase virtual memory size limit.
|
# Inrease the number of simultaneous connections.
|
||||||
# Since imap-login processes handle TLS connections
|
|
||||||
# even after logging users in
|
|
||||||
# and many connections are handled by each process,
|
|
||||||
# memory size limit should be increased.
|
|
||||||
#
|
#
|
||||||
# Otherwise the whole process eventually dies
|
# As of Dovecot 2.3.19.1 the default is 100 processes.
|
||||||
# with an error similar to
|
# Combined with `service_count = 1` it means only 100 connections
|
||||||
# imap-login: Fatal: master: service(imap-login):
|
# can be handled simultaneously.
|
||||||
# child 1422951 returned error 83
|
process_limit = 10000
|
||||||
# (Out of memory (service imap-login { vsz_limit=256 MB },
|
|
||||||
# you may need to increase it)
|
|
||||||
# and takes down all its TLS connections at once.
|
|
||||||
vsz_limit = 1G
|
|
||||||
|
|
||||||
# Avoid startup latency for new connections.
|
# Avoid startup latency for new connections.
|
||||||
#
|
|
||||||
# Should be set to at least the number of CPU cores
|
|
||||||
# according to the documentation.
|
|
||||||
process_min_avail = 10
|
process_min_avail = 10
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -247,8 +206,8 @@ service anvil {
|
|||||||
}
|
}
|
||||||
|
|
||||||
ssl = required
|
ssl = required
|
||||||
ssl_cert = <{{ config.tls_cert_path }}
|
ssl_cert = </var/lib/acme/live/{{ config.mail_domain }}/fullchain
|
||||||
ssl_key = <{{ config.tls_key_path }}
|
ssl_key = </var/lib/acme/live/{{ config.mail_domain }}/privkey
|
||||||
ssl_dh = </usr/share/dovecot/dh.pem
|
ssl_dh = </usr/share/dovecot/dh.pem
|
||||||
ssl_min_protocol = TLSv1.3
|
ssl_min_protocol = TLSv1.3
|
||||||
ssl_prefer_server_ciphers = yes
|
ssl_prefer_server_ciphers = yes
|
||||||
@@ -271,186 +230,5 @@ protocol imap {
|
|||||||
# sort -sn <(sed 's/ / C: /' *.in) <(sed 's/ / S: /' cat *.out)
|
# sort -sn <(sed 's/ / C: /' *.in) <(sed 's/ / S: /' cat *.out)
|
||||||
|
|
||||||
rawlog_dir = %h
|
rawlog_dir = %h
|
||||||
|
|
||||||
# Disable fsync for IMAP. May lose IMAP changes like setting flags.
|
|
||||||
mail_fsync = never
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if not config.imap_compress %}
|
|
||||||
# Hibernate IDLE users to save memory and CPU resources
|
|
||||||
# NOTE: this will have no effect if imap_zlib plugin is used
|
|
||||||
imap_hibernate_timeout = 30s
|
|
||||||
service imap {
|
|
||||||
# Note that this change will allow any process running as
|
|
||||||
# $default_internal_user (dovecot) to access mails as any other user.
|
|
||||||
# This may be insecure in some installations, which is why this isn't
|
|
||||||
# done by default.
|
|
||||||
unix_listener imap-master {
|
|
||||||
user = $default_internal_user
|
|
||||||
}
|
|
||||||
}
|
|
||||||
# The following is the default already in v2.3.1+:
|
|
||||||
service imap {
|
|
||||||
extra_groups = $default_internal_group
|
|
||||||
}
|
|
||||||
service imap-hibernate {
|
|
||||||
unix_listener imap-hibernate {
|
|
||||||
mode = 0660
|
|
||||||
group = $default_internal_group
|
|
||||||
}
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if config.mtail_address %}
|
|
||||||
#
|
|
||||||
# Dovecot Statistics
|
|
||||||
#
|
|
||||||
# OpenMetrics endpoint at http://{{- config.mtail_address}}:3904/metrics
|
|
||||||
service stats {
|
|
||||||
inet_listener http {
|
|
||||||
port = 3904
|
|
||||||
address = {{- config.mtail_address}}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
# IMAP Command Metrics
|
|
||||||
# - Bytes in/out for compression efficiency analysis
|
|
||||||
# - Lock wait time for contention debugging
|
|
||||||
# - Grouped by command name and reply state
|
|
||||||
metric imap_command {
|
|
||||||
filter = event=imap_command_finished
|
|
||||||
fields = bytes_in bytes_out lock_wait_usecs running_usecs
|
|
||||||
group_by = cmd_name tagged_reply_state
|
|
||||||
}
|
|
||||||
|
|
||||||
# Duration buckets for latency histograms (base 10: 10us, 100us, 1ms, 10ms, 100ms, 1s, 10s, 100s)
|
|
||||||
metric imap_command_duration {
|
|
||||||
filter = event=imap_command_finished
|
|
||||||
group_by = cmd_name duration:exponential:1:8:10
|
|
||||||
}
|
|
||||||
|
|
||||||
# Slow command outliers (>1 second = 1000000 usecs)
|
|
||||||
# Useful for alerting without high cardinality
|
|
||||||
metric imap_command_slow {
|
|
||||||
filter = event=imap_command_finished AND duration>1000000 AND NOT cmd_name=IDLE
|
|
||||||
group_by = cmd_name
|
|
||||||
}
|
|
||||||
|
|
||||||
# IDLE-specific Metrics
|
|
||||||
|
|
||||||
metric imap_idle {
|
|
||||||
filter = event=imap_command_finished AND cmd_name=IDLE
|
|
||||||
fields = bytes_in bytes_out running_usecs
|
|
||||||
group_by = tagged_reply_state
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_idle_duration {
|
|
||||||
filter = event=imap_command_finished AND cmd_name=IDLE
|
|
||||||
# Base 10: 100ms to 27h (covers short wakeups to long idle sessions)
|
|
||||||
group_by = duration:exponential:5:11:10
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_idle_commands {
|
|
||||||
filter = event=imap_command_finished AND cmd_name=IDLE
|
|
||||||
group_by = tagged_reply_state
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_idle_failed {
|
|
||||||
filter = event=imap_command_finished AND cmd_name=IDLE AND NOT tagged_reply_state=OK
|
|
||||||
}
|
|
||||||
|
|
||||||
# Hibernation Metrics (requires imap_hibernate_timeout)
|
|
||||||
|
|
||||||
metric imap_hibernated {
|
|
||||||
filter = event=imap_client_hibernated
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_hibernated_failed {
|
|
||||||
filter = event=imap_client_hibernated AND error=*
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_unhibernated {
|
|
||||||
filter = event=imap_client_unhibernated
|
|
||||||
fields = hibernation_usecs
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_unhibernated_reason {
|
|
||||||
filter = event=imap_client_unhibernated
|
|
||||||
group_by = reason
|
|
||||||
fields = hibernation_usecs
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_unhibernated_reason_sleep {
|
|
||||||
filter = event=imap_client_unhibernated
|
|
||||||
group_by = reason hibernation_usecs:exponential:4:8:10
|
|
||||||
}
|
|
||||||
|
|
||||||
metric imap_unhibernated_failed {
|
|
||||||
filter = event=imap_client_unhibernated AND error=*
|
|
||||||
}
|
|
||||||
|
|
||||||
# Hibernation duration buckets (how long clients stayed hibernated)
|
|
||||||
# Base 10: 100ms to 27h
|
|
||||||
metric imap_hibernation_duration {
|
|
||||||
filter = event=imap_client_unhibernated
|
|
||||||
group_by = reason duration:exponential:5:11:10
|
|
||||||
}
|
|
||||||
|
|
||||||
# Authentication / Login Metrics
|
|
||||||
|
|
||||||
metric auth_request {
|
|
||||||
filter = event=auth_request_finished
|
|
||||||
group_by = success
|
|
||||||
}
|
|
||||||
|
|
||||||
metric auth_request_duration {
|
|
||||||
filter = event=auth_request_finished
|
|
||||||
group_by = success duration:exponential:2:6:10
|
|
||||||
}
|
|
||||||
|
|
||||||
metric auth_failed {
|
|
||||||
filter = event=auth_request_finished AND success=no
|
|
||||||
}
|
|
||||||
|
|
||||||
# Passdb cache effectiveness
|
|
||||||
metric auth_passdb {
|
|
||||||
filter = event=auth_passdb_request_finished
|
|
||||||
group_by = result cache
|
|
||||||
}
|
|
||||||
|
|
||||||
# Master login (post-auth userdb lookup)
|
|
||||||
metric auth_master_login {
|
|
||||||
filter = event=auth_master_client_login_finished
|
|
||||||
}
|
|
||||||
|
|
||||||
metric auth_master_login_failed {
|
|
||||||
filter = event=auth_master_client_login_finished AND error=*
|
|
||||||
}
|
|
||||||
|
|
||||||
# Mail Delivery (LMTP) - affects IDLE wakeup latency
|
|
||||||
|
|
||||||
metric mail_delivery {
|
|
||||||
filter = event=mail_delivery_finished
|
|
||||||
}
|
|
||||||
|
|
||||||
metric mail_delivery_duration {
|
|
||||||
filter = event=mail_delivery_finished
|
|
||||||
group_by = duration:exponential:3:7:10
|
|
||||||
}
|
|
||||||
|
|
||||||
metric mail_delivery_failed {
|
|
||||||
filter = event=mail_delivery_finished AND error=*
|
|
||||||
}
|
|
||||||
|
|
||||||
# Connection Events
|
|
||||||
|
|
||||||
metric client_connected {
|
|
||||||
filter = event=client_connection_connected AND category="service:imap"
|
|
||||||
}
|
|
||||||
|
|
||||||
metric client_disconnected {
|
|
||||||
filter = event=client_connection_disconnected AND category="service:imap"
|
|
||||||
fields = bytes_in bytes_out
|
|
||||||
}
|
}
|
||||||
{% endif %}
|
{% endif %}
|
||||||
|
|||||||
14
cmdeploy/src/cmdeploy/dovecot/expunge.cron.j2
Normal file
14
cmdeploy/src/cmdeploy/dovecot/expunge.cron.j2
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
# delete already seen big mails after 7 days, in the INBOX
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/cur/*' -mtime +7 -size +200k -type f -delete
|
||||||
|
# delete all mails after {{ config.delete_mails_after }} days, in the Inbox
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/cur/*' -mtime +{{ config.delete_mails_after }} -type f -delete
|
||||||
|
# or in any IMAP subfolder
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/.*/cur/*' -mtime +{{ config.delete_mails_after }} -type f -delete
|
||||||
|
# even if they are unseen
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/new/*' -mtime +{{ config.delete_mails_after }} -type f -delete
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/.*/new/*' -mtime +{{ config.delete_mails_after }} -type f -delete
|
||||||
|
# or only temporary (but then they shouldn't be around after {{ config.delete_mails_after }} days anyway).
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/tmp/*' -mtime +{{ config.delete_mails_after }} -type f -delete
|
||||||
|
2 0 * * * vmail find {{ config.mailboxes_dir }} -path '*/.*/tmp/*' -mtime +{{ config.delete_mails_after }} -type f -delete
|
||||||
|
3 0 * * * vmail find {{ config.mailboxes_dir }} -name 'maildirsize' -type f -delete
|
||||||
|
4 0 * * * vmail /usr/local/lib/chatmaild/venv/bin/delete_inactive_users /usr/local/lib/chatmaild/chatmail.ini
|
||||||
@@ -2,6 +2,15 @@ function dovecot_lua_notify_begin_txn(user)
|
|||||||
return user
|
return user
|
||||||
end
|
end
|
||||||
|
|
||||||
|
function contains(v, needle)
|
||||||
|
for _, keyword in ipairs(v) do
|
||||||
|
if keyword == needle then
|
||||||
|
return true
|
||||||
|
end
|
||||||
|
end
|
||||||
|
return false
|
||||||
|
end
|
||||||
|
|
||||||
function dovecot_lua_notify_event_message_new(user, event)
|
function dovecot_lua_notify_event_message_new(user, event)
|
||||||
local mbox = user:mailbox(event.mailbox)
|
local mbox = user:mailbox(event.mailbox)
|
||||||
mbox:sync()
|
mbox:sync()
|
||||||
|
|||||||
44
cmdeploy/src/cmdeploy/external/deployer.py
vendored
44
cmdeploy/src/cmdeploy/external/deployer.py
vendored
@@ -1,44 +0,0 @@
|
|||||||
|
|
||||||
from pyinfra import host
|
|
||||||
from pyinfra.facts.files import File
|
|
||||||
|
|
||||||
from ..basedeploy import Deployer
|
|
||||||
|
|
||||||
|
|
||||||
class ExternalTlsDeployer(Deployer):
|
|
||||||
"""Expects TLS certificates to be managed on the server.
|
|
||||||
|
|
||||||
Validates that the configured certificate and key files
|
|
||||||
exist on the remote host. Installs a systemd path unit
|
|
||||||
that watches the certificate file and automatically
|
|
||||||
restarts/reloads affected services when it changes.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, cert_path, key_path):
|
|
||||||
self.cert_path = cert_path
|
|
||||||
self.key_path = key_path
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
# Verify cert and key exist on the remote host using pyinfra facts.
|
|
||||||
for path in (self.cert_path, self.key_path):
|
|
||||||
if host.get_fact(File, path=path) is None:
|
|
||||||
raise Exception(f"External TLS file not found on server: {path}")
|
|
||||||
|
|
||||||
self.ensure_systemd_unit(
|
|
||||||
"external/tls-cert-reload.path.j2",
|
|
||||||
cert_path=self.cert_path,
|
|
||||||
)
|
|
||||||
self.ensure_systemd_unit(
|
|
||||||
"external/tls-cert-reload.service",
|
|
||||||
)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
# No explicit reload needed here: dovecot/nginx read the cert
|
|
||||||
# on startup, and the .path watcher handles live changes.
|
|
||||||
self.ensure_service(
|
|
||||||
"tls-cert-reload.path",
|
|
||||||
running=True,
|
|
||||||
enabled=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@@ -1,15 +0,0 @@
|
|||||||
# Watch the TLS certificate file for changes.
|
|
||||||
# When the cert is updated (e.g. renewed by an external process),
|
|
||||||
# this triggers tls-cert-reload.service to reload the affected services.
|
|
||||||
#
|
|
||||||
# NOTE: changes to the certificates are not detected if they cross bind-mount boundaries.
|
|
||||||
# After cert renewal, you must then trigger the reload explicitly:
|
|
||||||
# systemctl start tls-cert-reload.service
|
|
||||||
[Unit]
|
|
||||||
Description=Watch TLS certificate for changes
|
|
||||||
|
|
||||||
[Path]
|
|
||||||
PathChanged={{ cert_path }}
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
@@ -1,15 +0,0 @@
|
|||||||
# Reload services that cache the TLS certificate.
|
|
||||||
#
|
|
||||||
# dovecot: caches the cert at startup; reload re-reads SSL certs
|
|
||||||
# without dropping existing connections.
|
|
||||||
# nginx: caches the cert at startup; reload gracefully picks up
|
|
||||||
# the new cert for new connections.
|
|
||||||
# postfix: reads the cert fresh on each TLS handshake,
|
|
||||||
# does NOT need a reload/restart.
|
|
||||||
[Unit]
|
|
||||||
Description=Reload TLS services after certificate change
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=oneshot
|
|
||||||
ExecStart=/bin/systemctl try-reload-or-restart dovecot
|
|
||||||
ExecStart=/bin/systemctl try-reload-or-restart nginx
|
|
||||||
@@ -1,40 +0,0 @@
|
|||||||
import os
|
|
||||||
|
|
||||||
from pyinfra import facts, host
|
|
||||||
|
|
||||||
from cmdeploy.basedeploy import Deployer
|
|
||||||
|
|
||||||
|
|
||||||
class FiltermailDeployer(Deployer):
|
|
||||||
services = ["filtermail", "filtermail-incoming", "filtermail-transport"]
|
|
||||||
bin_path = "/usr/local/bin/filtermail"
|
|
||||||
config_path = "/usr/local/lib/chatmaild/chatmail.ini"
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
local_bin = os.environ.get("CHATMAIL_FILTERMAIL_BINARY")
|
|
||||||
if local_bin:
|
|
||||||
self.put_executable(
|
|
||||||
src=local_bin,
|
|
||||||
dest=self.bin_path,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
|
|
||||||
arch = host.get_fact(facts.server.Arch)
|
|
||||||
url = f"https://github.com/chatmail/filtermail/releases/download/v0.6.4/filtermail-{arch}"
|
|
||||||
sha256sum = {
|
|
||||||
"x86_64": "5295115952c72e4c4ec3c85546e094b4155a4c702c82bd71fcdcb744dc73adf6",
|
|
||||||
"aarch64": "6892244f17b8f26ccb465766e96028e7222b3c8adefca9fc6bfe9ff332ca8dff",
|
|
||||||
}[arch]
|
|
||||||
self.download_executable(url, self.bin_path, sha256sum)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
for service in self.services:
|
|
||||||
self.ensure_systemd_unit(
|
|
||||||
f"filtermail/{service}.service.j2",
|
|
||||||
bin_path=self.bin_path,
|
|
||||||
config_path=self.config_path,
|
|
||||||
)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
for service in self.services:
|
|
||||||
self.ensure_service(f"{service}.service")
|
|
||||||
@@ -1,11 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=Chatmail transport service
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
ExecStart={{ bin_path }} {{ config_path }} transport
|
|
||||||
Restart=always
|
|
||||||
RestartSec=30
|
|
||||||
User=vmail
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
@@ -1,11 +1,5 @@
|
|||||||
enable_relay = true
|
enable_relay = true
|
||||||
http_bind_addr = "[::]:3340"
|
http_bind_addr = "[::]:3340"
|
||||||
|
enable_stun = true
|
||||||
# Disable built-in STUN server in iroh-relay 0.35
|
|
||||||
# as we deploy our own TURN server instead.
|
|
||||||
# STUN server is going to be removed in iroh-relay 1.0
|
|
||||||
# and this line can be removed after upgrade.
|
|
||||||
enable_stun = false
|
|
||||||
|
|
||||||
enable_metrics = false
|
enable_metrics = false
|
||||||
metrics_bind_addr = "127.0.0.1:9092"
|
metrics_bind_addr = "127.0.0.1:9092"
|
||||||
|
|||||||
1
cmdeploy/src/cmdeploy/metrics.cron.j2
Normal file
1
cmdeploy/src/cmdeploy/metrics.cron.j2
Normal file
@@ -0,0 +1 @@
|
|||||||
|
*/5 * * * * root {{ config.execpath }} {{ config.mailboxes_dir }} >/var/www/html/metrics
|
||||||
@@ -44,45 +44,21 @@ counter warning_count
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
counter filtered_outgoing_mail_count
|
counter filtered_mail_count
|
||||||
|
|
||||||
counter outgoing_encrypted_mail_count
|
counter encrypted_mail_count
|
||||||
/Outgoing: Filtering encrypted mail\./ {
|
/Filtering encrypted mail\./ {
|
||||||
outgoing_encrypted_mail_count++
|
encrypted_mail_count++
|
||||||
filtered_outgoing_mail_count++
|
filtered_mail_count++
|
||||||
}
|
}
|
||||||
|
|
||||||
counter outgoing_unencrypted_mail_count
|
counter unencrypted_mail_count
|
||||||
/Outgoing: Filtering unencrypted mail\./ {
|
/Filtering unencrypted mail\./ {
|
||||||
outgoing_unencrypted_mail_count++
|
unencrypted_mail_count++
|
||||||
filtered_outgoing_mail_count++
|
filtered_mail_count++
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
counter filtered_incoming_mail_count
|
|
||||||
|
|
||||||
counter incoming_encrypted_mail_count
|
|
||||||
/Incoming: Filtering encrypted mail\./ {
|
|
||||||
incoming_encrypted_mail_count++
|
|
||||||
filtered_incoming_mail_count++
|
|
||||||
}
|
|
||||||
|
|
||||||
counter incoming_unencrypted_mail_count
|
|
||||||
/Incoming: Filtering unencrypted mail\./ {
|
|
||||||
incoming_unencrypted_mail_count++
|
|
||||||
filtered_incoming_mail_count++
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
counter rejected_unencrypted_mail_count
|
counter rejected_unencrypted_mail_count
|
||||||
/Rejected unencrypted mail/ {
|
/Rejected unencrypted mail\./ {
|
||||||
rejected_unencrypted_mail_count++
|
rejected_unencrypted_mail_count++
|
||||||
}
|
}
|
||||||
|
|
||||||
counter quota_expire_runs
|
|
||||||
counter quota_expire_removed_files
|
|
||||||
|
|
||||||
/quota-expire: removed (?P<count>\d+) message\(s\)/ {
|
|
||||||
quota_expire_runs++
|
|
||||||
quota_expire_removed_files += $count
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,44 +0,0 @@
|
|||||||
from pyinfra import facts, host
|
|
||||||
from pyinfra.operations import apt
|
|
||||||
|
|
||||||
from cmdeploy.basedeploy import Deployer
|
|
||||||
|
|
||||||
|
|
||||||
class MtailDeployer(Deployer):
|
|
||||||
def __init__(self, mtail_address):
|
|
||||||
self.mtail_address = mtail_address
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
# Uninstall mtail package to install a static binary.
|
|
||||||
apt.packages(name="Uninstall mtail", packages=["mtail"], present=False)
|
|
||||||
|
|
||||||
(url, sha256sum) = {
|
|
||||||
"x86_64": (
|
|
||||||
"https://github.com/google/mtail/releases/download/v3.0.8/mtail_3.0.8_linux_amd64.tar.gz",
|
|
||||||
"d55cb601049c5e61eabab29998dbbcea95d480e5448544f9470337ba2eea882e",
|
|
||||||
),
|
|
||||||
"aarch64": (
|
|
||||||
"https://github.com/google/mtail/releases/download/v3.0.8/mtail_3.0.8_linux_arm64.tar.gz",
|
|
||||||
"f748db8ad2a1e0b63684d4c8868cf6a373a20f7e6922e5ece601fff0ee00eb1a",
|
|
||||||
),
|
|
||||||
}[host.get_fact(facts.server.Arch)]
|
|
||||||
self.download_executable(
|
|
||||||
url,
|
|
||||||
"/usr/local/bin/mtail",
|
|
||||||
sha256sum,
|
|
||||||
extract="gunzip | tar -xf - mtail -O",
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
# Using our own systemd unit instead of `/usr/lib/systemd/system/mtail.service`.
|
|
||||||
# This allows to read from journalctl instead of log files.
|
|
||||||
self.ensure_systemd_unit(
|
|
||||||
"mtail/mtail.service.j2",
|
|
||||||
address=self.mtail_address or "127.0.0.1",
|
|
||||||
port=3903,
|
|
||||||
)
|
|
||||||
self.put_file("mtail/delivered_mail.mtail", "/etc/mtail/delivered_mail.mtail")
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
active = bool(self.mtail_address)
|
|
||||||
self.ensure_service("mtail.service", running=active, enabled=active)
|
|
||||||
@@ -1,13 +1,10 @@
|
|||||||
[Unit]
|
[Unit]
|
||||||
Description=mtail
|
Description=mtail
|
||||||
After=network-online.target
|
|
||||||
Wants=network-online.target
|
|
||||||
|
|
||||||
[Service]
|
[Service]
|
||||||
Type=simple
|
Type=simple
|
||||||
ExecStart=/bin/sh -c "journalctl -f -o short-iso -n 0 | /usr/local/bin/mtail --address={{ address }} --port={{ port }} --progs /etc/mtail --logtostderr --logs -"
|
ExecStart=/bin/sh -c "journalctl -f -o short-iso -n 0 | /usr/bin/mtail --address={{ address }} --port={{ port }} --progs /etc/mtail --logtostderr --logs /dev/stdin"
|
||||||
Restart=on-failure
|
Restart=on-failure
|
||||||
RestartSec=2s
|
|
||||||
|
|
||||||
[Install]
|
[Install]
|
||||||
WantedBy=multi-user.target
|
WantedBy=multi-user.target
|
||||||
|
|||||||
@@ -1,47 +1,47 @@
|
|||||||
<?xml version="1.0" encoding="UTF-8"?>
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
|
||||||
<clientConfig version="1.1">
|
<clientConfig version="1.1">
|
||||||
<emailProvider id="{{ config.mail_domain }}">
|
<emailProvider id="{{ config.domain_name }}">
|
||||||
<domain>{{ config.mail_domain }}</domain>
|
<domain>{{ config.domain_name }}</domain>
|
||||||
<displayName>{{ config.mail_domain }} chatmail</displayName>
|
<displayName>{{ config.domain_name }} chatmail</displayName>
|
||||||
<displayShortName>{{ config.mail_domain }}</displayShortName>
|
<displayShortName>{{ config.domain_name }}</displayShortName>
|
||||||
<incomingServer type="imap">
|
<incomingServer type="imap">
|
||||||
<hostname>{{ config.mail_domain }}</hostname>
|
<hostname>{{ config.domain_name }}</hostname>
|
||||||
<port>993</port>
|
<port>993</port>
|
||||||
<socketType>SSL</socketType>
|
<socketType>SSL</socketType>
|
||||||
<authentication>password-cleartext</authentication>
|
<authentication>password-cleartext</authentication>
|
||||||
<username>%EMAILADDRESS%</username>
|
<username>%EMAILADDRESS%</username>
|
||||||
</incomingServer>
|
</incomingServer>
|
||||||
<incomingServer type="imap">
|
<incomingServer type="imap">
|
||||||
<hostname>{{ config.mail_domain }}</hostname>
|
<hostname>{{ config.domain_name }}</hostname>
|
||||||
<port>143</port>
|
<port>143</port>
|
||||||
<socketType>STARTTLS</socketType>
|
<socketType>STARTTLS</socketType>
|
||||||
<authentication>password-cleartext</authentication>
|
<authentication>password-cleartext</authentication>
|
||||||
<username>%EMAILADDRESS%</username>
|
<username>%EMAILADDRESS%</username>
|
||||||
</incomingServer>
|
</incomingServer>
|
||||||
<incomingServer type="imap">
|
<incomingServer type="imap">
|
||||||
<hostname>{{ config.mail_domain }}</hostname>
|
<hostname>{{ config.domain_name }}</hostname>
|
||||||
<port>443</port>
|
<port>443</port>
|
||||||
<socketType>SSL</socketType>
|
<socketType>SSL</socketType>
|
||||||
<authentication>password-cleartext</authentication>
|
<authentication>password-cleartext</authentication>
|
||||||
<username>%EMAILADDRESS%</username>
|
<username>%EMAILADDRESS%</username>
|
||||||
</incomingServer>
|
</incomingServer>
|
||||||
<outgoingServer type="smtp">
|
<outgoingServer type="smtp">
|
||||||
<hostname>{{ config.mail_domain }}</hostname>
|
<hostname>{{ config.domain_name }}</hostname>
|
||||||
<port>465</port>
|
<port>465</port>
|
||||||
<socketType>SSL</socketType>
|
<socketType>SSL</socketType>
|
||||||
<authentication>password-cleartext</authentication>
|
<authentication>password-cleartext</authentication>
|
||||||
<username>%EMAILADDRESS%</username>
|
<username>%EMAILADDRESS%</username>
|
||||||
</outgoingServer>
|
</outgoingServer>
|
||||||
<outgoingServer type="smtp">
|
<outgoingServer type="smtp">
|
||||||
<hostname>{{ config.mail_domain }}</hostname>
|
<hostname>{{ config.domain_name }}</hostname>
|
||||||
<port>587</port>
|
<port>587</port>
|
||||||
<socketType>STARTTLS</socketType>
|
<socketType>STARTTLS</socketType>
|
||||||
<authentication>password-cleartext</authentication>
|
<authentication>password-cleartext</authentication>
|
||||||
<username>%EMAILADDRESS%</username>
|
<username>%EMAILADDRESS%</username>
|
||||||
</outgoingServer>
|
</outgoingServer>
|
||||||
<outgoingServer type="smtp">
|
<outgoingServer type="smtp">
|
||||||
<hostname>{{ config.mail_domain }}</hostname>
|
<hostname>{{ config.domain_name }}</hostname>
|
||||||
<port>443</port>
|
<port>443</port>
|
||||||
<socketType>SSL</socketType>
|
<socketType>SSL</socketType>
|
||||||
<authentication>password-cleartext</authentication>
|
<authentication>password-cleartext</authentication>
|
||||||
|
|||||||
@@ -1,80 +0,0 @@
|
|||||||
from chatmaild.config import Config
|
|
||||||
from pyinfra.operations import apt
|
|
||||||
|
|
||||||
from cmdeploy.basedeploy import (
|
|
||||||
Deployer,
|
|
||||||
get_resource,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class NginxDeployer(Deployer):
|
|
||||||
def __init__(self, config):
|
|
||||||
self.config = config
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
#
|
|
||||||
# If we allow nginx to start up on install, it will grab port
|
|
||||||
# 80, which then will block acmetool from listening on the port.
|
|
||||||
# That in turn prevents getting certificates, which then causes
|
|
||||||
# an error when we try to start nginx on the custom config
|
|
||||||
# that leaves port 80 open but also requires certificates to
|
|
||||||
# be present. To avoid getting into that interlocking mess,
|
|
||||||
# we use policy-rc.d to prevent nginx from starting up when it
|
|
||||||
# is installed.
|
|
||||||
#
|
|
||||||
# This approach allows us to avoid performing any explicit
|
|
||||||
# systemd operations during the install stage (as opposed to
|
|
||||||
# allowing it to start and then forcing it to stop), which allows
|
|
||||||
# the install stage to run in non-systemd environments like a
|
|
||||||
# container image build.
|
|
||||||
#
|
|
||||||
# For documentation about policy-rc.d, see:
|
|
||||||
# https://people.debian.org/~hmh/invokerc.d-policyrc.d-specification.txt
|
|
||||||
#
|
|
||||||
self.put_executable(src="policy-rc.d", dest="/usr/sbin/policy-rc.d")
|
|
||||||
|
|
||||||
apt.packages(
|
|
||||||
name="Install nginx",
|
|
||||||
packages=["nginx", "libnginx-mod-stream"],
|
|
||||||
)
|
|
||||||
|
|
||||||
self.remove_file("/usr/sbin/policy-rc.d")
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
_configure_nginx(self, self.config)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service("nginx.service")
|
|
||||||
|
|
||||||
|
|
||||||
def _configure_nginx(deployer, config: Config, debug: bool = False):
|
|
||||||
"""Configures nginx HTTP server."""
|
|
||||||
|
|
||||||
deployer.put_template(
|
|
||||||
"nginx/nginx.conf.j2",
|
|
||||||
"/etc/nginx/nginx.conf",
|
|
||||||
config=config,
|
|
||||||
disable_ipv6=config.disable_ipv6,
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.put_template(
|
|
||||||
"nginx/autoconfig.xml.j2",
|
|
||||||
"/var/www/html/.well-known/autoconfig/mail/config-v1.1.xml",
|
|
||||||
config=config,
|
|
||||||
)
|
|
||||||
|
|
||||||
deployer.put_template(
|
|
||||||
"nginx/mta-sts.txt.j2",
|
|
||||||
"/var/www/html/.well-known/mta-sts.txt",
|
|
||||||
config=config,
|
|
||||||
)
|
|
||||||
|
|
||||||
# install CGI newemail script
|
|
||||||
#
|
|
||||||
cgi_dir = "/usr/lib/cgi-bin"
|
|
||||||
deployer.ensure_directory(cgi_dir)
|
|
||||||
|
|
||||||
deployer.put_executable(
|
|
||||||
src=get_resource("newemail.py", pkg="chatmaild").open("rb"),
|
|
||||||
dest=f"{cgi_dir}/newemail.py",
|
|
||||||
)
|
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
version: STSv1
|
version: STSv1
|
||||||
mode: enforce
|
mode: enforce
|
||||||
mx: {{ config.mail_domain }}
|
mx: {{ config.domain_name }}
|
||||||
max_age: 2419200
|
max_age: 2419200
|
||||||
|
|||||||
@@ -2,25 +2,11 @@ load_module modules/ngx_stream_module.so;
|
|||||||
|
|
||||||
user www-data;
|
user www-data;
|
||||||
worker_processes auto;
|
worker_processes auto;
|
||||||
|
|
||||||
# Increase the number of connections
|
|
||||||
# that a worker process can open
|
|
||||||
# to avoid errors such as
|
|
||||||
# accept4() failed (24: Too many open files)
|
|
||||||
# and
|
|
||||||
# socket() failed (24: Too many open files) while connecting to upstream
|
|
||||||
# in the logs.
|
|
||||||
# <https://nginx.org/en/docs/ngx_core_module.html#worker_rlimit_nofile>
|
|
||||||
worker_rlimit_nofile 2048;
|
|
||||||
pid /run/nginx.pid;
|
pid /run/nginx.pid;
|
||||||
error_log syslog:server=unix:/dev/log,facility=local3;
|
error_log syslog:server=unix:/dev/log,facility=local3;
|
||||||
|
|
||||||
events {
|
events {
|
||||||
# Increase to avoid errors such as
|
worker_connections 768;
|
||||||
# 768 worker_connections are not enough while connecting to upstream
|
|
||||||
# in the logs.
|
|
||||||
# <https://nginx.org/en/docs/ngx_core_module.html#worker_connections>
|
|
||||||
worker_connections 2048;
|
|
||||||
# multi_accept on;
|
# multi_accept on;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -42,9 +28,6 @@ stream {
|
|||||||
}
|
}
|
||||||
|
|
||||||
http {
|
http {
|
||||||
{% if config.tls_cert_mode == "self" %}
|
|
||||||
limit_req_zone $binary_remote_addr zone=newaccount:10m rate=2r/s;
|
|
||||||
{% endif %}
|
|
||||||
sendfile on;
|
sendfile on;
|
||||||
tcp_nopush on;
|
tcp_nopush on;
|
||||||
|
|
||||||
@@ -54,10 +37,10 @@ http {
|
|||||||
include /etc/nginx/mime.types;
|
include /etc/nginx/mime.types;
|
||||||
default_type application/octet-stream;
|
default_type application/octet-stream;
|
||||||
|
|
||||||
ssl_protocols TLSv1.2 TLSv1.3;
|
ssl_protocols TLSv1 TLSv1.1 TLSv1.2 TLSv1.3;
|
||||||
ssl_prefer_server_ciphers on;
|
ssl_prefer_server_ciphers on;
|
||||||
ssl_certificate {{ config.tls_cert_path }};
|
ssl_certificate /var/lib/acme/live/{{ config.domain_name }}/fullchain;
|
||||||
ssl_certificate_key {{ config.tls_key_path }};
|
ssl_certificate_key /var/lib/acme/live/{{ config.domain_name }}/privkey;
|
||||||
|
|
||||||
gzip on;
|
gzip on;
|
||||||
|
|
||||||
@@ -69,30 +52,26 @@ http {
|
|||||||
|
|
||||||
index index.html index.htm;
|
index index.html index.htm;
|
||||||
|
|
||||||
server_name {{ config.mail_domain }} mta-sts.{{ config.mail_domain }};
|
server_name _;
|
||||||
|
|
||||||
access_log syslog:server=unix:/dev/log,facility=local7;
|
access_log syslog:server=unix:/dev/log,facility=local7;
|
||||||
|
|
||||||
location /mxdeliv {
|
|
||||||
proxy_pass http://127.0.0.1:{{ config.filtermail_http_port_incoming }};
|
|
||||||
}
|
|
||||||
|
|
||||||
location / {
|
location / {
|
||||||
# First attempt to serve request as file, then
|
# First attempt to serve request as file, then
|
||||||
# as directory, then fall back to displaying a 404.
|
# as directory, then fall back to displaying a 404.
|
||||||
try_files $uri $uri/ =404;
|
try_files $uri $uri/ =404;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
location /metrics {
|
||||||
|
default_type text/plain;
|
||||||
|
}
|
||||||
|
|
||||||
location /new {
|
location /new {
|
||||||
{% if config.tls_cert_mode != "self" %}
|
|
||||||
if ($request_method = GET) {
|
if ($request_method = GET) {
|
||||||
# Redirect to Delta Chat,
|
# Redirect to Delta Chat,
|
||||||
# which will in turn do a POST request.
|
# which will in turn do a POST request.
|
||||||
return 301 dcaccount:https://{{ config.mail_domain }}/new;
|
return 301 dcaccount:https://{{ config.domain_name }}/new;
|
||||||
}
|
}
|
||||||
{% else %}
|
|
||||||
limit_req zone=newaccount burst=5 nodelay;
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
fastcgi_pass unix:/run/fcgiwrap.socket;
|
fastcgi_pass unix:/run/fcgiwrap.socket;
|
||||||
include /etc/nginx/fastcgi_params;
|
include /etc/nginx/fastcgi_params;
|
||||||
@@ -106,11 +85,9 @@ http {
|
|||||||
#
|
#
|
||||||
# Redirects are only for browsers.
|
# Redirects are only for browsers.
|
||||||
location /cgi-bin/newemail.py {
|
location /cgi-bin/newemail.py {
|
||||||
{% if config.tls_cert_mode != "self" %}
|
|
||||||
if ($request_method = GET) {
|
if ($request_method = GET) {
|
||||||
return 301 dcaccount:https://{{ config.mail_domain }}/new;
|
return 301 dcaccount:https://{{ config.domain_name }}/new;
|
||||||
}
|
}
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
fastcgi_pass unix:/run/fcgiwrap.socket;
|
fastcgi_pass unix:/run/fcgiwrap.socket;
|
||||||
include /etc/nginx/fastcgi_params;
|
include /etc/nginx/fastcgi_params;
|
||||||
@@ -141,29 +118,8 @@ http {
|
|||||||
# Redirect www. to non-www
|
# Redirect www. to non-www
|
||||||
server {
|
server {
|
||||||
listen 127.0.0.1:8443 ssl;
|
listen 127.0.0.1:8443 ssl;
|
||||||
server_name www.{{ config.mail_domain }};
|
server_name www.{{ config.domain_name }};
|
||||||
return 301 $scheme://{{ config.mail_domain }}$request_uri;
|
return 301 $scheme://{{ config.domain_name }}$request_uri;
|
||||||
access_log syslog:server=unix:/dev/log,facility=local7;
|
access_log syslog:server=unix:/dev/log,facility=local7;
|
||||||
}
|
}
|
||||||
|
|
||||||
server {
|
|
||||||
listen 80;
|
|
||||||
{% if not disable_ipv6 %}
|
|
||||||
listen [::]:80;
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if config.tls_cert_mode == "acme" %}
|
|
||||||
location /.well-known/acme-challenge/ {
|
|
||||||
proxy_pass http://acmetool;
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
return 301 https://$host$request_uri;
|
|
||||||
}
|
|
||||||
|
|
||||||
{% if config.tls_cert_mode == "acme" %}
|
|
||||||
upstream acmetool {
|
|
||||||
server 127.0.0.1:402;
|
|
||||||
}
|
|
||||||
{% endif %}
|
|
||||||
}
|
}
|
||||||
|
|||||||
BIN
cmdeploy/src/cmdeploy/obs-home-deltachat.gpg
Normal file
BIN
cmdeploy/src/cmdeploy/obs-home-deltachat.gpg
Normal file
Binary file not shown.
@@ -1,86 +0,0 @@
|
|||||||
"""
|
|
||||||
Installs OpenDKIM
|
|
||||||
"""
|
|
||||||
|
|
||||||
from pyinfra import host
|
|
||||||
from pyinfra.facts.files import File
|
|
||||||
from pyinfra.operations import apt, files, server
|
|
||||||
|
|
||||||
from cmdeploy.basedeploy import Deployer
|
|
||||||
|
|
||||||
|
|
||||||
class OpendkimDeployer(Deployer):
|
|
||||||
required_users = [("opendkim", None, ["opendkim"])]
|
|
||||||
|
|
||||||
def __init__(self, mail_domain):
|
|
||||||
self.mail_domain = mail_domain
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
apt.packages(
|
|
||||||
name="apt install opendkim opendkim-tools",
|
|
||||||
packages=["opendkim", "opendkim-tools"],
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
domain = self.mail_domain
|
|
||||||
dkim_selector = "opendkim"
|
|
||||||
"""Configures OpenDKIM"""
|
|
||||||
|
|
||||||
self.put_template(
|
|
||||||
"opendkim/opendkim.conf",
|
|
||||||
"/etc/opendkim.conf",
|
|
||||||
config={"domain_name": domain, "opendkim_selector": dkim_selector},
|
|
||||||
)
|
|
||||||
|
|
||||||
self.remove_file("/etc/opendkim/screen.lua")
|
|
||||||
self.remove_file("/etc/opendkim/final.lua")
|
|
||||||
|
|
||||||
self.ensure_directory(
|
|
||||||
"/etc/opendkim",
|
|
||||||
owner="opendkim",
|
|
||||||
mode="750",
|
|
||||||
)
|
|
||||||
|
|
||||||
self.put_template(
|
|
||||||
"opendkim/KeyTable",
|
|
||||||
"/etc/dkimkeys/KeyTable",
|
|
||||||
owner="opendkim",
|
|
||||||
config={"domain_name": domain, "opendkim_selector": dkim_selector},
|
|
||||||
)
|
|
||||||
|
|
||||||
self.put_template(
|
|
||||||
"opendkim/SigningTable",
|
|
||||||
"/etc/dkimkeys/SigningTable",
|
|
||||||
owner="opendkim",
|
|
||||||
config={"domain_name": domain, "opendkim_selector": dkim_selector},
|
|
||||||
)
|
|
||||||
self.ensure_directory(
|
|
||||||
"/var/spool/postfix/opendkim",
|
|
||||||
owner="opendkim",
|
|
||||||
mode="750",
|
|
||||||
)
|
|
||||||
|
|
||||||
if not host.get_fact(File, f"/etc/dkimkeys/{dkim_selector}.private"):
|
|
||||||
server.shell(
|
|
||||||
name="Generate OpenDKIM domain keys",
|
|
||||||
commands=[
|
|
||||||
f"/usr/sbin/opendkim-genkey -D /etc/dkimkeys -d {domain} -s {dkim_selector}"
|
|
||||||
],
|
|
||||||
_use_su_login=True,
|
|
||||||
_su_user="opendkim",
|
|
||||||
)
|
|
||||||
|
|
||||||
self.put_file(
|
|
||||||
"opendkim/systemd.conf",
|
|
||||||
"/etc/systemd/system/opendkim.service.d/10-prevent-memory-leak.conf",
|
|
||||||
)
|
|
||||||
|
|
||||||
files.file(
|
|
||||||
name="chown opendkim: /etc/dkimkeys/opendkim.private",
|
|
||||||
path="/etc/dkimkeys/opendkim.private",
|
|
||||||
user="opendkim",
|
|
||||||
group="opendkim",
|
|
||||||
)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
self.ensure_service("opendkim.service")
|
|
||||||
28
cmdeploy/src/cmdeploy/opendkim/final.lua
Normal file
28
cmdeploy/src/cmdeploy/opendkim/final.lua
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
if odkim.internal_ip(ctx) == 1 then
|
||||||
|
-- Outgoing message will be signed,
|
||||||
|
-- no need to look for signatures.
|
||||||
|
return nil
|
||||||
|
end
|
||||||
|
|
||||||
|
nsigs = odkim.get_sigcount(ctx)
|
||||||
|
if nsigs == nil then
|
||||||
|
return nil
|
||||||
|
end
|
||||||
|
|
||||||
|
for i = 1, nsigs do
|
||||||
|
sig = odkim.get_sighandle(ctx, i - 1)
|
||||||
|
sigres = odkim.sig_result(sig)
|
||||||
|
|
||||||
|
-- All signatures that do not correspond to From:
|
||||||
|
-- were ignored in screen.lua and return sigres -1.
|
||||||
|
--
|
||||||
|
-- Any valid signature that was not ignored like this
|
||||||
|
-- means the message is acceptable.
|
||||||
|
if sigres == 0 then
|
||||||
|
return nil
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
odkim.set_reply(ctx, "554", "5.7.1", "No valid DKIM signature found")
|
||||||
|
odkim.set_result(ctx, SMFIS_REJECT)
|
||||||
|
return nil
|
||||||
@@ -13,7 +13,6 @@ OversignHeaders From
|
|||||||
On-BadSignature reject
|
On-BadSignature reject
|
||||||
On-KeyNotFound reject
|
On-KeyNotFound reject
|
||||||
On-NoSignature reject
|
On-NoSignature reject
|
||||||
DNSTimeout 60
|
|
||||||
|
|
||||||
# Signing domain, selector, and key (required). For example, perform signing
|
# Signing domain, selector, and key (required). For example, perform signing
|
||||||
# for domain "example.com" with selector "2020" (2020._domainkey.example.com),
|
# for domain "example.com" with selector "2020" (2020._domainkey.example.com),
|
||||||
@@ -45,6 +44,12 @@ SignHeaders *,+autocrypt,+content-type
|
|||||||
# Default is empty.
|
# Default is empty.
|
||||||
OversignHeaders from,reply-to,subject,date,to,cc,resent-date,resent-from,resent-sender,resent-to,resent-cc,in-reply-to,references,list-id,list-help,list-unsubscribe,list-subscribe,list-post,list-owner,list-archive,autocrypt
|
OversignHeaders from,reply-to,subject,date,to,cc,resent-date,resent-from,resent-sender,resent-to,resent-cc,in-reply-to,references,list-id,list-help,list-unsubscribe,list-subscribe,list-post,list-owner,list-archive,autocrypt
|
||||||
|
|
||||||
|
# Script to ignore signatures that do not correspond to the From: domain.
|
||||||
|
ScreenPolicyScript /etc/opendkim/screen.lua
|
||||||
|
|
||||||
|
# Script to reject mails without a valid DKIM signature.
|
||||||
|
FinalPolicyScript /etc/opendkim/final.lua
|
||||||
|
|
||||||
# In Debian, opendkim runs as user "opendkim". A umask of 007 is required when
|
# In Debian, opendkim runs as user "opendkim". A umask of 007 is required when
|
||||||
# using a local socket with MTAs that access the socket as a non-privileged
|
# using a local socket with MTAs that access the socket as a non-privileged
|
||||||
# user (for example, Postfix). You may need to add user "postfix" to group
|
# user (for example, Postfix). You may need to add user "postfix" to group
|
||||||
@@ -59,9 +64,3 @@ PidFile /run/opendkim/opendkim.pid
|
|||||||
# The trust anchor enables DNSSEC. In Debian, the trust anchor file is provided
|
# The trust anchor enables DNSSEC. In Debian, the trust anchor file is provided
|
||||||
# by the package dns-root-data.
|
# by the package dns-root-data.
|
||||||
TrustAnchorFile /usr/share/dns/root.key
|
TrustAnchorFile /usr/share/dns/root.key
|
||||||
|
|
||||||
# Sign messages when `-o milter_macro_daemon_name=ORIGINATING` is set.
|
|
||||||
MTA ORIGINATING
|
|
||||||
|
|
||||||
# No hosts are treated as internal, ORIGINATING daemon name should be set explicitly.
|
|
||||||
InternalHosts -
|
|
||||||
|
|||||||
21
cmdeploy/src/cmdeploy/opendkim/screen.lua
Normal file
21
cmdeploy/src/cmdeploy/opendkim/screen.lua
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
-- Ignore signatures that do not correspond to the From: domain.
|
||||||
|
|
||||||
|
from_domain = odkim.get_fromdomain(ctx)
|
||||||
|
if from_domain == nil then
|
||||||
|
return nil
|
||||||
|
end
|
||||||
|
|
||||||
|
n = odkim.get_sigcount(ctx)
|
||||||
|
if n == nil then
|
||||||
|
return nil
|
||||||
|
end
|
||||||
|
|
||||||
|
for i = 1, n do
|
||||||
|
sig = odkim.get_sighandle(ctx, i - 1)
|
||||||
|
sig_domain = odkim.sig_getdomain(sig)
|
||||||
|
if from_domain ~= sig_domain then
|
||||||
|
odkim.sig_ignore(sig)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
return nil
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
#!/bin/sh
|
|
||||||
echo "All runlevel operations denied by policy" >&2
|
|
||||||
exit 101
|
|
||||||
@@ -1,75 +0,0 @@
|
|||||||
from pyinfra.operations import apt, server
|
|
||||||
|
|
||||||
from cmdeploy.basedeploy import Deployer
|
|
||||||
|
|
||||||
|
|
||||||
class PostfixDeployer(Deployer):
|
|
||||||
required_users = [("postfix", None, ["opendkim"])]
|
|
||||||
|
|
||||||
def __init__(self, config, disable_mail):
|
|
||||||
self.config = config
|
|
||||||
self.disable_mail = disable_mail
|
|
||||||
|
|
||||||
def install(self):
|
|
||||||
apt.packages(
|
|
||||||
name="Install Postfix",
|
|
||||||
packages="postfix",
|
|
||||||
)
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
config = self.config
|
|
||||||
|
|
||||||
self.put_template(
|
|
||||||
"postfix/main.cf.j2",
|
|
||||||
"/etc/postfix/main.cf",
|
|
||||||
config=config,
|
|
||||||
disable_ipv6=config.disable_ipv6,
|
|
||||||
)
|
|
||||||
|
|
||||||
self.put_template(
|
|
||||||
"postfix/master.cf.j2",
|
|
||||||
"/etc/postfix/master.cf",
|
|
||||||
debug=False,
|
|
||||||
config=config,
|
|
||||||
)
|
|
||||||
|
|
||||||
self.put_file(
|
|
||||||
"postfix/submission_header_cleanup",
|
|
||||||
"/etc/postfix/submission_header_cleanup",
|
|
||||||
)
|
|
||||||
self.put_file("postfix/lmtp_header_cleanup", "/etc/postfix/lmtp_header_cleanup")
|
|
||||||
|
|
||||||
res = self.put_file(
|
|
||||||
"postfix/smtp_tls_policy_map", "/etc/postfix/smtp_tls_policy_map"
|
|
||||||
)
|
|
||||||
tls_policy_changed = res.changed
|
|
||||||
if tls_policy_changed:
|
|
||||||
server.shell(
|
|
||||||
commands=["postmap /etc/postfix/smtp_tls_policy_map"],
|
|
||||||
)
|
|
||||||
|
|
||||||
# Login map that 1:1 maps email address to login.
|
|
||||||
self.put_file("postfix/login_map", "/etc/postfix/login_map")
|
|
||||||
|
|
||||||
self.put_file(
|
|
||||||
"service/10_restart_on_failure.conf",
|
|
||||||
"/etc/systemd/system/postfix@.service.d/10_restart.conf",
|
|
||||||
)
|
|
||||||
|
|
||||||
# Validate postfix configuration before restart
|
|
||||||
if self.need_restart:
|
|
||||||
server.shell(
|
|
||||||
name="Validate postfix configuration",
|
|
||||||
# Extract stderr and quit with error if non-zero
|
|
||||||
commands=[
|
|
||||||
"""bash -c 'w=$(postconf 2>&1 >/dev/null); [[ -z "$w" ]] || { echo "$w"; false; }'"""
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
active = not self.disable_mail
|
|
||||||
self.ensure_service(
|
|
||||||
"postfix.service",
|
|
||||||
running=active,
|
|
||||||
enabled=active,
|
|
||||||
)
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
/^DKIM-Signature:/ IGNORE
|
|
||||||
/^Authentication-Results:/ IGNORE
|
|
||||||
/^Received:/ IGNORE
|
|
||||||
@@ -15,8 +15,8 @@ readme_directory = no
|
|||||||
compatibility_level = 3.6
|
compatibility_level = 3.6
|
||||||
|
|
||||||
# TLS parameters
|
# TLS parameters
|
||||||
smtpd_tls_cert_file={{ config.tls_cert_path }}
|
smtpd_tls_cert_file=/var/lib/acme/live/{{ config.mail_domain }}/fullchain
|
||||||
smtpd_tls_key_file={{ config.tls_key_path }}
|
smtpd_tls_key_file=/var/lib/acme/live/{{ config.mail_domain }}/privkey
|
||||||
smtpd_tls_security_level=may
|
smtpd_tls_security_level=may
|
||||||
|
|
||||||
smtp_tls_CApath=/etc/ssl/certs
|
smtp_tls_CApath=/etc/ssl/certs
|
||||||
@@ -25,9 +25,8 @@ smtp_tls_security_level=verify
|
|||||||
# <https://www.postfix.org/postconf.5.html#smtp_tls_servername>
|
# <https://www.postfix.org/postconf.5.html#smtp_tls_servername>
|
||||||
smtp_tls_servername = hostname
|
smtp_tls_servername = hostname
|
||||||
smtp_tls_session_cache_database = btree:${data_directory}/smtp_scache
|
smtp_tls_session_cache_database = btree:${data_directory}/smtp_scache
|
||||||
smtp_tls_policy_maps = regexp:/etc/postfix/smtp_tls_policy_map
|
smtp_tls_policy_maps = inline:{nauta.cu=may}
|
||||||
smtp_tls_protocols = >=TLSv1.2
|
smtpd_tls_protocols = >=TLSv1.2
|
||||||
smtp_tls_mandatory_protocols = >=TLSv1.2
|
|
||||||
|
|
||||||
# Disable anonymous cipher suites
|
# Disable anonymous cipher suites
|
||||||
# and known insecure algorithms.
|
# and known insecure algorithms.
|
||||||
@@ -64,11 +63,7 @@ alias_database = hash:/etc/aliases
|
|||||||
mydestination =
|
mydestination =
|
||||||
|
|
||||||
relayhost =
|
relayhost =
|
||||||
{% if disable_ipv6 %}
|
|
||||||
mynetworks = 127.0.0.0/8
|
|
||||||
{% else %}
|
|
||||||
mynetworks = 127.0.0.0/8 [::ffff:127.0.0.0]/104 [::1]/128
|
mynetworks = 127.0.0.0/8 [::ffff:127.0.0.0]/104 [::1]/128
|
||||||
{% endif %}
|
|
||||||
mailbox_size_limit = 0
|
mailbox_size_limit = 0
|
||||||
message_size_limit = {{config.max_message_size}}
|
message_size_limit = {{config.max_message_size}}
|
||||||
recipient_delimiter = +
|
recipient_delimiter = +
|
||||||
@@ -81,7 +76,6 @@ inet_protocols = all
|
|||||||
|
|
||||||
virtual_transport = lmtp:unix:private/dovecot-lmtp
|
virtual_transport = lmtp:unix:private/dovecot-lmtp
|
||||||
virtual_mailbox_domains = {{ config.mail_domain }}
|
virtual_mailbox_domains = {{ config.mail_domain }}
|
||||||
lmtp_header_checks = regexp:/etc/postfix/lmtp_header_cleanup
|
|
||||||
|
|
||||||
mua_client_restrictions = permit_sasl_authenticated, reject
|
mua_client_restrictions = permit_sasl_authenticated, reject
|
||||||
mua_sender_restrictions = reject_sender_login_mismatch, permit_sasl_authenticated, reject
|
mua_sender_restrictions = reject_sender_login_mismatch, permit_sasl_authenticated, reject
|
||||||
@@ -93,12 +87,3 @@ smtpd_sender_login_maps = regexp:/etc/postfix/login_map
|
|||||||
# Do not lookup SMTP client hostnames to reduce delays
|
# Do not lookup SMTP client hostnames to reduce delays
|
||||||
# and avoid unnecessary DNS requests.
|
# and avoid unnecessary DNS requests.
|
||||||
smtpd_peername_lookup = no
|
smtpd_peername_lookup = no
|
||||||
|
|
||||||
# Use filtermail-transport to relay messages.
|
|
||||||
# We can't force postfix to split messages per destination,
|
|
||||||
# when specifying a custom next-hop,
|
|
||||||
# so instead this is handled in filtermail.
|
|
||||||
# We use LMTP instead SMTP so we can communicate per-recipient errors back to postfix.
|
|
||||||
default_transport = lmtp-filtermail:inet:[127.0.0.1]:{{ config.filtermail_lmtp_port_transport }}
|
|
||||||
lmtp-filtermail_initial_destination_concurrency=10000
|
|
||||||
lmtp-filtermail_destination_concurrency_limit=10000
|
|
||||||
|
|||||||
@@ -14,8 +14,6 @@ smtp inet n - y - - smtpd -v
|
|||||||
{%- else %}
|
{%- else %}
|
||||||
smtp inet n - y - - smtpd
|
smtp inet n - y - - smtpd
|
||||||
{%- endif %}
|
{%- endif %}
|
||||||
-o smtpd_tls_security_level=encrypt
|
|
||||||
-o smtpd_tls_mandatory_protocols=>=TLSv1.2
|
|
||||||
-o smtpd_proxy_filter=127.0.0.1:{{ config.filtermail_smtp_port_incoming }}
|
-o smtpd_proxy_filter=127.0.0.1:{{ config.filtermail_smtp_port_incoming }}
|
||||||
submission inet n - y - 5000 smtpd
|
submission inet n - y - 5000 smtpd
|
||||||
-o syslog_name=postfix/submission
|
-o syslog_name=postfix/submission
|
||||||
@@ -31,6 +29,7 @@ submission inet n - y - 5000 smtpd
|
|||||||
-o smtpd_sender_restrictions=$mua_sender_restrictions
|
-o smtpd_sender_restrictions=$mua_sender_restrictions
|
||||||
-o smtpd_recipient_restrictions=
|
-o smtpd_recipient_restrictions=
|
||||||
-o smtpd_relay_restrictions=permit_sasl_authenticated,reject
|
-o smtpd_relay_restrictions=permit_sasl_authenticated,reject
|
||||||
|
-o milter_macro_daemon_name=ORIGINATING
|
||||||
-o smtpd_client_connection_count_limit=1000
|
-o smtpd_client_connection_count_limit=1000
|
||||||
-o smtpd_proxy_filter=127.0.0.1:{{ config.filtermail_smtp_port }}
|
-o smtpd_proxy_filter=127.0.0.1:{{ config.filtermail_smtp_port }}
|
||||||
smtps inet n - y - 5000 smtpd
|
smtps inet n - y - 5000 smtpd
|
||||||
@@ -48,6 +47,7 @@ smtps inet n - y - 5000 smtpd
|
|||||||
-o smtpd_recipient_restrictions=
|
-o smtpd_recipient_restrictions=
|
||||||
-o smtpd_relay_restrictions=permit_sasl_authenticated,reject
|
-o smtpd_relay_restrictions=permit_sasl_authenticated,reject
|
||||||
-o smtpd_client_connection_count_limit=1000
|
-o smtpd_client_connection_count_limit=1000
|
||||||
|
-o milter_macro_daemon_name=ORIGINATING
|
||||||
-o smtpd_proxy_filter=127.0.0.1:{{ config.filtermail_smtp_port }}
|
-o smtpd_proxy_filter=127.0.0.1:{{ config.filtermail_smtp_port }}
|
||||||
#628 inet n - y - - qmqpd
|
#628 inet n - y - - qmqpd
|
||||||
pickup unix n - y 60 1 pickup
|
pickup unix n - y 60 1 pickup
|
||||||
@@ -77,15 +77,15 @@ scache unix - - y - 1 scache
|
|||||||
postlog unix-dgram n - n - 1 postlogd
|
postlog unix-dgram n - n - 1 postlogd
|
||||||
filter unix - n n - - lmtp
|
filter unix - n n - - lmtp
|
||||||
# Local SMTP server for reinjecting outgoing filtered mail.
|
# Local SMTP server for reinjecting outgoing filtered mail.
|
||||||
127.0.0.1:{{ config.postfix_reinject_port }} inet n - n - 100 smtpd
|
127.0.0.1:{{ config.postfix_reinject_port }} inet n - n - 10 smtpd
|
||||||
-o syslog_name=postfix/reinject
|
-o syslog_name=postfix/reinject
|
||||||
-o milter_macro_daemon_name=ORIGINATING
|
|
||||||
-o smtpd_milters=unix:opendkim/opendkim.sock
|
-o smtpd_milters=unix:opendkim/opendkim.sock
|
||||||
-o cleanup_service_name=authclean
|
-o cleanup_service_name=authclean
|
||||||
|
|
||||||
# Local SMTP server for reinjecting incoming filtered mail
|
# Local SMTP server for reinjecting incoming filtered mail
|
||||||
127.0.0.1:{{ config.postfix_reinject_port_incoming }} inet n - n - 100 smtpd
|
127.0.0.1:{{ config.postfix_reinject_port_incoming }} inet n - n - 10 smtpd
|
||||||
-o syslog_name=postfix/reinject_incoming
|
-o syslog_name=postfix/reinject_incoming
|
||||||
|
-o smtpd_milters=unix:opendkim/opendkim.sock
|
||||||
|
|
||||||
# Cleanup `Received` headers for authenticated mail
|
# Cleanup `Received` headers for authenticated mail
|
||||||
# to avoid leaking client IP.
|
# to avoid leaking client IP.
|
||||||
@@ -100,8 +100,3 @@ filter unix - n n - - lmtp
|
|||||||
# cannot send unprotected Subject.
|
# cannot send unprotected Subject.
|
||||||
authclean unix n - - - 0 cleanup
|
authclean unix n - - - 0 cleanup
|
||||||
-o header_checks=regexp:/etc/postfix/submission_header_cleanup
|
-o header_checks=regexp:/etc/postfix/submission_header_cleanup
|
||||||
|
|
||||||
lmtp-filtermail unix - - y - 10000 lmtp
|
|
||||||
-o syslog_name=postfix/lmtp-filtermail
|
|
||||||
-o lmtp_header_checks=
|
|
||||||
-o lmtp_tls_security_level=none
|
|
||||||
|
|||||||
@@ -1,3 +0,0 @@
|
|||||||
/^\[[^]]+\]$/ encrypt
|
|
||||||
/^_/ encrypt
|
|
||||||
/^nauta\.cu$/ may
|
|
||||||
@@ -12,55 +12,48 @@ All functions of this module
|
|||||||
|
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from .rshell import CalledProcessError, log_progress, shell
|
from .rshell import CalledProcessError, shell
|
||||||
|
|
||||||
|
|
||||||
def perform_initial_checks(mail_domain, pre_command=""):
|
def perform_initial_checks(mail_domain):
|
||||||
"""Collecting initial DNS settings."""
|
"""Collecting initial DNS settings."""
|
||||||
assert mail_domain
|
assert mail_domain
|
||||||
if not shell("dig", fail_ok=True, print=log_progress):
|
if not shell("dig", fail_ok=True):
|
||||||
shell("apt-get update && apt-get install -y dnsutils", print=log_progress)
|
shell("apt-get install -y dnsutils")
|
||||||
A = query_dns("A", mail_domain)
|
A = query_dns("A", mail_domain)
|
||||||
AAAA = query_dns("AAAA", mail_domain)
|
AAAA = query_dns("AAAA", mail_domain)
|
||||||
MTA_STS = query_dns("CNAME", f"mta-sts.{mail_domain}")
|
MTA_STS = query_dns("CNAME", f"mta-sts.{mail_domain}")
|
||||||
WWW = query_dns("CNAME", f"www.{mail_domain}")
|
WWW = query_dns("CNAME", f"www.{mail_domain}")
|
||||||
|
|
||||||
res = dict(mail_domain=mail_domain, A=A, AAAA=AAAA, MTA_STS=MTA_STS, WWW=WWW)
|
res = dict(mail_domain=mail_domain, A=A, AAAA=AAAA, MTA_STS=MTA_STS, WWW=WWW)
|
||||||
res["acme_account_url"] = shell(
|
res["acme_account_url"] = shell("acmetool account-url", fail_ok=True)
|
||||||
pre_command + "acmetool account-url", fail_ok=True, print=log_progress
|
|
||||||
)
|
|
||||||
res["dkim_entry"], res["web_dkim_entry"] = get_dkim_entry(
|
res["dkim_entry"], res["web_dkim_entry"] = get_dkim_entry(
|
||||||
mail_domain, pre_command, dkim_selector="opendkim"
|
mail_domain, dkim_selector="opendkim"
|
||||||
)
|
)
|
||||||
|
|
||||||
if not MTA_STS or not WWW or (not A and not AAAA):
|
if not MTA_STS or not WWW or (not A and not AAAA):
|
||||||
return res
|
return res
|
||||||
|
|
||||||
# parse out sts-id if exists, example: "v=STSv1; id=2090123"
|
# parse out sts-id if exists, example: "v=STSv1; id=2090123"
|
||||||
mta_sts_txt = query_dns("TXT", f"_mta-sts.{mail_domain}")
|
parts = query_dns("TXT", f"_mta-sts.{mail_domain}").split("id=")
|
||||||
if not mta_sts_txt:
|
|
||||||
return res
|
|
||||||
parts = mta_sts_txt.split("id=")
|
|
||||||
res["sts_id"] = parts[1].rstrip('"') if len(parts) == 2 else ""
|
res["sts_id"] = parts[1].rstrip('"') if len(parts) == 2 else ""
|
||||||
return res
|
return res
|
||||||
|
|
||||||
|
|
||||||
def get_dkim_entry(mail_domain, pre_command, dkim_selector):
|
def get_dkim_entry(mail_domain, dkim_selector):
|
||||||
try:
|
try:
|
||||||
dkim_pubkey = shell(
|
dkim_pubkey = shell(
|
||||||
f"{pre_command}openssl rsa -in /etc/dkimkeys/{dkim_selector}.private "
|
f"openssl rsa -in /etc/dkimkeys/{dkim_selector}.private "
|
||||||
"-pubout 2>/dev/null | awk '/-/{next}{printf(\"%s\",$0)}'",
|
"-pubout 2>/dev/null | awk '/-/{next}{printf(\"%s\",$0)}'"
|
||||||
print=log_progress,
|
|
||||||
)
|
)
|
||||||
except CalledProcessError:
|
except CalledProcessError:
|
||||||
return None, None
|
return
|
||||||
dkim_value_raw = f"v=DKIM1;k=rsa;p={dkim_pubkey};s=email;t=s"
|
dkim_value_raw = f"v=DKIM1;k=rsa;p={dkim_pubkey};s=email;t=s"
|
||||||
dkim_value = '" "'.join(re.findall(".{1,255}", dkim_value_raw))
|
dkim_value = '" "'.join(re.findall(".{1,255}", dkim_value_raw))
|
||||||
web_dkim_value = "".join(re.findall(".{1,255}", dkim_value_raw))
|
web_dkim_value = "".join(re.findall(".{1,255}", dkim_value_raw))
|
||||||
name = f"{dkim_selector}._domainkey.{mail_domain}."
|
|
||||||
return (
|
return (
|
||||||
f'{name:<40} 3600 IN TXT "{dkim_value}"',
|
f'{dkim_selector}._domainkey.{mail_domain}. TXT "{dkim_value}"',
|
||||||
f'{name:<40} 3600 IN TXT "{web_dkim_value}"',
|
f'{dkim_selector}._domainkey.{mail_domain}. TXT "{web_dkim_value}"',
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@@ -68,9 +61,9 @@ def query_dns(typ, domain):
|
|||||||
# Get autoritative nameserver from the SOA record.
|
# Get autoritative nameserver from the SOA record.
|
||||||
soa_answers = [
|
soa_answers = [
|
||||||
x.split()
|
x.split()
|
||||||
for x in shell(
|
for x in shell(f"dig -r -q {domain} -t SOA +noall +authority +answer").split(
|
||||||
f"dig -r -q {domain} -t SOA +noall +authority +answer", print=log_progress
|
"\n"
|
||||||
).split("\n")
|
)
|
||||||
]
|
]
|
||||||
soa = [a for a in soa_answers if len(a) >= 3 and a[3] == "SOA"]
|
soa = [a for a in soa_answers if len(a) >= 3 and a[3] == "SOA"]
|
||||||
if not soa:
|
if not soa:
|
||||||
@@ -78,11 +71,13 @@ def query_dns(typ, domain):
|
|||||||
ns = soa[0][4]
|
ns = soa[0][4]
|
||||||
|
|
||||||
# Query authoritative nameserver directly to bypass DNS cache.
|
# Query authoritative nameserver directly to bypass DNS cache.
|
||||||
res = shell(f"dig @{ns} -r -q {domain} -t {typ} +short", print=log_progress)
|
res = shell(f"dig @{ns} -r -q {domain} -t {typ} +short")
|
||||||
return next((line for line in res.split("\n") if not line.startswith(";")), "")
|
if res:
|
||||||
|
return res.split("\n")[0]
|
||||||
|
return ""
|
||||||
|
|
||||||
|
|
||||||
def check_zonefile(zonefile, verbose=True):
|
def check_zonefile(zonefile, mail_domain):
|
||||||
"""Check expected zone file entries."""
|
"""Check expected zone file entries."""
|
||||||
required = True
|
required = True
|
||||||
required_diff = []
|
required_diff = []
|
||||||
@@ -94,8 +89,8 @@ def check_zonefile(zonefile, verbose=True):
|
|||||||
continue
|
continue
|
||||||
if not zf_line.strip() or zf_line.startswith(";"):
|
if not zf_line.strip() or zf_line.startswith(";"):
|
||||||
continue
|
continue
|
||||||
print(f"dns-checking {zf_line!r}") if verbose else log_progress("")
|
print(f"dns-checking {zf_line!r}")
|
||||||
zf_domain, _ttl, _in, zf_typ, zf_value = zf_line.split(None, 4)
|
zf_domain, zf_typ, zf_value = zf_line.split(maxsplit=2)
|
||||||
zf_domain = zf_domain.rstrip(".")
|
zf_domain = zf_domain.rstrip(".")
|
||||||
zf_value = zf_value.strip()
|
zf_value = zf_value.strip()
|
||||||
query_value = query_dns(zf_typ, zf_domain)
|
query_value = query_dns(zf_typ, zf_domain)
|
||||||
|
|||||||
@@ -1,13 +1,7 @@
|
|||||||
import sys
|
|
||||||
from subprocess import DEVNULL, CalledProcessError, check_output
|
from subprocess import DEVNULL, CalledProcessError, check_output
|
||||||
|
|
||||||
|
|
||||||
def log_progress(data):
|
def shell(command, fail_ok=False):
|
||||||
sys.stderr.write(".")
|
|
||||||
sys.stderr.flush()
|
|
||||||
|
|
||||||
|
|
||||||
def shell(command, fail_ok=False, print=print):
|
|
||||||
print(f"$ {command}")
|
print(f"$ {command}")
|
||||||
args = dict(shell=True)
|
args = dict(shell=True)
|
||||||
if fail_ok:
|
if fail_ok:
|
||||||
@@ -40,5 +34,5 @@ def dovecot_recalc_quota(user):
|
|||||||
#
|
#
|
||||||
for line in output.split("\n"):
|
for line in output.split("\n"):
|
||||||
parts = line.split()
|
parts = line.split()
|
||||||
if len(parts) >= 6 and parts[2] == "STORAGE":
|
if parts[2] == "STORAGE":
|
||||||
return dict(value=int(parts[3]), limit=int(parts[4]), percent=int(parts[5]))
|
return dict(value=int(parts[3]), limit=int(parts[4]), percent=int(parts[5]))
|
||||||
|
|||||||
@@ -1,52 +0,0 @@
|
|||||||
import shlex
|
|
||||||
|
|
||||||
from pyinfra.operations import server
|
|
||||||
|
|
||||||
from ..basedeploy import Deployer
|
|
||||||
|
|
||||||
|
|
||||||
def openssl_selfsigned_args(domain, cert_path, key_path, days=36500):
|
|
||||||
"""Return the openssl argument list for a self-signed certificate.
|
|
||||||
|
|
||||||
The certificate uses an EC P-256 key with SAN entries for *domain*,
|
|
||||||
``www.<domain>`` and ``mta-sts.<domain>``.
|
|
||||||
"""
|
|
||||||
return [
|
|
||||||
"openssl", "req", "-x509",
|
|
||||||
"-newkey", "ec", "-pkeyopt", "ec_paramgen_curve:P-256",
|
|
||||||
"-noenc", "-days", str(days),
|
|
||||||
"-keyout", str(key_path),
|
|
||||||
"-out", str(cert_path),
|
|
||||||
"-subj", f"/CN={domain}",
|
|
||||||
# Mark as end-entity cert so it cannot be used as a CA to sign others.
|
|
||||||
"-addext", "basicConstraints=critical,CA:FALSE",
|
|
||||||
"-addext", "extendedKeyUsage=serverAuth,clientAuth",
|
|
||||||
"-addext",
|
|
||||||
f"subjectAltName=DNS:{domain},DNS:www.{domain},DNS:mta-sts.{domain}",
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
class SelfSignedTlsDeployer(Deployer):
|
|
||||||
"""Generates a self-signed TLS certificate for all chatmail endpoints."""
|
|
||||||
|
|
||||||
def __init__(self, mail_domain):
|
|
||||||
self.mail_domain = mail_domain
|
|
||||||
self.cert_path = "/etc/ssl/certs/mailserver.pem"
|
|
||||||
self.key_path = "/etc/ssl/private/mailserver.key"
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def configure(self):
|
|
||||||
args = openssl_selfsigned_args(
|
|
||||||
self.mail_domain, self.cert_path, self.key_path,
|
|
||||||
)
|
|
||||||
cmd = shlex.join(args)
|
|
||||||
server.shell(
|
|
||||||
name="Generate self-signed TLS certificate if not present",
|
|
||||||
commands=[f"[ -f {self.cert_path} ] || {cmd}"],
|
|
||||||
)
|
|
||||||
|
|
||||||
def activate(self):
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
[Service]
|
|
||||||
Restart=always
|
|
||||||
RestartSec=30
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=chatmail mail storage expiration job
|
|
||||||
After=network.target
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=oneshot
|
|
||||||
User=vmail
|
|
||||||
ExecStart=/usr/local/lib/chatmaild/venv/bin/chatmail-expire /usr/local/lib/chatmaild/chatmail.ini -v --remove
|
|
||||||
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=Run Daily chatmail-expire job
|
|
||||||
|
|
||||||
[Timer]
|
|
||||||
OnCalendar=*-*-* 00:02:00
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=timers.target
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=chatmail file system storage reporting job
|
|
||||||
After=network.target
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
Type=oneshot
|
|
||||||
User=vmail
|
|
||||||
ExecStart=/usr/local/lib/chatmaild/venv/bin/chatmail-fsreport /usr/local/lib/chatmaild/chatmail.ini
|
|
||||||
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description=Run Daily Chatmail fsreport Job
|
|
||||||
|
|
||||||
[Timer]
|
|
||||||
OnCalendar=*-*-* 08:02:00
|
|
||||||
Persistent=true
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=timers.target
|
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user