Compare commits
62 Commits
2.2.2
...
ext/new_sy
Author | SHA1 | Date | |
---|---|---|---|
0f3c37bf6a | |||
|
ae5864cd91 | ||
|
e456724caf | ||
|
ed9dcc4061 | ||
|
ea8e386939 | ||
|
e0ec887118 | ||
|
a71d32ba77 | ||
|
a0f6fe9b0d | ||
|
c25647156a | ||
|
e7c4c12a98 | ||
|
90b2b5301c | ||
|
0d93a26e6d | ||
|
c29fc0f0eb | ||
|
e421c851c9 | ||
|
5b2b45233a | ||
|
888f7a4209 | ||
|
0c301a49c7 | ||
|
1fda2dd3b7 | ||
|
c4a20efe5e | ||
|
fc45f1b090 | ||
|
1480507d76 | ||
|
a1ab1e8e0a | ||
|
1e5033b461 | ||
|
7323851c6e | ||
|
08db73e55b | ||
|
9fba20475b | ||
|
9af5fce014 | ||
|
9843e14c1a | ||
|
60e6f1e23c | ||
|
6cdbcc69c7 | ||
|
ed7c714738 | ||
|
a9d783192b | ||
|
2bb5a734d1 | ||
|
9aa5c4cca9 | ||
|
9c4faab5d8 | ||
|
53c4ffdc4e | ||
|
e4144e923a | ||
|
791361c10d | ||
|
7c94bd4744 | ||
|
4b5eecd7e7 | ||
|
a6968fb7e9 | ||
|
d4853b1154 | ||
|
89df4b2425 | ||
|
0f89121b98 | ||
|
8a40ca185b | ||
|
5baeb42623 | ||
|
072e5f66cb | ||
|
b2f41d689b | ||
|
9b4aff58c7 | ||
|
a20e41574d | ||
|
72977d65ae | ||
|
7555fff1a5 | ||
|
aed12e5536 | ||
|
75efd9921d | ||
|
9219bd4723 | ||
|
73526be2ac | ||
|
b827efca2c | ||
|
6b7a4c8a23 | ||
|
47f6239268 | ||
|
0d6f65b469 | ||
|
be915aed94 | ||
|
ce938bb4a5 |
71
.github/workflows/codeql-analysis.yml
vendored
Normal file
71
.github/workflows/codeql-analysis.yml
vendored
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
# For most projects, this workflow file will not need changing; you simply need
|
||||||
|
# to commit it to your repository.
|
||||||
|
#
|
||||||
|
# You may wish to alter this file to override the set of languages analyzed,
|
||||||
|
# or to provide custom queries or build logic.
|
||||||
|
name: "CodeQL"
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches: [master]
|
||||||
|
pull_request:
|
||||||
|
# The branches below must be a subset of the branches above
|
||||||
|
branches: [master]
|
||||||
|
schedule:
|
||||||
|
- cron: '0 3 * * 6'
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
analyze:
|
||||||
|
name: Analyze
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
# Override automatic language detection by changing the below list
|
||||||
|
# Supported options are ['csharp', 'cpp', 'go', 'java', 'javascript', 'python']
|
||||||
|
language: ['java']
|
||||||
|
# Learn more...
|
||||||
|
# https://docs.github.com/en/github/finding-security-vulnerabilities-and-errors-in-your-code/configuring-code-scanning#overriding-automatic-language-detection
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v2
|
||||||
|
with:
|
||||||
|
# We must fetch at least the immediate parents so that if this is
|
||||||
|
# a pull request then we can checkout the head.
|
||||||
|
fetch-depth: 2
|
||||||
|
|
||||||
|
# If this run was triggered by a pull request event, then checkout
|
||||||
|
# the head of the pull request instead of the merge commit.
|
||||||
|
- run: git checkout HEAD^2
|
||||||
|
if: ${{ github.event_name == 'pull_request' }}
|
||||||
|
|
||||||
|
# Initializes the CodeQL tools for scanning.
|
||||||
|
- name: Initialize CodeQL
|
||||||
|
uses: github/codeql-action/init@v1
|
||||||
|
with:
|
||||||
|
languages: ${{ matrix.language }}
|
||||||
|
# If you wish to specify custom queries, you can do so here or in a config file.
|
||||||
|
# By default, queries listed here will override any specified in a config file.
|
||||||
|
# Prefix the list here with "+" to use these queries and those in the config file.
|
||||||
|
# queries: ./path/to/local/query, your-org/your-repo/queries@main
|
||||||
|
|
||||||
|
# Autobuild attempts to build any compiled languages (C/C++, C#, or Java).
|
||||||
|
# If this step fails, then you should remove it and run the build manually (see below)
|
||||||
|
- name: Autobuild
|
||||||
|
uses: github/codeql-action/autobuild@v1
|
||||||
|
|
||||||
|
# ℹ️ Command-line programs to run using the OS shell.
|
||||||
|
# 📚 https://git.io/JvXDl
|
||||||
|
|
||||||
|
# ✏️ If the Autobuild fails above, remove it and uncomment the following three lines
|
||||||
|
# and modify them (or add more) to build your code if your project
|
||||||
|
# uses a compiled language
|
||||||
|
|
||||||
|
#- run: |
|
||||||
|
# make bootstrap
|
||||||
|
# make release
|
||||||
|
|
||||||
|
- name: Perform CodeQL Analysis
|
||||||
|
uses: github/codeql-action/analyze@v1
|
10
Dockerfile
10
Dockerfile
@@ -1,3 +1,11 @@
|
|||||||
|
FROM --platform=$BUILDPLATFORM openjdk:8-jre-alpine AS builder
|
||||||
|
|
||||||
|
RUN apk update && apk add gradle git && rm -rf /var/lib/apk/* /var/cache/apk/*
|
||||||
|
|
||||||
|
WORKDIR /ma1sd
|
||||||
|
COPY . .
|
||||||
|
RUN ./gradlew shadowJar
|
||||||
|
|
||||||
FROM openjdk:8-jre-alpine
|
FROM openjdk:8-jre-alpine
|
||||||
|
|
||||||
RUN apk update && apk add bash && rm -rf /var/lib/apk/* /var/cache/apk/*
|
RUN apk update && apk add bash && rm -rf /var/lib/apk/* /var/cache/apk/*
|
||||||
@@ -15,4 +23,4 @@ CMD [ "/start.sh" ]
|
|||||||
|
|
||||||
ADD src/docker/start.sh /start.sh
|
ADD src/docker/start.sh /start.sh
|
||||||
ADD src/script/ma1sd /app/ma1sd
|
ADD src/script/ma1sd /app/ma1sd
|
||||||
ADD build/libs/ma1sd.jar /app/ma1sd.jar
|
COPY --from=builder /ma1sd/build/libs/ma1sd.jar /app/ma1sd.jar
|
||||||
|
16
DockerfileX
Normal file
16
DockerfileX
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
FROM --platform=$BUILDPLATFORM openjdk:11.0.7-jre-slim
|
||||||
|
|
||||||
|
VOLUME /etc/ma1sd
|
||||||
|
VOLUME /var/ma1sd
|
||||||
|
EXPOSE 8090
|
||||||
|
|
||||||
|
ENV JAVA_OPTS=""
|
||||||
|
ENV CONF_FILE_PATH="/etc/ma1sd/ma1sd.yaml"
|
||||||
|
ENV SIGN_KEY_PATH="/var/ma1sd/sign.key"
|
||||||
|
ENV SQLITE_DATABASE_PATH="/var/ma1sd/ma1sd.db"
|
||||||
|
|
||||||
|
CMD [ "/start.sh" ]
|
||||||
|
|
||||||
|
ADD src/docker/start.sh /start.sh
|
||||||
|
ADD src/script/ma1sd /app/ma1sd
|
||||||
|
ADD build/libs/ma1sd.jar /app/ma1sd.jar
|
95
build.gradle
95
build.gradle
@@ -20,7 +20,7 @@
|
|||||||
|
|
||||||
import java.util.regex.Pattern
|
import java.util.regex.Pattern
|
||||||
|
|
||||||
apply plugin: 'java'
|
apply plugin: 'java-library'
|
||||||
apply plugin: 'application'
|
apply plugin: 'application'
|
||||||
apply plugin: 'com.github.johnrengelman.shadow'
|
apply plugin: 'com.github.johnrengelman.shadow'
|
||||||
apply plugin: 'idea'
|
apply plugin: 'idea'
|
||||||
@@ -73,90 +73,94 @@ String gitVersion() {
|
|||||||
|
|
||||||
buildscript {
|
buildscript {
|
||||||
repositories {
|
repositories {
|
||||||
jcenter()
|
gradlePluginPortal()
|
||||||
|
mavenCentral()
|
||||||
}
|
}
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
classpath 'com.github.jengelman.gradle.plugins:shadow:5.1.0'
|
classpath 'com.github.jengelman.gradle.plugins:shadow:6.1.0'
|
||||||
classpath 'com.github.ben-manes:gradle-versions-plugin:0.27.0'
|
classpath 'com.github.ben-manes:gradle-versions-plugin:0.38.0'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
repositories {
|
repositories {
|
||||||
jcenter()
|
mavenCentral()
|
||||||
}
|
}
|
||||||
|
|
||||||
dependencies {
|
dependencies {
|
||||||
// Logging
|
// Logging
|
||||||
compile 'org.slf4j:slf4j-simple:1.7.25'
|
api 'org.slf4j:slf4j-simple:1.7.25'
|
||||||
|
|
||||||
// Easy file management
|
// Easy file management
|
||||||
compile 'commons-io:commons-io:2.6'
|
api 'commons-io:commons-io:2.8.0'
|
||||||
|
|
||||||
// Config management
|
// Config management
|
||||||
compile 'org.yaml:snakeyaml:1.25'
|
api 'org.yaml:snakeyaml:1.28'
|
||||||
|
|
||||||
// Dependencies from old Matrix-java-sdk
|
// Dependencies from old Matrix-java-sdk
|
||||||
compile 'org.apache.commons:commons-lang3:3.9'
|
api 'org.apache.commons:commons-lang3:3.12.0'
|
||||||
compile 'com.squareup.okhttp3:okhttp:4.2.2'
|
api 'com.squareup.okhttp3:okhttp:4.2.2'
|
||||||
compile 'commons-codec:commons-codec:1.13'
|
api 'commons-codec:commons-codec:1.15'
|
||||||
|
|
||||||
// ORMLite
|
// ORMLite
|
||||||
compile 'com.j256.ormlite:ormlite-jdbc:5.1'
|
api 'com.j256.ormlite:ormlite-jdbc:5.3'
|
||||||
|
|
||||||
// ed25519 handling
|
// ed25519 handling
|
||||||
compile 'net.i2p.crypto:eddsa:0.3.0'
|
api 'net.i2p.crypto:eddsa:0.3.0'
|
||||||
|
|
||||||
// LDAP connector
|
// LDAP connector
|
||||||
compile 'org.apache.directory.api:api-all:1.0.3'
|
api 'org.apache.directory.api:api-all:1.0.3'
|
||||||
|
|
||||||
// DNS lookups
|
// DNS lookups
|
||||||
compile 'dnsjava:dnsjava:2.1.9'
|
api 'dnsjava:dnsjava:2.1.9'
|
||||||
|
|
||||||
// HTTP connections
|
// HTTP connections
|
||||||
compile 'org.apache.httpcomponents:httpclient:4.5.10'
|
api 'org.apache.httpcomponents:httpclient:4.5.13'
|
||||||
|
|
||||||
// Phone numbers validation
|
// Phone numbers validation
|
||||||
compile 'com.googlecode.libphonenumber:libphonenumber:8.10.22'
|
api 'com.googlecode.libphonenumber:libphonenumber:8.12.21'
|
||||||
|
|
||||||
// E-mail sending
|
// E-mail sending
|
||||||
compile 'javax.mail:javax.mail-api:1.6.2'
|
api 'javax.mail:javax.mail-api:1.6.2'
|
||||||
compile 'com.sun.mail:javax.mail:1.6.2'
|
api 'com.sun.mail:javax.mail:1.6.2'
|
||||||
|
|
||||||
// Google Firebase Authentication backend
|
// Google Firebase Authentication backend
|
||||||
compile 'com.google.firebase:firebase-admin:5.3.0'
|
api 'com.google.firebase:firebase-admin:5.3.0'
|
||||||
|
|
||||||
// Connection Pool
|
// Connection Pool
|
||||||
compile 'com.mchange:c3p0:0.9.5.4'
|
api 'com.mchange:c3p0:0.9.5.5'
|
||||||
|
|
||||||
// SQLite
|
// SQLite
|
||||||
compile 'org.xerial:sqlite-jdbc:3.28.0'
|
api 'org.xerial:sqlite-jdbc:3.34.0'
|
||||||
|
|
||||||
// PostgreSQL
|
// PostgreSQL
|
||||||
compile 'org.postgresql:postgresql:42.2.8'
|
api 'org.postgresql:postgresql:42.2.19'
|
||||||
|
|
||||||
// MariaDB/MySQL
|
// MariaDB/MySQL
|
||||||
compile 'org.mariadb.jdbc:mariadb-java-client:2.5.1'
|
api 'org.mariadb.jdbc:mariadb-java-client:2.7.2'
|
||||||
|
|
||||||
|
// UNIX sockets
|
||||||
|
api 'com.kohlschutter.junixsocket:junixsocket-core:2.3.3'
|
||||||
|
|
||||||
// Twilio SDK for SMS
|
// Twilio SDK for SMS
|
||||||
compile 'com.twilio.sdk:twilio:7.45.0'
|
api 'com.twilio.sdk:twilio:7.45.0'
|
||||||
|
|
||||||
// SendGrid SDK to send emails from GCE
|
// SendGrid SDK to send emails from GCE
|
||||||
compile 'com.sendgrid:sendgrid-java:2.2.2'
|
api 'com.sendgrid:sendgrid-java:2.2.2'
|
||||||
|
|
||||||
// ZT-Exec for exec identity store
|
// ZT-Exec for exec identity store
|
||||||
compile 'org.zeroturnaround:zt-exec:1.11'
|
api 'org.zeroturnaround:zt-exec:1.12'
|
||||||
|
|
||||||
// HTTP server
|
// HTTP server
|
||||||
compile 'io.undertow:undertow-core:2.0.27.Final'
|
api 'io.undertow:undertow-core:2.2.7.Final'
|
||||||
|
|
||||||
// Command parser for AS interface
|
// Command parser for AS interface
|
||||||
implementation 'commons-cli:commons-cli:1.4'
|
api 'commons-cli:commons-cli:1.4'
|
||||||
|
|
||||||
testCompile 'junit:junit:4.13-rc-1'
|
testImplementation 'junit:junit:4.13.2'
|
||||||
testCompile 'com.github.tomakehurst:wiremock:2.25.1'
|
testImplementation 'com.github.tomakehurst:wiremock:2.27.2'
|
||||||
testCompile 'com.unboundid:unboundid-ldapsdk:4.0.12'
|
testImplementation 'com.unboundid:unboundid-ldapsdk:4.0.12'
|
||||||
testCompile 'com.icegreen:greenmail:1.5.11'
|
testImplementation 'com.icegreen:greenmail:1.5.11'
|
||||||
}
|
}
|
||||||
|
|
||||||
jar {
|
jar {
|
||||||
@@ -239,7 +243,7 @@ task debBuild(dependsOn: shadowJar) {
|
|||||||
|
|
||||||
ant.chmod(
|
ant.chmod(
|
||||||
file: "${debBuildDebianPath}/postinst",
|
file: "${debBuildDebianPath}/postinst",
|
||||||
perm: 'a+x'
|
perm: '0755'
|
||||||
)
|
)
|
||||||
|
|
||||||
ant.chmod(
|
ant.chmod(
|
||||||
@@ -264,7 +268,7 @@ task debBuild(dependsOn: shadowJar) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
task dockerBuild(type: Exec, dependsOn: shadowJar) {
|
task dockerBuild(type: Exec) {
|
||||||
commandLine 'docker', 'build', '-t', dockerImageTag, project.rootDir
|
commandLine 'docker', 'build', '-t', dockerImageTag, project.rootDir
|
||||||
|
|
||||||
doLast {
|
doLast {
|
||||||
@@ -274,6 +278,15 @@ task dockerBuild(type: Exec, dependsOn: shadowJar) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
task dockerBuildX(type: Exec, dependsOn: shadowJar) {
|
||||||
|
commandLine 'docker', 'buildx', 'build', '--push', '--platform', 'linux/arm64,linux/amd64,linux/arm/v7', '-t', dockerImageTag , project.rootDir
|
||||||
|
doLast {
|
||||||
|
exec {
|
||||||
|
commandLine 'docker', 'buildx', 'build', '--push', '--platform', 'linux/arm64,linux/amd64,linux/arm/v7', '-t', "${dockerImageName}:latest-dev", project.rootDir
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
task dockerPush(type: Exec) {
|
task dockerPush(type: Exec) {
|
||||||
commandLine 'docker', 'push', dockerImageTag
|
commandLine 'docker', 'push', dockerImageTag
|
||||||
|
|
||||||
@@ -283,3 +296,15 @@ task dockerPush(type: Exec) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
task dockerPushX(type: Exec) {
|
||||||
|
commandLine 'docker', 'push', dockerImageTag
|
||||||
|
|
||||||
|
doLast {
|
||||||
|
exec {
|
||||||
|
commandLine 'docker', 'push', "${dockerImageName}:latest-dev"
|
||||||
|
commandLine 'docker', 'push', "${dockerImageName}:latest-amd64-dev"
|
||||||
|
commandLine 'docker', 'push', "${dockerImageName}:latest-arm64-dev"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@@ -110,7 +110,7 @@ For sql provider (i.e. for the `synapseSql`):
|
|||||||
```.yaml
|
```.yaml
|
||||||
synapseSql:
|
synapseSql:
|
||||||
lookup:
|
lookup:
|
||||||
query: 'select user_id as mxid, medium, address from user_threepids' # query for retrive 3PIDs for hashes.
|
query: 'select user_id as mxid, medium, address from user_threepid_id_server' # query for retrive 3PIDs for hashes.
|
||||||
```
|
```
|
||||||
|
|
||||||
For general sql provider:
|
For general sql provider:
|
||||||
|
@@ -9,6 +9,8 @@
|
|||||||
## Binaries
|
## Binaries
|
||||||
### Requirements
|
### Requirements
|
||||||
- JDK 1.8
|
- JDK 1.8
|
||||||
|
- OpenJDK 11
|
||||||
|
- OpenJDK 14
|
||||||
|
|
||||||
### Build
|
### Build
|
||||||
```bash
|
```bash
|
||||||
@@ -70,5 +72,13 @@ Then follow the instruction in the [Debian package](install/debian.md) document.
|
|||||||
```
|
```
|
||||||
Then follow the instructions in the [Docker install](install/docker.md#configure) document.
|
Then follow the instructions in the [Docker install](install/docker.md#configure) document.
|
||||||
|
|
||||||
|
### Multi-platform builds
|
||||||
|
|
||||||
|
Provided with experimental docker feature [buildx](https://docs.docker.com/buildx/working-with-buildx/)
|
||||||
|
To build the arm64 and amd64 images run:
|
||||||
|
```bash
|
||||||
|
./gradlew dockerBuildX
|
||||||
|
```
|
||||||
|
|
||||||
## Next steps
|
## Next steps
|
||||||
- [Integrate with your infrastructure](getting-started.md#integrate)
|
- [Integrate with your infrastructure](getting-started.md#integrate)
|
||||||
|
@@ -58,7 +58,53 @@ Commonly the `server.publicUrl` should be the same value as the `trusted_third_p
|
|||||||
|
|
||||||
## Storage
|
## Storage
|
||||||
### SQLite
|
### SQLite
|
||||||
`storage.provider.sqlite.database`: Absolute location of the SQLite database
|
```yaml
|
||||||
|
storage:
|
||||||
|
backend: sqlite # default
|
||||||
|
provider:
|
||||||
|
sqlite:
|
||||||
|
database: /var/lib/ma1sd/store.db # Absolute location of the SQLite database
|
||||||
|
```
|
||||||
|
|
||||||
|
### Postgresql
|
||||||
|
```yaml
|
||||||
|
storage:
|
||||||
|
backend: postgresql
|
||||||
|
provider:
|
||||||
|
postgresql:
|
||||||
|
database: //localhost:5432/ma1sd
|
||||||
|
username: ma1sd
|
||||||
|
password: secret_password
|
||||||
|
```
|
||||||
|
See [the migration instruction](migration-to-postgresql.md) from sqlite to postgresql
|
||||||
|
|
||||||
|
|
||||||
|
## Logging
|
||||||
|
```yaml
|
||||||
|
logging:
|
||||||
|
root: error # default level for all loggers (apps and thirdparty libraries)
|
||||||
|
app: info # log level only for the ma1sd
|
||||||
|
requests: false # log request and response
|
||||||
|
```
|
||||||
|
|
||||||
|
Possible value: `trace`, `debug`, `info`, `warn`, `error`, `off`.
|
||||||
|
|
||||||
|
Default value for root level: `info`.
|
||||||
|
|
||||||
|
Value for app level can be specified via `MA1SD_LOG_LEVEL` environment variable, configuration or start options.
|
||||||
|
|
||||||
|
Default value for app level: `info`.
|
||||||
|
|
||||||
|
| start option | equivalent configuration |
|
||||||
|
| --- | --- |
|
||||||
|
| | app: info |
|
||||||
|
| -v | app: debug |
|
||||||
|
| -vv | app: trace |
|
||||||
|
|
||||||
|
#### WARNING
|
||||||
|
|
||||||
|
The setting `logging.requests` *MUST NOT* be used in production due it prints full unmasked request and response into the log and can be cause of the data leak.
|
||||||
|
This setting can be used only to testing and debugging errors.
|
||||||
|
|
||||||
## Identity stores
|
## Identity stores
|
||||||
See the [Identity stores](stores/README.md) for specific configuration
|
See the [Identity stores](stores/README.md) for specific configuration
|
||||||
|
@@ -56,8 +56,7 @@ Accounts cannot currently migrate/move from one server to another.
|
|||||||
See a [brief explanation document](concepts.md) about Matrix and ma1sd concepts and vocabulary.
|
See a [brief explanation document](concepts.md) about Matrix and ma1sd concepts and vocabulary.
|
||||||
|
|
||||||
### I already use the synapse LDAP3 auth provider. Why should I care about ma1sd?
|
### I already use the synapse LDAP3 auth provider. Why should I care about ma1sd?
|
||||||
The [synapse LDAP3 auth provider](https://github.com/matrix-org/matrix-synapse-ldap3) is not longer maintained despite
|
The [synapse LDAP3 auth provider](https://github.com/matrix-org/matrix-synapse-ldap3) only handles one specific flow: validate credentials at login.
|
||||||
saying so and only handles on specific flow: validate credentials at login.
|
|
||||||
|
|
||||||
It does not:
|
It does not:
|
||||||
- Auto-provision user profiles
|
- Auto-provision user profiles
|
||||||
|
@@ -1,5 +1,5 @@
|
|||||||
# Identity
|
# Identity
|
||||||
Implementation of the [Identity Service API r0.2.0](https://matrix.org/docs/spec/identity_service/r0.2.0.html).
|
Implementation of the [Identity Service API r0.3.0](https://matrix.org/docs/spec/identity_service/r0.3.0.html).
|
||||||
|
|
||||||
- [Lookups](#lookups)
|
- [Lookups](#lookups)
|
||||||
- [Invitations](#invitations)
|
- [Invitations](#invitations)
|
||||||
|
@@ -121,15 +121,13 @@ server {
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
### Synapse
|
### Synapse (Deprecated with synapse v1.4.0)
|
||||||
Add your ma1sd domain into the `homeserver.yaml` at `trusted_third_party_id_servers` and restart synapse.
|
Add your ma1sd domain into the `homeserver.yaml` at `trusted_third_party_id_servers` and restart synapse.
|
||||||
In a typical configuration, you would end up with something similar to:
|
In a typical configuration, you would end up with something similar to:
|
||||||
```yaml
|
```yaml
|
||||||
trusted_third_party_id_servers:
|
trusted_third_party_id_servers:
|
||||||
- matrix.example.org
|
- matrix.example.org
|
||||||
```
|
```
|
||||||
It is **highly recommended** to remove `matrix.org` and `vector.im` (or any other default entry) from your configuration
|
|
||||||
so only your own Identity server is authoritative for your HS.
|
|
||||||
|
|
||||||
## Validate (Under reconstruction)
|
## Validate (Under reconstruction)
|
||||||
**NOTE:** In case your homeserver has no working federation, step 5 will not happen. If step 4 took place, consider
|
**NOTE:** In case your homeserver has no working federation, step 5 will not happen. If step 4 took place, consider
|
||||||
|
41
docs/migration-to-postgresql.md
Normal file
41
docs/migration-to-postgresql.md
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
# Migration from sqlite to postgresql
|
||||||
|
|
||||||
|
Starting from the version 2.3.0 ma1sd support postgresql for internal storage in addition to sqlite (parameters `storage.backend`).
|
||||||
|
|
||||||
|
#### Migration steps
|
||||||
|
|
||||||
|
1. create the postgresql database and user for ma1sd storage
|
||||||
|
2. create a backup for sqlite storage (default location: /var/lib/ma1sd/store.db)
|
||||||
|
3. migrate data from sqlite to postgresql
|
||||||
|
4. change ma1sd configuration to use the postgresql
|
||||||
|
|
||||||
|
For data migration is it possible to use https://pgloader.io tool.
|
||||||
|
|
||||||
|
Example of the migration command:
|
||||||
|
```shell script
|
||||||
|
pgloader --with "quote identifiers" /path/to/store.db pgsql://ma1sd_user:ma1sd_password@host:port/database
|
||||||
|
```
|
||||||
|
or (short version for database on localhost)
|
||||||
|
```shell script
|
||||||
|
pgloader --with "quote identifiers" /path/to/store.db pgsql://ma1sd_user:ma1sd_password@localhost/ma1sd
|
||||||
|
```
|
||||||
|
|
||||||
|
An option `--with "quote identifies"` used to create case sensitive tables.
|
||||||
|
ma1sd_user - postgresql user for ma1sd.
|
||||||
|
ma1sd_password - password of the postgresql user.
|
||||||
|
host - postgresql host
|
||||||
|
post - database port (default 5432)
|
||||||
|
database - database name.
|
||||||
|
|
||||||
|
|
||||||
|
Configuration example for postgresql storage:
|
||||||
|
```yaml
|
||||||
|
storage:
|
||||||
|
backend: postgresql
|
||||||
|
provider:
|
||||||
|
postgresql:
|
||||||
|
database: '//localhost/ma1sd' # or full variant //192.168.1.100:5432/ma1sd_database
|
||||||
|
username: 'ma1sd_user'
|
||||||
|
password: 'ma1sd_password'
|
||||||
|
```
|
||||||
|
|
@@ -1,7 +1,10 @@
|
|||||||
# Operations Guide
|
# Operations Guide
|
||||||
- [Overview](#overview)
|
- [Operations Guide](#operations-guide)
|
||||||
- [Maintenance](#maintenance)
|
- [Overview](#overview)
|
||||||
- [Backuo](#backup)
|
- [Maintenance](#maintenance)
|
||||||
|
- [Backup](#backup)
|
||||||
|
- [Run](#run)
|
||||||
|
- [Restore](#restore)
|
||||||
|
|
||||||
## Overview
|
## Overview
|
||||||
This document gives various information for the day-to-day management and operations of ma1sd.
|
This document gives various information for the day-to-day management and operations of ma1sd.
|
||||||
|
@@ -136,7 +136,7 @@ sql:
|
|||||||
|
|
||||||
```
|
```
|
||||||
For the `role` query, `type` can be used to tell ma1sd how to inject the User ID in the query:
|
For the `role` query, `type` can be used to tell ma1sd how to inject the User ID in the query:
|
||||||
- `localpart` will extract and set only the localpart.
|
- `uid` will extract and set only the localpart.
|
||||||
- `mxid` will use the ID as-is.
|
- `mxid` will use the ID as-is.
|
||||||
|
|
||||||
On each query, the first parameter `?` is set as a string with the corresponding ID format.
|
On each query, the first parameter `?` is set as a string with the corresponding ID format.
|
||||||
|
2
gradle/wrapper/gradle-wrapper.properties
vendored
2
gradle/wrapper/gradle-wrapper.properties
vendored
@@ -1,5 +1,5 @@
|
|||||||
#Thu Dec 05 22:39:36 MSK 2019
|
#Thu Dec 05 22:39:36 MSK 2019
|
||||||
distributionUrl=https\://services.gradle.org/distributions/gradle-6.0-all.zip
|
distributionUrl=https\://services.gradle.org/distributions/gradle-7.0-all.zip
|
||||||
distributionBase=GRADLE_USER_HOME
|
distributionBase=GRADLE_USER_HOME
|
||||||
distributionPath=wrapper/dists
|
distributionPath=wrapper/dists
|
||||||
zipStorePath=wrapper/dists
|
zipStorePath=wrapper/dists
|
||||||
|
@@ -22,7 +22,7 @@
|
|||||||
matrix:
|
matrix:
|
||||||
domain: ''
|
domain: ''
|
||||||
v1: true # deprecated
|
v1: true # deprecated
|
||||||
v2: false # MSC2140 API v2. Disabled by default in order to preserve backward compatibility.
|
v2: true # MSC2140 API v2. Riot require enabled V2 API.
|
||||||
|
|
||||||
|
|
||||||
################
|
################
|
||||||
@@ -51,10 +51,39 @@ key:
|
|||||||
# - /var/lib/ma1sd/store.db
|
# - /var/lib/ma1sd/store.db
|
||||||
#
|
#
|
||||||
storage:
|
storage:
|
||||||
|
# backend: sqlite # or postgresql
|
||||||
provider:
|
provider:
|
||||||
sqlite:
|
sqlite:
|
||||||
database: '/path/to/ma1sd.db'
|
database: '/path/to/ma1sd.db'
|
||||||
|
# postgresql:
|
||||||
|
# # Wrap all string values with quotes to avoid yaml parsing mistakes
|
||||||
|
# database: '//localhost/ma1sd' # or full variant //192.168.1.100:5432/ma1sd_database
|
||||||
|
# username: 'ma1sd_user'
|
||||||
|
# password: 'ma1sd_password'
|
||||||
|
#
|
||||||
|
# # Pool configuration for postgresql backend.
|
||||||
|
# #######
|
||||||
|
# # Enable or disable pooling
|
||||||
|
# pool: false
|
||||||
|
#
|
||||||
|
# #######
|
||||||
|
# # Check database connection before get from pool
|
||||||
|
# testBeforeGetFromPool: false # or true
|
||||||
|
#
|
||||||
|
# #######
|
||||||
|
# # There is an internal thread which checks each of the database connections as a keep-alive mechanism. This set the
|
||||||
|
# # number of milliseconds it sleeps between checks -- default is 30000. To disable the checking thread, set this to
|
||||||
|
# # 0 before you start using the connection source.
|
||||||
|
# checkConnectionsEveryMillis: 30000
|
||||||
|
#
|
||||||
|
# #######
|
||||||
|
# # Set the number of connections that can be unused in the available list.
|
||||||
|
# maxConnectionsFree: 5
|
||||||
|
#
|
||||||
|
# #######
|
||||||
|
# # Set the number of milliseconds that a connection can stay open before being closed. Set to 9223372036854775807 to have
|
||||||
|
# # the connections never expire.
|
||||||
|
# maxConnectionAgeMillis: 3600000
|
||||||
|
|
||||||
###################
|
###################
|
||||||
# Identity Stores #
|
# Identity Stores #
|
||||||
@@ -129,12 +158,15 @@ threepid:
|
|||||||
### hash lookup for synapseSql provider.
|
### hash lookup for synapseSql provider.
|
||||||
# synapseSql:
|
# synapseSql:
|
||||||
# lookup:
|
# lookup:
|
||||||
# query: 'select user_id as mxid, medium, address from user_threepids' # query for retrive 3PIDs for hashes.
|
# query: 'select user_id as mxid, medium, address from user_threepid_id_server' # query for retrive 3PIDs for hashes.
|
||||||
|
# legacyRoomNames: false # use the old query to get room names.
|
||||||
|
|
||||||
### hash lookup for ldap provider (with example of the ldap configuration)
|
### hash lookup for ldap provider (with example of the ldap configuration)
|
||||||
# ldap:
|
# ldap:
|
||||||
# enabled: true
|
# enabled: true
|
||||||
# lookup: true # hash lookup
|
# lookup: true # hash lookup
|
||||||
|
# activeDirectory: false
|
||||||
|
# defaultDomain: ''
|
||||||
# connection:
|
# connection:
|
||||||
# host: 'ldap.domain.tld'
|
# host: 'ldap.domain.tld'
|
||||||
# port: 389
|
# port: 389
|
||||||
@@ -167,3 +199,21 @@ threepid:
|
|||||||
# - '/_matrix/identity/v2/hash_details'
|
# - '/_matrix/identity/v2/hash_details'
|
||||||
# - '/_matrix/identity/v2/lookup'
|
# - '/_matrix/identity/v2/lookup'
|
||||||
#
|
#
|
||||||
|
|
||||||
|
# logging:
|
||||||
|
# root: error # default level for all loggers (apps and thirdparty libraries)
|
||||||
|
# app: info # log level only for the ma1sd
|
||||||
|
# requests: false # or true to dump full requests and responses
|
||||||
|
|
||||||
|
|
||||||
|
# Config invitation manager
|
||||||
|
#invite:
|
||||||
|
# fullDisplayName: true # print full name of the invited user (default false)
|
||||||
|
# resolution:
|
||||||
|
# timer: 10
|
||||||
|
# period: seconds # search invites every 10 seconds (by default 5 minutes)
|
||||||
|
|
||||||
|
|
||||||
|
# Internal API
|
||||||
|
#internal:
|
||||||
|
# enabled: true # default to false
|
||||||
|
@@ -27,7 +27,7 @@ public class ThreePid implements _ThreePid {
|
|||||||
|
|
||||||
public ThreePid(String medium, String address) {
|
public ThreePid(String medium, String address) {
|
||||||
this.medium = medium;
|
this.medium = medium;
|
||||||
this.address = address;
|
this.address = address.toLowerCase();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@@ -23,11 +23,13 @@ package io.kamax.mxisd;
|
|||||||
import io.kamax.mxisd.config.MatrixConfig;
|
import io.kamax.mxisd.config.MatrixConfig;
|
||||||
import io.kamax.mxisd.config.MxisdConfig;
|
import io.kamax.mxisd.config.MxisdConfig;
|
||||||
import io.kamax.mxisd.config.PolicyConfig;
|
import io.kamax.mxisd.config.PolicyConfig;
|
||||||
|
import io.kamax.mxisd.config.ServerConfig;
|
||||||
import io.kamax.mxisd.http.undertow.handler.ApiHandler;
|
import io.kamax.mxisd.http.undertow.handler.ApiHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.AuthorizationHandler;
|
import io.kamax.mxisd.http.undertow.handler.AuthorizationHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.CheckTermsHandler;
|
import io.kamax.mxisd.http.undertow.handler.CheckTermsHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.InternalInfoHandler;
|
import io.kamax.mxisd.http.undertow.handler.InternalInfoHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.OptionsHandler;
|
import io.kamax.mxisd.http.undertow.handler.OptionsHandler;
|
||||||
|
import io.kamax.mxisd.http.undertow.handler.RequestDumpingHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.SaneHandler;
|
import io.kamax.mxisd.http.undertow.handler.SaneHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.as.v1.AsNotFoundHandler;
|
import io.kamax.mxisd.http.undertow.handler.as.v1.AsNotFoundHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.as.v1.AsTransactionHandler;
|
import io.kamax.mxisd.http.undertow.handler.as.v1.AsTransactionHandler;
|
||||||
@@ -56,6 +58,7 @@ import io.kamax.mxisd.http.undertow.handler.identity.v1.BulkLookupHandler;
|
|||||||
import io.kamax.mxisd.http.undertow.handler.identity.v1.SingleLookupHandler;
|
import io.kamax.mxisd.http.undertow.handler.identity.v1.SingleLookupHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.identity.v2.HashDetailsHandler;
|
import io.kamax.mxisd.http.undertow.handler.identity.v2.HashDetailsHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.identity.v2.HashLookupHandler;
|
import io.kamax.mxisd.http.undertow.handler.identity.v2.HashLookupHandler;
|
||||||
|
import io.kamax.mxisd.http.undertow.handler.internal.InternalInviteManagerHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.invite.v1.RoomInviteHandler;
|
import io.kamax.mxisd.http.undertow.handler.invite.v1.RoomInviteHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.profile.v1.InternalProfileHandler;
|
import io.kamax.mxisd.http.undertow.handler.profile.v1.InternalProfileHandler;
|
||||||
import io.kamax.mxisd.http.undertow.handler.profile.v1.ProfileHandler;
|
import io.kamax.mxisd.http.undertow.handler.profile.v1.ProfileHandler;
|
||||||
@@ -99,9 +102,9 @@ public class HttpMxisd {
|
|||||||
public void start() {
|
public void start() {
|
||||||
m.start();
|
m.start();
|
||||||
|
|
||||||
HttpHandler asUserHandler = SaneHandler.around(new AsUserHandler(m.getAs()));
|
HttpHandler asUserHandler = sane(new AsUserHandler(m.getAs()));
|
||||||
HttpHandler asTxnHandler = SaneHandler.around(new AsTransactionHandler(m.getAs()));
|
HttpHandler asTxnHandler = sane(new AsTransactionHandler(m.getAs()));
|
||||||
HttpHandler asNotFoundHandler = SaneHandler.around(new AsNotFoundHandler(m.getAs()));
|
HttpHandler asNotFoundHandler = sane(new AsNotFoundHandler(m.getAs()));
|
||||||
|
|
||||||
final RoutingHandler handler = Handlers.routing()
|
final RoutingHandler handler = Handlers.routing()
|
||||||
.add("OPTIONS", "/**", sane(new OptionsHandler()))
|
.add("OPTIONS", "/**", sane(new OptionsHandler()))
|
||||||
@@ -145,7 +148,13 @@ public class HttpMxisd {
|
|||||||
termsEndpoints(handler);
|
termsEndpoints(handler);
|
||||||
hashEndpoints(handler);
|
hashEndpoints(handler);
|
||||||
accountEndpoints(handler);
|
accountEndpoints(handler);
|
||||||
httpSrv = Undertow.builder().addHttpListener(m.getConfig().getServer().getPort(), "0.0.0.0").setHandler(handler).build();
|
|
||||||
|
if (m.getConfig().getInternal().isEnabled()) {
|
||||||
|
handler.get(InternalInviteManagerHandler.PATH, new InternalInviteManagerHandler(m.getInvite()));
|
||||||
|
}
|
||||||
|
|
||||||
|
ServerConfig serverConfig = m.getConfig().getServer();
|
||||||
|
httpSrv = Undertow.builder().addHttpListener(serverConfig.getPort(), serverConfig.getHostname()).setHandler(handler).build();
|
||||||
|
|
||||||
httpSrv.start();
|
httpSrv.start();
|
||||||
}
|
}
|
||||||
@@ -265,6 +274,11 @@ public class HttpMxisd {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private HttpHandler sane(HttpHandler httpHandler) {
|
private HttpHandler sane(HttpHandler httpHandler) {
|
||||||
return SaneHandler.around(httpHandler);
|
SaneHandler handler = SaneHandler.around(httpHandler);
|
||||||
|
if (m.getConfig().getLogging().isRequests()) {
|
||||||
|
return new RequestDumpingHandler(handler);
|
||||||
|
} else {
|
||||||
|
return handler;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -27,6 +27,7 @@ import io.kamax.mxisd.auth.AuthProviders;
|
|||||||
import io.kamax.mxisd.backend.IdentityStoreSupplier;
|
import io.kamax.mxisd.backend.IdentityStoreSupplier;
|
||||||
import io.kamax.mxisd.backend.sql.synapse.Synapse;
|
import io.kamax.mxisd.backend.sql.synapse.Synapse;
|
||||||
import io.kamax.mxisd.config.MxisdConfig;
|
import io.kamax.mxisd.config.MxisdConfig;
|
||||||
|
import io.kamax.mxisd.config.StorageConfig;
|
||||||
import io.kamax.mxisd.crypto.CryptoFactory;
|
import io.kamax.mxisd.crypto.CryptoFactory;
|
||||||
import io.kamax.mxisd.crypto.KeyManager;
|
import io.kamax.mxisd.crypto.KeyManager;
|
||||||
import io.kamax.mxisd.crypto.SignatureManager;
|
import io.kamax.mxisd.crypto.SignatureManager;
|
||||||
@@ -66,7 +67,7 @@ public class Mxisd {
|
|||||||
public static final String Version = StringUtils.defaultIfBlank(Mxisd.class.getPackage().getImplementationVersion(), "UNKNOWN");
|
public static final String Version = StringUtils.defaultIfBlank(Mxisd.class.getPackage().getImplementationVersion(), "UNKNOWN");
|
||||||
public static final String Agent = Name + "/" + Version;
|
public static final String Agent = Name + "/" + Version;
|
||||||
|
|
||||||
private MxisdConfig cfg;
|
private final MxisdConfig cfg;
|
||||||
|
|
||||||
private CloseableHttpClient httpClient;
|
private CloseableHttpClient httpClient;
|
||||||
private IRemoteIdentityServerFetcher srvFetcher;
|
private IRemoteIdentityServerFetcher srvFetcher;
|
||||||
@@ -109,7 +110,10 @@ public class Mxisd {
|
|||||||
IdentityServerUtils.setHttpClient(httpClient);
|
IdentityServerUtils.setHttpClient(httpClient);
|
||||||
srvFetcher = new RemoteIdentityServerFetcher(httpClient);
|
srvFetcher = new RemoteIdentityServerFetcher(httpClient);
|
||||||
|
|
||||||
store = new OrmLiteSqlStorage(cfg);
|
StorageConfig.BackendEnum storageBackend = cfg.getStorage().getBackend();
|
||||||
|
StorageConfig.Provider storageProvider = cfg.getStorage().getProvider();
|
||||||
|
store = new OrmLiteSqlStorage(storageBackend, storageProvider);
|
||||||
|
|
||||||
keyMgr = CryptoFactory.getKeyManager(cfg.getKey());
|
keyMgr = CryptoFactory.getKeyManager(cfg.getKey());
|
||||||
signMgr = CryptoFactory.getSignatureManager(cfg, keyMgr);
|
signMgr = CryptoFactory.getSignatureManager(cfg, keyMgr);
|
||||||
clientDns = new ClientDnsOverwrite(cfg.getDns().getOverwrite());
|
clientDns = new ClientDnsOverwrite(cfg.getDns().getOverwrite());
|
||||||
|
@@ -44,28 +44,43 @@ public class MxisdStandaloneExec {
|
|||||||
try {
|
try {
|
||||||
MxisdConfig cfg = null;
|
MxisdConfig cfg = null;
|
||||||
Iterator<String> argsIt = Arrays.asList(args).iterator();
|
Iterator<String> argsIt = Arrays.asList(args).iterator();
|
||||||
|
boolean dump = false;
|
||||||
|
boolean exit = false;
|
||||||
while (argsIt.hasNext()) {
|
while (argsIt.hasNext()) {
|
||||||
String arg = argsIt.next();
|
String arg = argsIt.next();
|
||||||
if (StringUtils.equalsAny(arg, "-h", "--help", "-?", "--usage")) {
|
switch (arg) {
|
||||||
|
case "-h":
|
||||||
|
case "--help":
|
||||||
|
case "-?":
|
||||||
|
case "--usage":
|
||||||
System.out.println("Available arguments:" + System.lineSeparator());
|
System.out.println("Available arguments:" + System.lineSeparator());
|
||||||
System.out.println(" -h, --help Show this help message");
|
System.out.println(" -h, --help Show this help message");
|
||||||
System.out.println(" --version Print the version then exit");
|
System.out.println(" --version Print the version then exit");
|
||||||
System.out.println(" -c, --config Set the configuration file location");
|
System.out.println(" -c, --config Set the configuration file location");
|
||||||
System.out.println(" -v Increase log level (log more info)");
|
System.out.println(" -v Increase log level (log more info)");
|
||||||
System.out.println(" -vv Further increase log level");
|
System.out.println(" -vv Further increase log level");
|
||||||
|
System.out.println(" --dump Dump the full ma1sd configuration");
|
||||||
|
System.out.println(" --dump-and-exit Dump the full ma1sd configuration and exit");
|
||||||
System.out.println(" ");
|
System.out.println(" ");
|
||||||
System.exit(0);
|
System.exit(0);
|
||||||
} else if (StringUtils.equals(arg, "-v")) {
|
return;
|
||||||
|
case "-v":
|
||||||
System.setProperty("org.slf4j.simpleLogger.log.io.kamax.mxisd", "debug");
|
System.setProperty("org.slf4j.simpleLogger.log.io.kamax.mxisd", "debug");
|
||||||
} else if (StringUtils.equals(arg, "-vv")) {
|
break;
|
||||||
|
case "-vv":
|
||||||
System.setProperty("org.slf4j.simpleLogger.log.io.kamax.mxisd", "trace");
|
System.setProperty("org.slf4j.simpleLogger.log.io.kamax.mxisd", "trace");
|
||||||
} else if (StringUtils.equalsAny(arg, "-c", "--config")) {
|
break;
|
||||||
|
case "-c":
|
||||||
|
case "--config":
|
||||||
String cfgFile = argsIt.next();
|
String cfgFile = argsIt.next();
|
||||||
cfg = YamlConfigLoader.loadFromFile(cfgFile);
|
cfg = YamlConfigLoader.loadFromFile(cfgFile);
|
||||||
} else if (StringUtils.equals("--version", arg)) {
|
break;
|
||||||
System.out.println(Mxisd.Version);
|
case "--dump-and-exit":
|
||||||
System.exit(0);
|
exit = true;
|
||||||
} else {
|
case "--dump":
|
||||||
|
dump = true;
|
||||||
|
break;
|
||||||
|
default:
|
||||||
System.err.println("Invalid argument: " + arg);
|
System.err.println("Invalid argument: " + arg);
|
||||||
System.err.println("Try '--help' for available arguments");
|
System.err.println("Try '--help' for available arguments");
|
||||||
System.exit(1);
|
System.exit(1);
|
||||||
@@ -76,6 +91,13 @@ public class MxisdStandaloneExec {
|
|||||||
cfg = YamlConfigLoader.tryLoadFromFile("ma1sd.yaml").orElseGet(MxisdConfig::new);
|
cfg = YamlConfigLoader.tryLoadFromFile("ma1sd.yaml").orElseGet(MxisdConfig::new);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (dump) {
|
||||||
|
YamlConfigLoader.dumpConfig(cfg);
|
||||||
|
if (exit) {
|
||||||
|
System.exit(0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
log.info("ma1sd starting");
|
log.info("ma1sd starting");
|
||||||
log.info("Version: {}", Mxisd.Version);
|
log.info("Version: {}", Mxisd.Version);
|
||||||
|
|
||||||
|
@@ -144,7 +144,13 @@ public class MembershipEventProcessor implements EventTypeProcessor {
|
|||||||
.collect(Collectors.toList());
|
.collect(Collectors.toList());
|
||||||
log.info("Found {} email(s) in identity store for {}", tpids.size(), inviteeId);
|
log.info("Found {} email(s) in identity store for {}", tpids.size(), inviteeId);
|
||||||
|
|
||||||
for (_ThreePid tpid : tpids) {
|
log.info("Removing duplicates from identity store");
|
||||||
|
List<_ThreePid> uniqueTpids = tpids.stream()
|
||||||
|
.distinct()
|
||||||
|
.collect(Collectors.toList());
|
||||||
|
log.info("There are {} unique email(s) in identity store for {}", uniqueTpids.size(), inviteeId);
|
||||||
|
|
||||||
|
for (_ThreePid tpid : uniqueTpids) {
|
||||||
log.info("Found Email to notify about room invitation: {}", tpid.getAddress());
|
log.info("Found Email to notify about room invitation: {}", tpid.getAddress());
|
||||||
Map<String, String> properties = new HashMap<>();
|
Map<String, String> properties = new HashMap<>();
|
||||||
profiler.getDisplayName(sender).ifPresent(name -> properties.put("sender_display_name", name));
|
profiler.getDisplayName(sender).ifPresent(name -> properties.put("sender_display_name", name));
|
||||||
|
@@ -54,6 +54,8 @@ public class LdapAuthProvider extends LdapBackend implements AuthenticatorProvid
|
|||||||
|
|
||||||
private transient final Logger log = LoggerFactory.getLogger(LdapAuthProvider.class);
|
private transient final Logger log = LoggerFactory.getLogger(LdapAuthProvider.class);
|
||||||
|
|
||||||
|
public static final char[] CHARACTERS_TO_ESCAPE = ",#+<>;\"=*\\\\".toCharArray();
|
||||||
|
|
||||||
private PhoneNumberUtil phoneUtil = PhoneNumberUtil.getInstance();
|
private PhoneNumberUtil phoneUtil = PhoneNumberUtil.getInstance();
|
||||||
|
|
||||||
public LdapAuthProvider(LdapConfig cfg, MatrixConfig mxCfg) {
|
public LdapAuthProvider(LdapConfig cfg, MatrixConfig mxCfg) {
|
||||||
@@ -94,7 +96,8 @@ public class LdapAuthProvider extends LdapBackend implements AuthenticatorProvid
|
|||||||
return BackendAuthResult.failure();
|
return BackendAuthResult.failure();
|
||||||
}
|
}
|
||||||
|
|
||||||
String userFilter = "(" + getUidAtt() + "=" + userFilterValue + ")";
|
String filteredValue = escape(userFilterValue);
|
||||||
|
String userFilter = "(" + getUidAtt() + "=" + filteredValue + ")";
|
||||||
userFilter = buildWithFilter(userFilter, getCfg().getAuth().getFilter());
|
userFilter = buildWithFilter(userFilter, getCfg().getAuth().getFilter());
|
||||||
|
|
||||||
Set<String> attributes = new HashSet<>();
|
Set<String> attributes = new HashSet<>();
|
||||||
@@ -162,8 +165,21 @@ public class LdapAuthProvider extends LdapBackend implements AuthenticatorProvid
|
|||||||
log.info("No match were found for {}", mxid);
|
log.info("No match were found for {}", mxid);
|
||||||
return BackendAuthResult.failure();
|
return BackendAuthResult.failure();
|
||||||
} catch (LdapException | IOException | CursorException e) {
|
} catch (LdapException | IOException | CursorException e) {
|
||||||
|
log.error("Unable to invoke query request: ", e);
|
||||||
throw new InternalServerError(e);
|
throw new InternalServerError(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private String escape(String raw) {
|
||||||
|
StringBuilder sb = new StringBuilder();
|
||||||
|
boolean escape;
|
||||||
|
for (char c : raw.toCharArray()) {
|
||||||
|
escape = false;
|
||||||
|
for (int i = 0; i < CHARACTERS_TO_ESCAPE.length && !escape; i++) {
|
||||||
|
escape = CHARACTERS_TO_ESCAPE[i] == c;
|
||||||
|
}
|
||||||
|
sb.append(escape ? "\\" + c : c);
|
||||||
|
}
|
||||||
|
return sb.toString();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@@ -20,6 +20,7 @@
|
|||||||
|
|
||||||
package io.kamax.mxisd.backend.ldap;
|
package io.kamax.mxisd.backend.ldap;
|
||||||
|
|
||||||
|
import io.kamax.matrix.MatrixID;
|
||||||
import io.kamax.matrix._MatrixID;
|
import io.kamax.matrix._MatrixID;
|
||||||
import io.kamax.mxisd.config.MatrixConfig;
|
import io.kamax.mxisd.config.MatrixConfig;
|
||||||
import io.kamax.mxisd.config.ldap.LdapConfig;
|
import io.kamax.mxisd.config.ldap.LdapConfig;
|
||||||
@@ -116,10 +117,20 @@ public abstract class LdapBackend {
|
|||||||
|
|
||||||
public String buildMatrixIdFromUid(String uid) {
|
public String buildMatrixIdFromUid(String uid) {
|
||||||
String uidType = getCfg().getAttribute().getUid().getType();
|
String uidType = getCfg().getAttribute().getUid().getType();
|
||||||
|
String localpart = uid.toLowerCase();
|
||||||
|
|
||||||
|
if (!StringUtils.equals(uid, localpart)) {
|
||||||
|
log.info("UID {} from LDAP has been changed to lowercase to match the Synapse specifications", uid);
|
||||||
|
}
|
||||||
|
|
||||||
if (StringUtils.equals(UID, uidType)) {
|
if (StringUtils.equals(UID, uidType)) {
|
||||||
return "@" + uid + ":" + mxCfg.getDomain();
|
if(getCfg().isActiveDirectory()) {
|
||||||
|
localpart = new UPN(uid.toLowerCase()).getMXID();
|
||||||
|
}
|
||||||
|
|
||||||
|
return "@" + localpart + ":" + mxCfg.getDomain();
|
||||||
} else if (StringUtils.equals(MATRIX_ID, uidType)) {
|
} else if (StringUtils.equals(MATRIX_ID, uidType)) {
|
||||||
return uid;
|
return localpart;
|
||||||
} else {
|
} else {
|
||||||
throw new IllegalArgumentException("Bind type " + uidType + " is not supported");
|
throw new IllegalArgumentException("Bind type " + uidType + " is not supported");
|
||||||
}
|
}
|
||||||
@@ -128,6 +139,10 @@ public abstract class LdapBackend {
|
|||||||
public String buildUidFromMatrixId(_MatrixID mxId) {
|
public String buildUidFromMatrixId(_MatrixID mxId) {
|
||||||
String uidType = getCfg().getAttribute().getUid().getType();
|
String uidType = getCfg().getAttribute().getUid().getType();
|
||||||
if (StringUtils.equals(UID, uidType)) {
|
if (StringUtils.equals(UID, uidType)) {
|
||||||
|
if(getCfg().isActiveDirectory()) {
|
||||||
|
return new UPN(mxId).getUPN();
|
||||||
|
}
|
||||||
|
|
||||||
return mxId.getLocalPart();
|
return mxId.getLocalPart();
|
||||||
} else if (StringUtils.equals(MATRIX_ID, uidType)) {
|
} else if (StringUtils.equals(MATRIX_ID, uidType)) {
|
||||||
return mxId.getId();
|
return mxId.getId();
|
||||||
@@ -169,4 +184,58 @@ public abstract class LdapBackend {
|
|||||||
return values;
|
return values;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private class UPN {
|
||||||
|
private String login;
|
||||||
|
private String domain;
|
||||||
|
|
||||||
|
public UPN(String userPrincipalName) {
|
||||||
|
String[] uidParts = userPrincipalName.split("@");
|
||||||
|
|
||||||
|
if (uidParts.length != 2) {
|
||||||
|
throw new IllegalArgumentException(String.format("Wrong userPrincipalName provided: %s", userPrincipalName));
|
||||||
|
}
|
||||||
|
|
||||||
|
this.login = uidParts[0];
|
||||||
|
this.domain = uidParts[1];
|
||||||
|
}
|
||||||
|
|
||||||
|
public UPN(_MatrixID mxid) {
|
||||||
|
String[] idParts = mxid.getLocalPart().split("/");
|
||||||
|
|
||||||
|
if (idParts.length != 2) {
|
||||||
|
if(idParts.length == 1 && !StringUtils.isEmpty(getCfg().getDefaultDomain())) {
|
||||||
|
throw new IllegalArgumentException(String.format(
|
||||||
|
"Local part of mxid %s does not contains domain separator and default domain is not configured",
|
||||||
|
mxid.getLocalPart()
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
this.domain = getCfg().getDefaultDomain();
|
||||||
|
} else {
|
||||||
|
this.domain = idParts[1];
|
||||||
|
}
|
||||||
|
|
||||||
|
this.login = idParts[0];
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getLogin() {
|
||||||
|
return login;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getDomain() {
|
||||||
|
return domain;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getMXID() {
|
||||||
|
if(StringUtils.equalsIgnoreCase(getCfg().getDefaultDomain(), this.domain)) {
|
||||||
|
return this.login;
|
||||||
|
}
|
||||||
|
|
||||||
|
return new StringBuilder(this.login).append("/").append(this.domain).toString();
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getUPN() {
|
||||||
|
return new StringBuilder(this.login).append("@").append(this.domain).toString();
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@@ -29,15 +29,19 @@ import java.util.Optional;
|
|||||||
|
|
||||||
public class Synapse {
|
public class Synapse {
|
||||||
|
|
||||||
private SqlConnectionPool pool;
|
private final SqlConnectionPool pool;
|
||||||
|
private final SynapseSqlProviderConfig providerConfig;
|
||||||
|
|
||||||
public Synapse(SynapseSqlProviderConfig sqlCfg) {
|
public Synapse(SynapseSqlProviderConfig sqlCfg) {
|
||||||
this.pool = new SqlConnectionPool(sqlCfg);
|
this.pool = new SqlConnectionPool(sqlCfg);
|
||||||
|
providerConfig = sqlCfg;
|
||||||
}
|
}
|
||||||
|
|
||||||
public Optional<String> getRoomName(String id) {
|
public Optional<String> getRoomName(String id) {
|
||||||
|
String query = providerConfig.isLegacyRoomNames() ? SynapseQueries.getLegacyRoomName() : SynapseQueries.getRoomName();
|
||||||
|
|
||||||
return pool.withConnFunction(conn -> {
|
return pool.withConnFunction(conn -> {
|
||||||
PreparedStatement stmt = conn.prepareStatement(SynapseQueries.getRoomName());
|
try (PreparedStatement stmt = conn.prepareStatement(query)) {
|
||||||
stmt.setString(1, id);
|
stmt.setString(1, id);
|
||||||
ResultSet rSet = stmt.executeQuery();
|
ResultSet rSet = stmt.executeQuery();
|
||||||
if (!rSet.next()) {
|
if (!rSet.next()) {
|
||||||
@@ -45,7 +49,7 @@ public class Synapse {
|
|||||||
}
|
}
|
||||||
|
|
||||||
return Optional.ofNullable(rSet.getString(1));
|
return Optional.ofNullable(rSet.getString(1));
|
||||||
|
}
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@@ -51,7 +51,7 @@ public class SynapseQueries {
|
|||||||
if (StringUtils.equals("sqlite", type)) {
|
if (StringUtils.equals("sqlite", type)) {
|
||||||
return "select " + getUserId(type, domain) + ", displayname from profiles p where displayname like ?";
|
return "select " + getUserId(type, domain) + ", displayname from profiles p where displayname like ?";
|
||||||
} else if (StringUtils.equals("postgresql", type)) {
|
} else if (StringUtils.equals("postgresql", type)) {
|
||||||
return "select " + getUserId(type, domain) + ", displayname from profiles p where displayname ilike ?";
|
return "SELECT u.name,p.displayname FROM users u JOIN profiles p ON u.name LIKE concat('@',p.user_id,':%') WHERE u.is_guest = 0 AND u.appservice_id IS NULL AND p.displayname LIKE ?";
|
||||||
} else {
|
} else {
|
||||||
throw new ConfigurationException("Invalid Synapse SQL type: " + type);
|
throw new ConfigurationException("Invalid Synapse SQL type: " + type);
|
||||||
}
|
}
|
||||||
@@ -72,7 +72,10 @@ public class SynapseQueries {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public static String getRoomName() {
|
public static String getRoomName() {
|
||||||
return "select r.name from room_names r, events e, (select r1.room_id,max(e1.origin_server_ts) ts from room_names r1, events e1 where r1.event_id = e1.event_id group by r1.room_id) rle where e.origin_server_ts = rle.ts and r.event_id = e.event_id and r.room_id = ?";
|
return "select name from room_stats_state where room_id = ? limit 1";
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static String getLegacyRoomName() {
|
||||||
|
return "select r.name from room_names r, events e, (select r1.room_id,max(e1.origin_server_ts) ts from room_names r1, events e1 where r1.event_id = e1.event_id group by r1.room_id) rle where e.origin_server_ts = rle.ts and r.event_id = e.event_id and r.room_id = ?";
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@@ -0,0 +1,5 @@
|
|||||||
|
package io.kamax.mxisd.config;
|
||||||
|
|
||||||
|
public interface DatabaseStorageConfig {
|
||||||
|
String getDatabase();
|
||||||
|
}
|
@@ -5,6 +5,7 @@ import org.slf4j.LoggerFactory;
|
|||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.Objects;
|
||||||
|
|
||||||
public class HashingConfig {
|
public class HashingConfig {
|
||||||
|
|
||||||
@@ -13,18 +14,19 @@ public class HashingConfig {
|
|||||||
private boolean enabled = false;
|
private boolean enabled = false;
|
||||||
private int pepperLength = 20;
|
private int pepperLength = 20;
|
||||||
private RotationPolicyEnum rotationPolicy;
|
private RotationPolicyEnum rotationPolicy;
|
||||||
private HashStorageEnum hashStorageType;
|
private HashStorageEnum hashStorageType = HashStorageEnum.in_memory;
|
||||||
private String delay = "10s";
|
private String delay = "10s";
|
||||||
private transient long delayInSeconds = 10;
|
private transient long delayInSeconds = 10;
|
||||||
private int requests = 10;
|
private int requests = 10;
|
||||||
private List<Algorithm> algorithms = new ArrayList<>();
|
private List<Algorithm> algorithms = new ArrayList<>();
|
||||||
|
|
||||||
public void build() {
|
public void build(MatrixConfig matrixConfig) {
|
||||||
if (isEnabled()) {
|
if (isEnabled()) {
|
||||||
LOGGER.info("--- Hash configuration ---");
|
LOGGER.info("--- Hash configuration ---");
|
||||||
LOGGER.info(" Pepper length: {}", getPepperLength());
|
LOGGER.info(" Pepper length: {}", getPepperLength());
|
||||||
LOGGER.info(" Rotation policy: {}", getRotationPolicy());
|
LOGGER.info(" Rotation policy: {}", getRotationPolicy());
|
||||||
LOGGER.info(" Hash storage type: {}", getHashStorageType());
|
LOGGER.info(" Hash storage type: {}", getHashStorageType());
|
||||||
|
Objects.requireNonNull(getHashStorageType(), "Storage type must be specified");
|
||||||
if (RotationPolicyEnum.per_seconds == getRotationPolicy()) {
|
if (RotationPolicyEnum.per_seconds == getRotationPolicy()) {
|
||||||
setDelayInSeconds(new DurationDeserializer().deserialize(getDelay()));
|
setDelayInSeconds(new DurationDeserializer().deserialize(getDelay()));
|
||||||
LOGGER.info(" Rotation delay: {}", getDelay());
|
LOGGER.info(" Rotation delay: {}", getDelay());
|
||||||
@@ -35,6 +37,9 @@ public class HashingConfig {
|
|||||||
}
|
}
|
||||||
LOGGER.info(" Algorithms: {}", getAlgorithms());
|
LOGGER.info(" Algorithms: {}", getAlgorithms());
|
||||||
} else {
|
} else {
|
||||||
|
if (matrixConfig.isV2()) {
|
||||||
|
LOGGER.warn("V2 enabled without the hash configuration.");
|
||||||
|
}
|
||||||
LOGGER.info("Hash configuration disabled, used only `none` pepper.");
|
LOGGER.info("Hash configuration disabled, used only `none` pepper.");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
24
src/main/java/io/kamax/mxisd/config/InternalAPIConfig.java
Normal file
24
src/main/java/io/kamax/mxisd/config/InternalAPIConfig.java
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
package io.kamax.mxisd.config;
|
||||||
|
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
public class InternalAPIConfig {
|
||||||
|
|
||||||
|
private final static Logger log = LoggerFactory.getLogger(InternalAPIConfig.class);
|
||||||
|
|
||||||
|
private boolean enabled = false;
|
||||||
|
|
||||||
|
public boolean isEnabled() {
|
||||||
|
return enabled;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setEnabled(boolean enabled) {
|
||||||
|
this.enabled = enabled;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void build() {
|
||||||
|
log.info("--- Internal API config ---");
|
||||||
|
log.info("Internal API enabled: {}", isEnabled());
|
||||||
|
}
|
||||||
|
}
|
@@ -67,6 +67,7 @@ public class InvitationConfig {
|
|||||||
|
|
||||||
private boolean recursive = true;
|
private boolean recursive = true;
|
||||||
private long timer = 5;
|
private long timer = 5;
|
||||||
|
private PeriodDimension period = PeriodDimension.minutes;
|
||||||
|
|
||||||
public boolean isRecursive() {
|
public boolean isRecursive() {
|
||||||
return recursive;
|
return recursive;
|
||||||
@@ -84,6 +85,13 @@ public class InvitationConfig {
|
|||||||
this.timer = timer;
|
this.timer = timer;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public PeriodDimension getPeriod() {
|
||||||
|
return period;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setPeriod(PeriodDimension period) {
|
||||||
|
this.period = period;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public static class SenderPolicy {
|
public static class SenderPolicy {
|
||||||
@@ -115,6 +123,7 @@ public class InvitationConfig {
|
|||||||
private Expiration expiration = new Expiration();
|
private Expiration expiration = new Expiration();
|
||||||
private Resolution resolution = new Resolution();
|
private Resolution resolution = new Resolution();
|
||||||
private Policies policy = new Policies();
|
private Policies policy = new Policies();
|
||||||
|
private boolean fullDisplayName = false;
|
||||||
|
|
||||||
public Expiration getExpiration() {
|
public Expiration getExpiration() {
|
||||||
return expiration;
|
return expiration;
|
||||||
@@ -140,11 +149,26 @@ public class InvitationConfig {
|
|||||||
this.policy = policy;
|
this.policy = policy;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public boolean isFullDisplayName() {
|
||||||
|
return fullDisplayName;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setFullDisplayName(boolean fullDisplayName) {
|
||||||
|
this.fullDisplayName = fullDisplayName;
|
||||||
|
}
|
||||||
|
|
||||||
public void build() {
|
public void build() {
|
||||||
log.info("--- Invite config ---");
|
log.info("--- Invite config ---");
|
||||||
log.info("Expiration: {}", GsonUtil.get().toJson(getExpiration()));
|
log.info("Expiration: {}", GsonUtil.get().toJson(getExpiration()));
|
||||||
log.info("Resolution: {}", GsonUtil.get().toJson(getResolution()));
|
log.info("Resolution: {}", GsonUtil.get().toJson(getResolution()));
|
||||||
log.info("Policies: {}", GsonUtil.get().toJson(getPolicy()));
|
log.info("Policies: {}", GsonUtil.get().toJson(getPolicy()));
|
||||||
|
log.info("Print full display name on invitation: {}", isFullDisplayName());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public enum PeriodDimension {
|
||||||
|
|
||||||
|
minutes,
|
||||||
|
|
||||||
|
seconds
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
60
src/main/java/io/kamax/mxisd/config/LoggingConfig.java
Normal file
60
src/main/java/io/kamax/mxisd/config/LoggingConfig.java
Normal file
@@ -0,0 +1,60 @@
|
|||||||
|
package io.kamax.mxisd.config;
|
||||||
|
|
||||||
|
import org.apache.commons.lang3.StringUtils;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
public class LoggingConfig {
|
||||||
|
|
||||||
|
private static final Logger LOGGER = LoggerFactory.getLogger("App");
|
||||||
|
|
||||||
|
private String root;
|
||||||
|
private String app;
|
||||||
|
private boolean requests = false;
|
||||||
|
|
||||||
|
public String getRoot() {
|
||||||
|
return root;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setRoot(String root) {
|
||||||
|
this.root = root;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getApp() {
|
||||||
|
return app;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setApp(String app) {
|
||||||
|
this.app = app;
|
||||||
|
}
|
||||||
|
|
||||||
|
public boolean isRequests() {
|
||||||
|
return requests;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setRequests(boolean requests) {
|
||||||
|
this.requests = requests;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void build() {
|
||||||
|
LOGGER.info("Logging config:");
|
||||||
|
if (StringUtils.isNotBlank(getRoot())) {
|
||||||
|
LOGGER.info(" Default log level: {}", getRoot());
|
||||||
|
System.setProperty("org.slf4j.simpleLogger.defaultLogLevel", getRoot());
|
||||||
|
}
|
||||||
|
|
||||||
|
String appLevel = System.getProperty("org.slf4j.simpleLogger.log.io.kamax.mxisd");
|
||||||
|
if (StringUtils.isNotBlank(appLevel)) {
|
||||||
|
LOGGER.info(" Logging level set by environment: {}", appLevel);
|
||||||
|
} else if (StringUtils.isNotBlank(getApp())) {
|
||||||
|
System.setProperty("org.slf4j.simpleLogger.log.io.kamax.mxisd", getApp());
|
||||||
|
LOGGER.info(" Logging level set by the configuration: {}", getApp());
|
||||||
|
} else {
|
||||||
|
LOGGER.info(" Logging level hasn't set, use default");
|
||||||
|
}
|
||||||
|
LOGGER.info(" Log requests: {}", isRequests());
|
||||||
|
if (isRequests()) {
|
||||||
|
LOGGER.warn(" Request dumping enabled, use this only to debug purposes, don't use it in the production.");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@@ -117,6 +117,8 @@ public class MxisdConfig {
|
|||||||
private WordpressConfig wordpress = new WordpressConfig();
|
private WordpressConfig wordpress = new WordpressConfig();
|
||||||
private PolicyConfig policy = new PolicyConfig();
|
private PolicyConfig policy = new PolicyConfig();
|
||||||
private HashingConfig hashing = new HashingConfig();
|
private HashingConfig hashing = new HashingConfig();
|
||||||
|
private LoggingConfig logging = new LoggingConfig();
|
||||||
|
private InternalAPIConfig internal = new InternalAPIConfig();
|
||||||
|
|
||||||
public AppServiceConfig getAppsvc() {
|
public AppServiceConfig getAppsvc() {
|
||||||
return appsvc;
|
return appsvc;
|
||||||
@@ -342,6 +344,14 @@ public class MxisdConfig {
|
|||||||
this.hashing = hashing;
|
this.hashing = hashing;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public LoggingConfig getLogging() {
|
||||||
|
return logging;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setLogging(LoggingConfig logging) {
|
||||||
|
this.logging = logging;
|
||||||
|
}
|
||||||
|
|
||||||
public MxisdConfig inMemory() {
|
public MxisdConfig inMemory() {
|
||||||
getKey().setPath(":memory:");
|
getKey().setPath(":memory:");
|
||||||
getStorage().getProvider().getSqlite().setDatabase(":memory:");
|
getStorage().getProvider().getSqlite().setDatabase(":memory:");
|
||||||
@@ -349,7 +359,17 @@ public class MxisdConfig {
|
|||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public InternalAPIConfig getInternal() {
|
||||||
|
return internal;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setInternal(InternalAPIConfig internal) {
|
||||||
|
this.internal = internal;
|
||||||
|
}
|
||||||
|
|
||||||
public MxisdConfig build() {
|
public MxisdConfig build() {
|
||||||
|
getLogging().build();
|
||||||
|
|
||||||
if (StringUtils.isBlank(getServer().getName())) {
|
if (StringUtils.isBlank(getServer().getName())) {
|
||||||
getServer().setName(getMatrix().getDomain());
|
getServer().setName(getMatrix().getDomain());
|
||||||
log.debug("server.name is empty, using matrix.domain");
|
log.debug("server.name is empty, using matrix.domain");
|
||||||
@@ -359,6 +379,7 @@ public class MxisdConfig {
|
|||||||
getAuth().build();
|
getAuth().build();
|
||||||
getAccountConfig().build();
|
getAccountConfig().build();
|
||||||
getDirectory().build();
|
getDirectory().build();
|
||||||
|
getDns().build();
|
||||||
getExec().build();
|
getExec().build();
|
||||||
getFirebase().build();
|
getFirebase().build();
|
||||||
getForward().build();
|
getForward().build();
|
||||||
@@ -381,7 +402,8 @@ public class MxisdConfig {
|
|||||||
getView().build();
|
getView().build();
|
||||||
getWordpress().build();
|
getWordpress().build();
|
||||||
getPolicy().build();
|
getPolicy().build();
|
||||||
getHashing().build();
|
getHashing().build(getMatrix());
|
||||||
|
getInternal().build();
|
||||||
|
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
105
src/main/java/io/kamax/mxisd/config/PostgresqlStorageConfig.java
Normal file
105
src/main/java/io/kamax/mxisd/config/PostgresqlStorageConfig.java
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
/*
|
||||||
|
* mxisd - Matrix Identity Server Daemon
|
||||||
|
* Copyright (C) 2017 Kamax Sarl
|
||||||
|
*
|
||||||
|
* https://www.kamax.io/
|
||||||
|
*
|
||||||
|
* This program is free software: you can redistribute it and/or modify
|
||||||
|
* it under the terms of the GNU Affero General Public License as
|
||||||
|
* published by the Free Software Foundation, either version 3 of the
|
||||||
|
* License, or (at your option) any later version.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope that it will be useful,
|
||||||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
* GNU Affero General Public License for more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU Affero General Public License
|
||||||
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package io.kamax.mxisd.config;
|
||||||
|
|
||||||
|
public class PostgresqlStorageConfig implements DatabaseStorageConfig {
|
||||||
|
|
||||||
|
private String database;
|
||||||
|
|
||||||
|
private String username;
|
||||||
|
|
||||||
|
private String password;
|
||||||
|
|
||||||
|
private boolean pool;
|
||||||
|
|
||||||
|
private int maxConnectionsFree = 1;
|
||||||
|
|
||||||
|
private long maxConnectionAgeMillis = 60 * 60 * 1000;
|
||||||
|
|
||||||
|
private long checkConnectionsEveryMillis = 30 * 1000;
|
||||||
|
|
||||||
|
private boolean testBeforeGetFromPool = false;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getDatabase() {
|
||||||
|
return database;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setDatabase(String database) {
|
||||||
|
this.database = database;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getUsername() {
|
||||||
|
return username;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setUsername(String username) {
|
||||||
|
this.username = username;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getPassword() {
|
||||||
|
return password;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setPassword(String password) {
|
||||||
|
this.password = password;
|
||||||
|
}
|
||||||
|
|
||||||
|
public boolean isPool() {
|
||||||
|
return pool;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setPool(boolean pool) {
|
||||||
|
this.pool = pool;
|
||||||
|
}
|
||||||
|
|
||||||
|
public int getMaxConnectionsFree() {
|
||||||
|
return maxConnectionsFree;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setMaxConnectionsFree(int maxConnectionsFree) {
|
||||||
|
this.maxConnectionsFree = maxConnectionsFree;
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getMaxConnectionAgeMillis() {
|
||||||
|
return maxConnectionAgeMillis;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setMaxConnectionAgeMillis(long maxConnectionAgeMillis) {
|
||||||
|
this.maxConnectionAgeMillis = maxConnectionAgeMillis;
|
||||||
|
}
|
||||||
|
|
||||||
|
public long getCheckConnectionsEveryMillis() {
|
||||||
|
return checkConnectionsEveryMillis;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setCheckConnectionsEveryMillis(long checkConnectionsEveryMillis) {
|
||||||
|
this.checkConnectionsEveryMillis = checkConnectionsEveryMillis;
|
||||||
|
}
|
||||||
|
|
||||||
|
public boolean isTestBeforeGetFromPool() {
|
||||||
|
return testBeforeGetFromPool;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setTestBeforeGetFromPool(boolean testBeforeGetFromPool) {
|
||||||
|
this.testBeforeGetFromPool = testBeforeGetFromPool;
|
||||||
|
}
|
||||||
|
}
|
@@ -20,10 +20,11 @@
|
|||||||
|
|
||||||
package io.kamax.mxisd.config;
|
package io.kamax.mxisd.config;
|
||||||
|
|
||||||
public class SQLiteStorageConfig {
|
public class SQLiteStorageConfig implements DatabaseStorageConfig {
|
||||||
|
|
||||||
private String database;
|
private String database;
|
||||||
|
|
||||||
|
@Override
|
||||||
public String getDatabase() {
|
public String getDatabase() {
|
||||||
return database;
|
return database;
|
||||||
}
|
}
|
||||||
|
@@ -34,6 +34,7 @@ public class ServerConfig {
|
|||||||
private String name;
|
private String name;
|
||||||
private int port = 8090;
|
private int port = 8090;
|
||||||
private String publicUrl;
|
private String publicUrl;
|
||||||
|
private String hostname;
|
||||||
|
|
||||||
public String getName() {
|
public String getName() {
|
||||||
return name;
|
return name;
|
||||||
@@ -59,6 +60,14 @@ public class ServerConfig {
|
|||||||
this.publicUrl = publicUrl;
|
this.publicUrl = publicUrl;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public String getHostname() {
|
||||||
|
return hostname;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setHostname(String hostname) {
|
||||||
|
this.hostname = hostname;
|
||||||
|
}
|
||||||
|
|
||||||
public void build() {
|
public void build() {
|
||||||
log.info("--- Server config ---");
|
log.info("--- Server config ---");
|
||||||
|
|
||||||
@@ -75,8 +84,13 @@ public class ServerConfig {
|
|||||||
log.warn("Public URL is not valid: {}", StringUtils.defaultIfBlank(e.getMessage(), "<no reason provided>"));
|
log.warn("Public URL is not valid: {}", StringUtils.defaultIfBlank(e.getMessage(), "<no reason provided>"));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (StringUtils.isBlank(getHostname())) {
|
||||||
|
setHostname("0.0.0.0");
|
||||||
|
}
|
||||||
|
|
||||||
log.info("Name: {}", getName());
|
log.info("Name: {}", getName());
|
||||||
log.info("Port: {}", getPort());
|
log.info("Port: {}", getPort());
|
||||||
log.info("Public URL: {}", getPublicUrl());
|
log.info("Public URL: {}", getPublicUrl());
|
||||||
|
log.info("Hostname: {}", getHostname());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -21,14 +21,21 @@
|
|||||||
package io.kamax.mxisd.config;
|
package io.kamax.mxisd.config;
|
||||||
|
|
||||||
import io.kamax.mxisd.exception.ConfigurationException;
|
import io.kamax.mxisd.exception.ConfigurationException;
|
||||||
import org.apache.commons.lang.StringUtils;
|
|
||||||
|
|
||||||
public class StorageConfig {
|
public class StorageConfig {
|
||||||
|
|
||||||
|
public enum BackendEnum {
|
||||||
|
sqlite,
|
||||||
|
|
||||||
|
postgresql
|
||||||
|
}
|
||||||
|
|
||||||
public static class Provider {
|
public static class Provider {
|
||||||
|
|
||||||
private SQLiteStorageConfig sqlite = new SQLiteStorageConfig();
|
private SQLiteStorageConfig sqlite = new SQLiteStorageConfig();
|
||||||
|
|
||||||
|
private PostgresqlStorageConfig postgresql = new PostgresqlStorageConfig();
|
||||||
|
|
||||||
public SQLiteStorageConfig getSqlite() {
|
public SQLiteStorageConfig getSqlite() {
|
||||||
return sqlite;
|
return sqlite;
|
||||||
}
|
}
|
||||||
@@ -37,16 +44,23 @@ public class StorageConfig {
|
|||||||
this.sqlite = sqlite;
|
this.sqlite = sqlite;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public PostgresqlStorageConfig getPostgresql() {
|
||||||
|
return postgresql;
|
||||||
}
|
}
|
||||||
|
|
||||||
private String backend = "sqlite";
|
public void setPostgresql(PostgresqlStorageConfig postgresql) {
|
||||||
|
this.postgresql = postgresql;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private BackendEnum backend = BackendEnum.sqlite; // or postgresql
|
||||||
private Provider provider = new Provider();
|
private Provider provider = new Provider();
|
||||||
|
|
||||||
public String getBackend() {
|
public BackendEnum getBackend() {
|
||||||
return backend;
|
return backend;
|
||||||
}
|
}
|
||||||
|
|
||||||
public void setBackend(String backend) {
|
public void setBackend(BackendEnum backend) {
|
||||||
this.backend = backend;
|
this.backend = backend;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -59,7 +73,7 @@ public class StorageConfig {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public void build() {
|
public void build() {
|
||||||
if (StringUtils.isBlank(getBackend())) {
|
if (getBackend() == null) {
|
||||||
throw new ConfigurationException("storage.backend");
|
throw new ConfigurationException("storage.backend");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -76,4 +76,14 @@ public class YamlConfigLoader {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static void dumpConfig(MxisdConfig cfg) {
|
||||||
|
Representer rep = new Representer();
|
||||||
|
rep.getPropertyUtils().setBeanAccess(BeanAccess.FIELD);
|
||||||
|
rep.getPropertyUtils().setAllowReadOnlyProperties(true);
|
||||||
|
rep.getPropertyUtils().setSkipMissingProperties(true);
|
||||||
|
|
||||||
|
Yaml yaml = new Yaml(new Constructor(MxisdConfig.class), rep);
|
||||||
|
String dump = yaml.dump(cfg);
|
||||||
|
log.info("Full configuration:\n{}", dump);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@@ -291,6 +291,9 @@ public abstract class LdapConfig {
|
|||||||
private boolean enabled;
|
private boolean enabled;
|
||||||
private String filter;
|
private String filter;
|
||||||
|
|
||||||
|
private boolean activeDirectory;
|
||||||
|
private String defaultDomain;
|
||||||
|
|
||||||
private Connection connection = new Connection();
|
private Connection connection = new Connection();
|
||||||
private Attribute attribute = new Attribute();
|
private Attribute attribute = new Attribute();
|
||||||
private Auth auth = new Auth();
|
private Auth auth = new Auth();
|
||||||
@@ -316,6 +319,22 @@ public abstract class LdapConfig {
|
|||||||
this.filter = filter;
|
this.filter = filter;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public boolean isActiveDirectory() {
|
||||||
|
return activeDirectory;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setActiveDirectory(boolean activeDirectory) {
|
||||||
|
this.activeDirectory = activeDirectory;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getDefaultDomain() {
|
||||||
|
return defaultDomain;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setDefaultDomain(String defaultDomain) {
|
||||||
|
this.defaultDomain = defaultDomain;
|
||||||
|
}
|
||||||
|
|
||||||
public Connection getConnection() {
|
public Connection getConnection() {
|
||||||
return connection;
|
return connection;
|
||||||
}
|
}
|
||||||
@@ -407,6 +426,15 @@ public abstract class LdapConfig {
|
|||||||
throw new ConfigurationException("ldap.identity.token");
|
throw new ConfigurationException("ldap.identity.token");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if(isActiveDirectory()) {
|
||||||
|
if(!StringUtils.equals(LdapBackend.UID, uidType)) {
|
||||||
|
throw new IllegalArgumentException(String.format(
|
||||||
|
"Attribute UID type should be set to %s in Active Directory mode",
|
||||||
|
LdapBackend.UID
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Build queries
|
// Build queries
|
||||||
attribute.getThreepid().forEach((k, v) -> {
|
attribute.getThreepid().forEach((k, v) -> {
|
||||||
if (StringUtils.isBlank(identity.getMedium().get(k))) {
|
if (StringUtils.isBlank(identity.getMedium().get(k))) {
|
||||||
|
@@ -45,4 +45,21 @@ public class MemoryThreePid implements _ThreePid {
|
|||||||
this.address = address;
|
this.address = address;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean equals(Object o) {
|
||||||
|
if (this == o) return true;
|
||||||
|
if (o == null || getClass() != o.getClass()) return false;
|
||||||
|
|
||||||
|
MemoryThreePid threePid = (MemoryThreePid) o;
|
||||||
|
|
||||||
|
if (!medium.equals(threePid.medium)) return false;
|
||||||
|
return address.equals(threePid.address);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int hashCode() {
|
||||||
|
int result = medium.hashCode();
|
||||||
|
result = 31 * result + address.hashCode();
|
||||||
|
return result;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@@ -125,7 +125,7 @@ public abstract class SqlConfig {
|
|||||||
}
|
}
|
||||||
|
|
||||||
public static class Lookup {
|
public static class Lookup {
|
||||||
private String query = "SELECT user_id AS mxid, medium, address from user_threepids";
|
private String query = "SELECT user_id AS mxid, medium, address from user_threepid_id_server";
|
||||||
|
|
||||||
public String getQuery() {
|
public String getQuery() {
|
||||||
return query;
|
return query;
|
||||||
@@ -140,7 +140,7 @@ public abstract class SqlConfig {
|
|||||||
|
|
||||||
private Boolean enabled;
|
private Boolean enabled;
|
||||||
private String type = "mxid";
|
private String type = "mxid";
|
||||||
private String query = "SELECT user_id AS uid FROM user_threepids WHERE medium = ? AND address = ?";
|
private String query = "SELECT user_id AS uid FROM user_threepid_id_server WHERE medium = ? AND address = ?";
|
||||||
private Map<String, String> medium = new HashMap<>();
|
private Map<String, String> medium = new HashMap<>();
|
||||||
|
|
||||||
public Boolean isEnabled() {
|
public Boolean isEnabled() {
|
||||||
|
@@ -24,9 +24,23 @@ import io.kamax.mxisd.UserIdType;
|
|||||||
import io.kamax.mxisd.backend.sql.synapse.SynapseQueries;
|
import io.kamax.mxisd.backend.sql.synapse.SynapseQueries;
|
||||||
import io.kamax.mxisd.config.sql.SqlConfig;
|
import io.kamax.mxisd.config.sql.SqlConfig;
|
||||||
import org.apache.commons.lang.StringUtils;
|
import org.apache.commons.lang.StringUtils;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
public class SynapseSqlProviderConfig extends SqlConfig {
|
public class SynapseSqlProviderConfig extends SqlConfig {
|
||||||
|
|
||||||
|
private transient final Logger log = LoggerFactory.getLogger(SynapseSqlProviderConfig.class);
|
||||||
|
|
||||||
|
private boolean legacyRoomNames = false;
|
||||||
|
|
||||||
|
public boolean isLegacyRoomNames() {
|
||||||
|
return legacyRoomNames;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setLegacyRoomNames(boolean legacyRoomNames) {
|
||||||
|
this.legacyRoomNames = legacyRoomNames;
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected String getProviderName() {
|
protected String getProviderName() {
|
||||||
return "Synapse SQL";
|
return "Synapse SQL";
|
||||||
@@ -42,7 +56,7 @@ public class SynapseSqlProviderConfig extends SqlConfig {
|
|||||||
|
|
||||||
if (getIdentity().isEnabled() && StringUtils.isBlank(getIdentity().getType())) {
|
if (getIdentity().isEnabled() && StringUtils.isBlank(getIdentity().getType())) {
|
||||||
getIdentity().setType("mxid");
|
getIdentity().setType("mxid");
|
||||||
getIdentity().setQuery("SELECT user_id AS uid FROM user_threepids WHERE medium = ? AND address = ?");
|
getIdentity().setQuery("SELECT user_id AS uid FROM user_threepid_id_server WHERE medium = ? AND address = ?");
|
||||||
}
|
}
|
||||||
|
|
||||||
if (getProfile().isEnabled()) {
|
if (getProfile().isEnabled()) {
|
||||||
@@ -65,4 +79,12 @@ public class SynapseSqlProviderConfig extends SqlConfig {
|
|||||||
printConfig();
|
printConfig();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
protected void printConfig() {
|
||||||
|
super.printConfig();
|
||||||
|
|
||||||
|
if (isEnabled()) {
|
||||||
|
log.info("Use legacy room name query: {}", isLegacyRoomNames());
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@@ -23,5 +23,6 @@ package io.kamax.mxisd.exception;
|
|||||||
public class InvalidParamException extends RuntimeException {
|
public class InvalidParamException extends RuntimeException {
|
||||||
|
|
||||||
public InvalidParamException() {
|
public InvalidParamException() {
|
||||||
|
super("The chosen hash algorithm is invalid or disallowed");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -23,5 +23,6 @@ package io.kamax.mxisd.exception;
|
|||||||
public class InvalidPepperException extends RuntimeException {
|
public class InvalidPepperException extends RuntimeException {
|
||||||
|
|
||||||
public InvalidPepperException() {
|
public InvalidPepperException() {
|
||||||
|
super("The provided pepper is invalid or expired");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -0,0 +1,28 @@
|
|||||||
|
/*
|
||||||
|
* ma1sd - Matrix Identity Server Daemon
|
||||||
|
* Copyright (C) 2020 Anatoliy SAblin
|
||||||
|
*
|
||||||
|
* https://www.github.com/ma1uta/ma1sd/
|
||||||
|
*
|
||||||
|
* This program is free software: you can redistribute it and/or modify
|
||||||
|
* it under the terms of the GNU Affero General Public License as
|
||||||
|
* published by the Free Software Foundation, either version 3 of the
|
||||||
|
* License, or (at your option) any later version.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope that it will be useful,
|
||||||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
* GNU Affero General Public License for more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU Affero General Public License
|
||||||
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package io.kamax.mxisd.exception;
|
||||||
|
|
||||||
|
public class TermsNotSignedException extends RuntimeException {
|
||||||
|
|
||||||
|
public TermsNotSignedException() {
|
||||||
|
super("Please accept our updated terms of service before continuing");
|
||||||
|
}
|
||||||
|
}
|
@@ -1,6 +1,9 @@
|
|||||||
package io.kamax.mxisd.hash;
|
package io.kamax.mxisd.hash;
|
||||||
|
|
||||||
import io.kamax.mxisd.config.HashingConfig;
|
import io.kamax.mxisd.config.HashingConfig;
|
||||||
|
import io.kamax.mxisd.hash.engine.Engine;
|
||||||
|
import io.kamax.mxisd.hash.engine.HashEngine;
|
||||||
|
import io.kamax.mxisd.hash.engine.NoneEngine;
|
||||||
import io.kamax.mxisd.hash.rotation.HashRotationStrategy;
|
import io.kamax.mxisd.hash.rotation.HashRotationStrategy;
|
||||||
import io.kamax.mxisd.hash.rotation.NoOpRotationStrategy;
|
import io.kamax.mxisd.hash.rotation.NoOpRotationStrategy;
|
||||||
import io.kamax.mxisd.hash.rotation.RotationPerRequests;
|
import io.kamax.mxisd.hash.rotation.RotationPerRequests;
|
||||||
@@ -21,7 +24,7 @@ public class HashManager {
|
|||||||
|
|
||||||
private static final Logger LOGGER = LoggerFactory.getLogger(HashManager.class);
|
private static final Logger LOGGER = LoggerFactory.getLogger(HashManager.class);
|
||||||
|
|
||||||
private HashEngine hashEngine;
|
private Engine engine;
|
||||||
private HashRotationStrategy rotationStrategy;
|
private HashRotationStrategy rotationStrategy;
|
||||||
private HashStorage hashStorage;
|
private HashStorage hashStorage;
|
||||||
private HashingConfig config;
|
private HashingConfig config;
|
||||||
@@ -32,7 +35,7 @@ public class HashManager {
|
|||||||
this.config = config;
|
this.config = config;
|
||||||
this.storage = storage;
|
this.storage = storage;
|
||||||
initStorage();
|
initStorage();
|
||||||
hashEngine = new HashEngine(providers, getHashStorage(), config);
|
engine = config.isEnabled() ? new HashEngine(providers, getHashStorage(), config) : new NoneEngine();
|
||||||
initRotationStrategy();
|
initRotationStrategy();
|
||||||
configured.set(true);
|
configured.set(true);
|
||||||
}
|
}
|
||||||
@@ -73,8 +76,8 @@ public class HashManager {
|
|||||||
this.rotationStrategy.register(getHashEngine());
|
this.rotationStrategy.register(getHashEngine());
|
||||||
}
|
}
|
||||||
|
|
||||||
public HashEngine getHashEngine() {
|
public Engine getHashEngine() {
|
||||||
return hashEngine;
|
return engine;
|
||||||
}
|
}
|
||||||
|
|
||||||
public HashRotationStrategy getRotationStrategy() {
|
public HashRotationStrategy getRotationStrategy() {
|
||||||
|
7
src/main/java/io/kamax/mxisd/hash/engine/Engine.java
Normal file
7
src/main/java/io/kamax/mxisd/hash/engine/Engine.java
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
package io.kamax.mxisd.hash.engine;
|
||||||
|
|
||||||
|
public interface Engine {
|
||||||
|
void updateHashes();
|
||||||
|
|
||||||
|
String getPepper();
|
||||||
|
}
|
@@ -1,4 +1,4 @@
|
|||||||
package io.kamax.mxisd.hash;
|
package io.kamax.mxisd.hash.engine;
|
||||||
|
|
||||||
import io.kamax.mxisd.config.HashingConfig;
|
import io.kamax.mxisd.config.HashingConfig;
|
||||||
import io.kamax.mxisd.hash.storage.HashStorage;
|
import io.kamax.mxisd.hash.storage.HashStorage;
|
||||||
@@ -12,7 +12,7 @@ import org.slf4j.LoggerFactory;
|
|||||||
import java.util.Base64;
|
import java.util.Base64;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
public class HashEngine {
|
public class HashEngine implements Engine {
|
||||||
|
|
||||||
private static final Logger LOGGER = LoggerFactory.getLogger(HashEngine.class);
|
private static final Logger LOGGER = LoggerFactory.getLogger(HashEngine.class);
|
||||||
|
|
||||||
@@ -28,6 +28,7 @@ public class HashEngine {
|
|||||||
this.config = config;
|
this.config = config;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public void updateHashes() {
|
public void updateHashes() {
|
||||||
LOGGER.info("Start update hashes.");
|
LOGGER.info("Start update hashes.");
|
||||||
synchronized (hashStorage) {
|
synchronized (hashStorage) {
|
||||||
@@ -48,6 +49,7 @@ public class HashEngine {
|
|||||||
LOGGER.info("Finish update hashes.");
|
LOGGER.info("Finish update hashes.");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public String getPepper() {
|
public String getPepper() {
|
||||||
synchronized (hashStorage) {
|
synchronized (hashStorage) {
|
||||||
return pepper;
|
return pepper;
|
19
src/main/java/io/kamax/mxisd/hash/engine/NoneEngine.java
Normal file
19
src/main/java/io/kamax/mxisd/hash/engine/NoneEngine.java
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
package io.kamax.mxisd.hash.engine;
|
||||||
|
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
public class NoneEngine implements Engine {
|
||||||
|
|
||||||
|
private static final Logger LOGGER = LoggerFactory.getLogger(NoneEngine.class);
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void updateHashes() {
|
||||||
|
LOGGER.info("Nothing to update.");
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String getPepper() {
|
||||||
|
return "";
|
||||||
|
}
|
||||||
|
}
|
@@ -1,12 +1,12 @@
|
|||||||
package io.kamax.mxisd.hash.rotation;
|
package io.kamax.mxisd.hash.rotation;
|
||||||
|
|
||||||
import io.kamax.mxisd.hash.HashEngine;
|
import io.kamax.mxisd.hash.engine.Engine;
|
||||||
|
|
||||||
public interface HashRotationStrategy {
|
public interface HashRotationStrategy {
|
||||||
|
|
||||||
void register(HashEngine hashEngine);
|
void register(Engine engine);
|
||||||
|
|
||||||
HashEngine getHashEngine();
|
Engine getHashEngine();
|
||||||
|
|
||||||
void newRequest();
|
void newRequest();
|
||||||
|
|
||||||
|
@@ -1,19 +1,19 @@
|
|||||||
package io.kamax.mxisd.hash.rotation;
|
package io.kamax.mxisd.hash.rotation;
|
||||||
|
|
||||||
import io.kamax.mxisd.hash.HashEngine;
|
import io.kamax.mxisd.hash.engine.Engine;
|
||||||
|
|
||||||
public class NoOpRotationStrategy implements HashRotationStrategy {
|
public class NoOpRotationStrategy implements HashRotationStrategy {
|
||||||
|
|
||||||
private HashEngine hashEngine;
|
private Engine engine;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void register(HashEngine hashEngine) {
|
public void register(Engine engine) {
|
||||||
this.hashEngine = hashEngine;
|
this.engine = engine;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public HashEngine getHashEngine() {
|
public Engine getHashEngine() {
|
||||||
return hashEngine;
|
return engine;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@@ -1,12 +1,12 @@
|
|||||||
package io.kamax.mxisd.hash.rotation;
|
package io.kamax.mxisd.hash.rotation;
|
||||||
|
|
||||||
import io.kamax.mxisd.hash.HashEngine;
|
import io.kamax.mxisd.hash.engine.Engine;
|
||||||
|
|
||||||
import java.util.concurrent.atomic.AtomicInteger;
|
import java.util.concurrent.atomic.AtomicInteger;
|
||||||
|
|
||||||
public class RotationPerRequests implements HashRotationStrategy {
|
public class RotationPerRequests implements HashRotationStrategy {
|
||||||
|
|
||||||
private HashEngine hashEngine;
|
private Engine engine;
|
||||||
private final AtomicInteger counter = new AtomicInteger(0);
|
private final AtomicInteger counter = new AtomicInteger(0);
|
||||||
private final int barrier;
|
private final int barrier;
|
||||||
|
|
||||||
@@ -15,14 +15,14 @@ public class RotationPerRequests implements HashRotationStrategy {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void register(HashEngine hashEngine) {
|
public void register(Engine engine) {
|
||||||
this.hashEngine = hashEngine;
|
this.engine = engine;
|
||||||
trigger();
|
trigger();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public HashEngine getHashEngine() {
|
public Engine getHashEngine() {
|
||||||
return hashEngine;
|
return engine;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@@ -1,6 +1,6 @@
|
|||||||
package io.kamax.mxisd.hash.rotation;
|
package io.kamax.mxisd.hash.rotation;
|
||||||
|
|
||||||
import io.kamax.mxisd.hash.HashEngine;
|
import io.kamax.mxisd.hash.engine.Engine;
|
||||||
|
|
||||||
import java.util.concurrent.Executors;
|
import java.util.concurrent.Executors;
|
||||||
import java.util.concurrent.ScheduledExecutorService;
|
import java.util.concurrent.ScheduledExecutorService;
|
||||||
@@ -9,7 +9,7 @@ import java.util.concurrent.TimeUnit;
|
|||||||
public class TimeBasedRotation implements HashRotationStrategy {
|
public class TimeBasedRotation implements HashRotationStrategy {
|
||||||
|
|
||||||
private final long delay;
|
private final long delay;
|
||||||
private HashEngine hashEngine;
|
private Engine engine;
|
||||||
private final ScheduledExecutorService executorService = Executors.newSingleThreadScheduledExecutor();
|
private final ScheduledExecutorService executorService = Executors.newSingleThreadScheduledExecutor();
|
||||||
|
|
||||||
public TimeBasedRotation(long delay) {
|
public TimeBasedRotation(long delay) {
|
||||||
@@ -17,15 +17,15 @@ public class TimeBasedRotation implements HashRotationStrategy {
|
|||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void register(HashEngine hashEngine) {
|
public void register(Engine engine) {
|
||||||
this.hashEngine = hashEngine;
|
this.engine = engine;
|
||||||
Runtime.getRuntime().addShutdownHook(new Thread(executorService::shutdown));
|
Runtime.getRuntime().addShutdownHook(new Thread(executorService::shutdown));
|
||||||
executorService.scheduleWithFixedDelay(this::trigger, 0, delay, TimeUnit.SECONDS);
|
executorService.scheduleWithFixedDelay(this::trigger, 0, delay, TimeUnit.SECONDS);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public HashEngine getHashEngine() {
|
public Engine getHashEngine() {
|
||||||
return hashEngine;
|
return engine;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@@ -20,11 +20,19 @@
|
|||||||
|
|
||||||
package io.kamax.mxisd.http;
|
package io.kamax.mxisd.http;
|
||||||
|
|
||||||
|
import static io.kamax.mxisd.util.RestClientUtils.urlEncode;
|
||||||
|
|
||||||
public class IsAPIv1 {
|
public class IsAPIv1 {
|
||||||
|
|
||||||
public static final String Base = "/_matrix/identity/api/v1";
|
public static final String Base = "/_matrix/identity/api/v1";
|
||||||
|
|
||||||
public static String getValidate(String medium, String sid, String secret, String token) {
|
public static String getValidate(String medium, String sid, String secret, String token) {
|
||||||
return String.format("%s/validate/%s/submitToken?sid=%s&client_secret=%s&token=%s", Base, medium, sid, secret, token);
|
return String.format("%s/validate/%s/submitToken?sid=%s&client_secret=%s&token=%s",
|
||||||
|
Base,
|
||||||
|
medium,
|
||||||
|
urlEncode(sid),
|
||||||
|
urlEncode(secret),
|
||||||
|
urlEncode(token)
|
||||||
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -0,0 +1,5 @@
|
|||||||
|
package io.kamax.mxisd.http.undertow.conduit;
|
||||||
|
|
||||||
|
public interface ConduitWithDump {
|
||||||
|
String dump();
|
||||||
|
}
|
@@ -0,0 +1,107 @@
|
|||||||
|
/*
|
||||||
|
* JBoss, Home of Professional Open Source.
|
||||||
|
* Copyright 2014 Red Hat, Inc., and individual contributors
|
||||||
|
* as indicated by the @author tags.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package io.kamax.mxisd.http.undertow.conduit;
|
||||||
|
|
||||||
|
import org.xnio.IoUtils;
|
||||||
|
import org.xnio.channels.StreamSourceChannel;
|
||||||
|
import org.xnio.conduits.AbstractStreamSinkConduit;
|
||||||
|
import org.xnio.conduits.ConduitWritableByteChannel;
|
||||||
|
import org.xnio.conduits.Conduits;
|
||||||
|
import org.xnio.conduits.StreamSinkConduit;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.nio.channels.FileChannel;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.concurrent.CopyOnWriteArrayList;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Conduit that saves all the data that is written through it and can dump it to the console
|
||||||
|
* <p>
|
||||||
|
* Obviously this should not be used in production.
|
||||||
|
*
|
||||||
|
* @author Stuart Douglas
|
||||||
|
*/
|
||||||
|
public class DebuggingStreamSinkConduit extends AbstractStreamSinkConduit<StreamSinkConduit> implements ConduitWithDump {
|
||||||
|
|
||||||
|
private final List<byte[]> data = new CopyOnWriteArrayList<>();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Construct a new instance.
|
||||||
|
*
|
||||||
|
* @param next the delegate conduit to set
|
||||||
|
*/
|
||||||
|
public DebuggingStreamSinkConduit(StreamSinkConduit next) {
|
||||||
|
super(next);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int write(ByteBuffer src) throws IOException {
|
||||||
|
int pos = src.position();
|
||||||
|
int res = super.write(src);
|
||||||
|
if (res > 0) {
|
||||||
|
byte[] d = new byte[res];
|
||||||
|
for (int i = 0; i < res; ++i) {
|
||||||
|
d[i] = src.get(i + pos);
|
||||||
|
}
|
||||||
|
data.add(d);
|
||||||
|
}
|
||||||
|
return res;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long write(ByteBuffer[] dsts, int offs, int len) throws IOException {
|
||||||
|
for (int i = offs; i < len; ++i) {
|
||||||
|
if (dsts[i].hasRemaining()) {
|
||||||
|
return write(dsts[i]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long transferFrom(final FileChannel src, final long position, final long count) throws IOException {
|
||||||
|
return src.transferTo(position, count, new ConduitWritableByteChannel(this));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long transferFrom(final StreamSourceChannel source, final long count, final ByteBuffer throughBuffer) throws IOException {
|
||||||
|
return IoUtils.transfer(source, count, throughBuffer, new ConduitWritableByteChannel(this));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int writeFinal(ByteBuffer src) throws IOException {
|
||||||
|
return Conduits.writeFinalBasic(this, src);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long writeFinal(ByteBuffer[] srcs, int offset, int length) throws IOException {
|
||||||
|
return Conduits.writeFinalBasic(this, srcs, offset, length);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String dump() {
|
||||||
|
StringBuilder sb = new StringBuilder();
|
||||||
|
for (byte[] datum : data) {
|
||||||
|
sb.append(new String(datum, StandardCharsets.UTF_8));
|
||||||
|
}
|
||||||
|
return sb.toString();
|
||||||
|
}
|
||||||
|
}
|
@@ -0,0 +1,95 @@
|
|||||||
|
/*
|
||||||
|
* JBoss, Home of Professional Open Source.
|
||||||
|
* Copyright 2014 Red Hat, Inc., and individual contributors
|
||||||
|
* as indicated by the @author tags.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package io.kamax.mxisd.http.undertow.conduit;
|
||||||
|
|
||||||
|
import org.xnio.IoUtils;
|
||||||
|
import org.xnio.channels.StreamSinkChannel;
|
||||||
|
import org.xnio.conduits.AbstractStreamSourceConduit;
|
||||||
|
import org.xnio.conduits.ConduitReadableByteChannel;
|
||||||
|
import org.xnio.conduits.StreamSourceConduit;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.nio.channels.FileChannel;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.concurrent.CopyOnWriteArrayList;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Conduit that saves all the data that is written through it and can dump it to the console
|
||||||
|
* <p>
|
||||||
|
* Obviously this should not be used in production.
|
||||||
|
*
|
||||||
|
* @author Stuart Douglas
|
||||||
|
*/
|
||||||
|
public class DebuggingStreamSourceConduit extends AbstractStreamSourceConduit<StreamSourceConduit> implements ConduitWithDump {
|
||||||
|
|
||||||
|
private final List<byte[]> data = new CopyOnWriteArrayList<>();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Construct a new instance.
|
||||||
|
*
|
||||||
|
* @param next the delegate conduit to set
|
||||||
|
*/
|
||||||
|
public DebuggingStreamSourceConduit(StreamSourceConduit next) {
|
||||||
|
super(next);
|
||||||
|
}
|
||||||
|
|
||||||
|
public long transferTo(final long position, final long count, final FileChannel target) throws IOException {
|
||||||
|
return target.transferFrom(new ConduitReadableByteChannel(this), position, count);
|
||||||
|
}
|
||||||
|
|
||||||
|
public long transferTo(final long count, final ByteBuffer throughBuffer, final StreamSinkChannel target) throws IOException {
|
||||||
|
return IoUtils.transfer(new ConduitReadableByteChannel(this), count, throughBuffer, target);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int read(ByteBuffer dst) throws IOException {
|
||||||
|
int pos = dst.position();
|
||||||
|
int res = super.read(dst);
|
||||||
|
if (res > 0) {
|
||||||
|
byte[] d = new byte[res];
|
||||||
|
for (int i = 0; i < res; ++i) {
|
||||||
|
d[i] = dst.get(i + pos);
|
||||||
|
}
|
||||||
|
data.add(d);
|
||||||
|
}
|
||||||
|
return res;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long read(ByteBuffer[] dsts, int offs, int len) throws IOException {
|
||||||
|
for (int i = offs; i < len; ++i) {
|
||||||
|
if (dsts[i].hasRemaining()) {
|
||||||
|
return read(dsts[i]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String dump() {
|
||||||
|
|
||||||
|
StringBuilder sb = new StringBuilder();
|
||||||
|
for (byte[] datum : data) {
|
||||||
|
sb.append(new String(datum, StandardCharsets.UTF_8));
|
||||||
|
}
|
||||||
|
return sb.toString();
|
||||||
|
}
|
||||||
|
}
|
@@ -0,0 +1,23 @@
|
|||||||
|
package io.kamax.mxisd.http.undertow.conduit;
|
||||||
|
|
||||||
|
import io.undertow.server.ConduitWrapper;
|
||||||
|
import io.undertow.server.HttpServerExchange;
|
||||||
|
import io.undertow.util.ConduitFactory;
|
||||||
|
import org.xnio.conduits.Conduit;
|
||||||
|
|
||||||
|
public abstract class LazyConduitWrapper<T extends Conduit> implements ConduitWrapper<T> {
|
||||||
|
|
||||||
|
private T conduit = null;
|
||||||
|
|
||||||
|
protected abstract T create(ConduitFactory<T> factory, HttpServerExchange exchange);
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public T wrap(ConduitFactory<T> factory, HttpServerExchange exchange) {
|
||||||
|
conduit = create(factory, exchange);
|
||||||
|
return conduit;
|
||||||
|
}
|
||||||
|
|
||||||
|
public T get() {
|
||||||
|
return conduit;
|
||||||
|
}
|
||||||
|
}
|
@@ -33,6 +33,7 @@ import io.kamax.mxisd.util.RestClientUtils;
|
|||||||
import io.undertow.server.HttpHandler;
|
import io.undertow.server.HttpHandler;
|
||||||
import io.undertow.server.HttpServerExchange;
|
import io.undertow.server.HttpServerExchange;
|
||||||
import io.undertow.server.handlers.form.FormData;
|
import io.undertow.server.handlers.form.FormData;
|
||||||
|
import io.undertow.util.Headers;
|
||||||
import io.undertow.util.HttpString;
|
import io.undertow.util.HttpString;
|
||||||
import org.apache.commons.io.IOUtils;
|
import org.apache.commons.io.IOUtils;
|
||||||
import org.apache.commons.lang3.StringUtils;
|
import org.apache.commons.lang3.StringUtils;
|
||||||
@@ -189,7 +190,7 @@ public abstract class BasicHttpHandler implements HttpHandler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
protected void respond(HttpServerExchange ex, int status, String errCode, String error) {
|
protected void respond(HttpServerExchange ex, int status, String errCode, String error) {
|
||||||
respond(ex, status, buildErrorBody(ex, errCode, error));
|
respond(ex, status, buildErrorBody(ex, errCode, error != null ? error : "An error has occurred"));
|
||||||
}
|
}
|
||||||
|
|
||||||
protected void handleException(HttpServerExchange exchange, HttpMatrixException ex) {
|
protected void handleException(HttpServerExchange exchange, HttpMatrixException ex) {
|
||||||
@@ -203,26 +204,34 @@ public abstract class BasicHttpHandler implements HttpHandler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
protected void proxyPost(HttpServerExchange exchange, JsonObject body, CloseableHttpClient client, ClientDnsOverwrite dns) {
|
protected void proxyPost(HttpServerExchange exchange, JsonObject body, CloseableHttpClient client, ClientDnsOverwrite dns) {
|
||||||
|
proxyPost(exchange, body, client, dns, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
protected void proxyPost(HttpServerExchange exchange, JsonObject body, CloseableHttpClient client, ClientDnsOverwrite dns,
|
||||||
|
boolean defaultJsonResponse) {
|
||||||
String target = dns.transform(URI.create(exchange.getRequestURL())).toString();
|
String target = dns.transform(URI.create(exchange.getRequestURL())).toString();
|
||||||
log.info("Requesting remote: {}", target);
|
log.info("Requesting remote: {}", target);
|
||||||
HttpPost req = RestClientUtils.post(target, GsonUtil.get(), body);
|
HttpPost req = RestClientUtils.post(target, GsonUtil.get(), body);
|
||||||
|
|
||||||
exchange.getRequestHeaders().forEach(header -> {
|
exchange.getRequestHeaders().forEach(header -> header.forEach(v -> {
|
||||||
header.forEach(v -> {
|
|
||||||
String name = header.getHeaderName().toString();
|
String name = header.getHeaderName().toString();
|
||||||
if (!StringUtils.startsWithIgnoreCase(name, "content-")) {
|
if (!StringUtils.startsWithIgnoreCase(name, "content-")) {
|
||||||
req.addHeader(name, v);
|
req.addHeader(name, v);
|
||||||
}
|
}
|
||||||
});
|
}));
|
||||||
});
|
|
||||||
|
|
||||||
|
boolean missingJsonResponse = true;
|
||||||
try (CloseableHttpResponse res = client.execute(req)) {
|
try (CloseableHttpResponse res = client.execute(req)) {
|
||||||
exchange.setStatusCode(res.getStatusLine().getStatusCode());
|
exchange.setStatusCode(res.getStatusLine().getStatusCode());
|
||||||
for (Header h : res.getAllHeaders()) {
|
for (Header h : res.getAllHeaders()) {
|
||||||
for (HeaderElement el : h.getElements()) {
|
for (HeaderElement el : h.getElements()) {
|
||||||
|
missingJsonResponse = !Headers.CONTENT_TYPE_STRING.equalsIgnoreCase(h.getName());
|
||||||
exchange.getResponseHeaders().add(HttpString.tryFromString(h.getName()), el.getValue());
|
exchange.getResponseHeaders().add(HttpString.tryFromString(h.getName()), el.getValue());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if (defaultJsonResponse && missingJsonResponse) {
|
||||||
|
exchange.getRequestHeaders().add(Headers.CONTENT_TYPE, "application/json");
|
||||||
|
}
|
||||||
res.getEntity().writeTo(exchange.getOutputStream());
|
res.getEntity().writeTo(exchange.getOutputStream());
|
||||||
exchange.endExchange();
|
exchange.endExchange();
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
|
@@ -23,6 +23,7 @@ package io.kamax.mxisd.http.undertow.handler;
|
|||||||
import io.kamax.mxisd.auth.AccountManager;
|
import io.kamax.mxisd.auth.AccountManager;
|
||||||
import io.kamax.mxisd.config.PolicyConfig;
|
import io.kamax.mxisd.config.PolicyConfig;
|
||||||
import io.kamax.mxisd.exception.InvalidCredentialsException;
|
import io.kamax.mxisd.exception.InvalidCredentialsException;
|
||||||
|
import io.kamax.mxisd.exception.TermsNotSignedException;
|
||||||
import io.undertow.server.HttpHandler;
|
import io.undertow.server.HttpHandler;
|
||||||
import io.undertow.server.HttpServerExchange;
|
import io.undertow.server.HttpServerExchange;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
@@ -66,7 +67,7 @@ public class CheckTermsHandler extends BasicHttpHandler {
|
|||||||
|
|
||||||
if (!accountManager.isTermAccepted(token, policies)) {
|
if (!accountManager.isTermAccepted(token, policies)) {
|
||||||
log.error("Non accepting request from: {}", exchange.getHostAndPort());
|
log.error("Non accepting request from: {}", exchange.getHostAndPort());
|
||||||
throw new InvalidCredentialsException();
|
throw new TermsNotSignedException();
|
||||||
}
|
}
|
||||||
log.trace("Access granted");
|
log.trace("Access granted");
|
||||||
child.handleRequest(exchange);
|
child.handleRequest(exchange);
|
||||||
|
@@ -0,0 +1,186 @@
|
|||||||
|
/*
|
||||||
|
* JBoss, Home of Professional Open Source.
|
||||||
|
* Copyright 2014 Red Hat, Inc., and individual contributors
|
||||||
|
* as indicated by the @author tags.
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package io.kamax.mxisd.http.undertow.handler;
|
||||||
|
|
||||||
|
import io.kamax.mxisd.http.undertow.conduit.ConduitWithDump;
|
||||||
|
import io.kamax.mxisd.http.undertow.conduit.DebuggingStreamSinkConduit;
|
||||||
|
import io.kamax.mxisd.http.undertow.conduit.DebuggingStreamSourceConduit;
|
||||||
|
import io.kamax.mxisd.http.undertow.conduit.LazyConduitWrapper;
|
||||||
|
import io.undertow.security.api.SecurityContext;
|
||||||
|
import io.undertow.server.HttpHandler;
|
||||||
|
import io.undertow.server.HttpServerExchange;
|
||||||
|
import io.undertow.server.handlers.Cookie;
|
||||||
|
import io.undertow.util.ConduitFactory;
|
||||||
|
import io.undertow.util.HeaderValues;
|
||||||
|
import io.undertow.util.Headers;
|
||||||
|
import io.undertow.util.LocaleUtils;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
import org.xnio.conduits.StreamSinkConduit;
|
||||||
|
import org.xnio.conduits.StreamSourceConduit;
|
||||||
|
|
||||||
|
import java.util.Deque;
|
||||||
|
import java.util.Iterator;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Handler that dumps a exchange to a log.
|
||||||
|
*
|
||||||
|
* @author Stuart Douglas
|
||||||
|
*/
|
||||||
|
public class RequestDumpingHandler implements HttpHandler {
|
||||||
|
|
||||||
|
private static final Logger LOGGER = LoggerFactory.getLogger(RequestDumpingHandler.class);
|
||||||
|
|
||||||
|
private final HttpHandler next;
|
||||||
|
|
||||||
|
public RequestDumpingHandler(HttpHandler next) {
|
||||||
|
this.next = next;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void handleRequest(HttpServerExchange exchange) throws Exception {
|
||||||
|
LazyConduitWrapper<StreamSourceConduit> requestConduitWrapper = new LazyConduitWrapper<StreamSourceConduit>() {
|
||||||
|
@Override
|
||||||
|
protected StreamSourceConduit create(ConduitFactory<StreamSourceConduit> factory, HttpServerExchange exchange) {
|
||||||
|
return new DebuggingStreamSourceConduit(factory.create());
|
||||||
|
}
|
||||||
|
};
|
||||||
|
LazyConduitWrapper<StreamSinkConduit> responseConduitWrapper = new LazyConduitWrapper<StreamSinkConduit>() {
|
||||||
|
@Override
|
||||||
|
protected StreamSinkConduit create(ConduitFactory<StreamSinkConduit> factory, HttpServerExchange exchange) {
|
||||||
|
return new DebuggingStreamSinkConduit(factory.create());
|
||||||
|
}
|
||||||
|
};
|
||||||
|
exchange.addRequestWrapper(requestConduitWrapper);
|
||||||
|
exchange.addResponseWrapper(responseConduitWrapper);
|
||||||
|
|
||||||
|
final StringBuilder sb = new StringBuilder();
|
||||||
|
// Log pre-service information
|
||||||
|
final SecurityContext sc = exchange.getSecurityContext();
|
||||||
|
sb.append("\n----------------------------REQUEST---------------------------\n");
|
||||||
|
sb.append(" URI=").append(exchange.getRequestURI()).append("\n");
|
||||||
|
sb.append(" characterEncoding=").append(exchange.getRequestHeaders().get(Headers.CONTENT_ENCODING)).append("\n");
|
||||||
|
sb.append(" contentLength=").append(exchange.getRequestContentLength()).append("\n");
|
||||||
|
sb.append(" contentType=").append(exchange.getRequestHeaders().get(Headers.CONTENT_TYPE)).append("\n");
|
||||||
|
//sb.append(" contextPath=" + exchange.getContextPath());
|
||||||
|
if (sc != null) {
|
||||||
|
if (sc.isAuthenticated()) {
|
||||||
|
sb.append(" authType=").append(sc.getMechanismName()).append("\n");
|
||||||
|
sb.append(" principle=").append(sc.getAuthenticatedAccount().getPrincipal()).append("\n");
|
||||||
|
} else {
|
||||||
|
sb.append(" authType=none\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Map<String, Cookie> cookies = exchange.getRequestCookies();
|
||||||
|
if (cookies != null) {
|
||||||
|
for (Map.Entry<String, Cookie> entry : cookies.entrySet()) {
|
||||||
|
Cookie cookie = entry.getValue();
|
||||||
|
sb.append(" cookie=").append(cookie.getName()).append("=").append(cookie.getValue()).append("\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (HeaderValues header : exchange.getRequestHeaders()) {
|
||||||
|
for (String value : header) {
|
||||||
|
sb.append(" header=").append(header.getHeaderName()).append("=").append(value).append("\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sb.append(" locale=").append(LocaleUtils.getLocalesFromHeader(exchange.getRequestHeaders().get(Headers.ACCEPT_LANGUAGE)))
|
||||||
|
.append("\n");
|
||||||
|
sb.append(" method=").append(exchange.getRequestMethod()).append("\n");
|
||||||
|
Map<String, Deque<String>> pnames = exchange.getQueryParameters();
|
||||||
|
for (Map.Entry<String, Deque<String>> entry : pnames.entrySet()) {
|
||||||
|
String pname = entry.getKey();
|
||||||
|
Iterator<String> pvalues = entry.getValue().iterator();
|
||||||
|
sb.append(" parameter=");
|
||||||
|
sb.append(pname);
|
||||||
|
sb.append('=');
|
||||||
|
while (pvalues.hasNext()) {
|
||||||
|
sb.append(pvalues.next());
|
||||||
|
if (pvalues.hasNext()) {
|
||||||
|
sb.append(", ");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sb.append("\n");
|
||||||
|
}
|
||||||
|
//sb.append(" pathInfo=" + exchange.getPathInfo());
|
||||||
|
sb.append(" protocol=").append(exchange.getProtocol()).append("\n");
|
||||||
|
sb.append(" queryString=").append(exchange.getQueryString()).append("\n");
|
||||||
|
sb.append(" remoteAddr=").append(exchange.getSourceAddress()).append("\n");
|
||||||
|
sb.append(" remoteHost=").append(exchange.getSourceAddress().getHostName()).append("\n");
|
||||||
|
//sb.append("requestedSessionId=" + exchange.getRequestedSessionId());
|
||||||
|
sb.append(" scheme=").append(exchange.getRequestScheme()).append("\n");
|
||||||
|
sb.append(" host=").append(exchange.getRequestHeaders().getFirst(Headers.HOST)).append("\n");
|
||||||
|
sb.append(" serverPort=").append(exchange.getDestinationAddress().getPort()).append("\n");
|
||||||
|
//sb.append(" servletPath=" + exchange.getServletPath());
|
||||||
|
sb.append(" isSecure=").append(exchange.isSecure()).append("\n");
|
||||||
|
|
||||||
|
exchange.addExchangeCompleteListener((exchange1, nextListener) -> {
|
||||||
|
StreamSourceConduit sourceConduit = requestConduitWrapper.get();
|
||||||
|
if (sourceConduit instanceof ConduitWithDump) {
|
||||||
|
ConduitWithDump conduitWithDump = (ConduitWithDump) sourceConduit;
|
||||||
|
sb.append("body=\n");
|
||||||
|
sb.append(conduitWithDump.dump()).append("\n");
|
||||||
|
}
|
||||||
|
|
||||||
|
// Log post-service information
|
||||||
|
sb.append("--------------------------RESPONSE--------------------------\n");
|
||||||
|
if (sc != null) {
|
||||||
|
if (sc.isAuthenticated()) {
|
||||||
|
sb.append(" authType=").append(sc.getMechanismName()).append("\n");
|
||||||
|
sb.append(" principle=").append(sc.getAuthenticatedAccount().getPrincipal()).append("\n");
|
||||||
|
} else {
|
||||||
|
sb.append(" authType=none\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sb.append(" contentLength=").append(exchange1.getResponseContentLength()).append("\n");
|
||||||
|
sb.append(" contentType=").append(exchange1.getResponseHeaders().getFirst(Headers.CONTENT_TYPE)).append("\n");
|
||||||
|
Map<String, Cookie> cookies1 = exchange1.getResponseCookies();
|
||||||
|
if (cookies1 != null) {
|
||||||
|
for (Cookie cookie : cookies1.values()) {
|
||||||
|
sb.append(" cookie=").append(cookie.getName()).append("=").append(cookie.getValue()).append("; domain=")
|
||||||
|
.append(cookie.getDomain()).append("; path=").append(cookie.getPath()).append("\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (HeaderValues header : exchange1.getResponseHeaders()) {
|
||||||
|
for (String value : header) {
|
||||||
|
sb.append(" header=").append(header.getHeaderName()).append("=").append(value).append("\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sb.append(" status=").append(exchange1.getStatusCode()).append("\n");
|
||||||
|
StreamSinkConduit streamSinkConduit = responseConduitWrapper.get();
|
||||||
|
if (streamSinkConduit instanceof ConduitWithDump) {
|
||||||
|
ConduitWithDump conduitWithDump = (ConduitWithDump) streamSinkConduit;
|
||||||
|
sb.append("body=\n");
|
||||||
|
sb.append(conduitWithDump.dump());
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
sb.append("\n==============================================================");
|
||||||
|
|
||||||
|
|
||||||
|
nextListener.proceed();
|
||||||
|
LOGGER.info(sb.toString());
|
||||||
|
});
|
||||||
|
|
||||||
|
|
||||||
|
// Perform the exchange
|
||||||
|
next.handleRequest(exchange);
|
||||||
|
}
|
||||||
|
}
|
@@ -82,7 +82,10 @@ public class SaneHandler extends BasicHttpHandler {
|
|||||||
} catch (InvalidJsonException e) {
|
} catch (InvalidJsonException e) {
|
||||||
respond(exchange, HttpStatus.SC_BAD_REQUEST, e.getErrorCode(), e.getError());
|
respond(exchange, HttpStatus.SC_BAD_REQUEST, e.getErrorCode(), e.getError());
|
||||||
} catch (InvalidCredentialsException e) {
|
} catch (InvalidCredentialsException e) {
|
||||||
|
log.error("Unauthorized: ", e);
|
||||||
respond(exchange, HttpStatus.SC_UNAUTHORIZED, "M_UNAUTHORIZED", e.getMessage());
|
respond(exchange, HttpStatus.SC_UNAUTHORIZED, "M_UNAUTHORIZED", e.getMessage());
|
||||||
|
} catch (TermsNotSignedException e) {
|
||||||
|
respond(exchange, HttpStatus.SC_FORBIDDEN, "M_TERMS_NOT_SIGNED", e.getMessage());
|
||||||
} catch (ObjectNotFoundException e) {
|
} catch (ObjectNotFoundException e) {
|
||||||
respond(exchange, HttpStatus.SC_NOT_FOUND, "M_NOT_FOUND", e.getMessage());
|
respond(exchange, HttpStatus.SC_NOT_FOUND, "M_NOT_FOUND", e.getMessage());
|
||||||
} catch (NotImplementedException e) {
|
} catch (NotImplementedException e) {
|
||||||
|
@@ -29,10 +29,7 @@ import java.util.Optional;
|
|||||||
public abstract class ApplicationServiceHandler extends BasicHttpHandler {
|
public abstract class ApplicationServiceHandler extends BasicHttpHandler {
|
||||||
|
|
||||||
protected String getToken(HttpServerExchange ex) {
|
protected String getToken(HttpServerExchange ex) {
|
||||||
return Optional.ofNullable(ex.getQueryParameters()
|
return getAccessToken(ex);
|
||||||
.getOrDefault("access_token", new LinkedList<>())
|
|
||||||
.peekFirst()
|
|
||||||
).orElse("");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@@ -31,6 +31,8 @@ public class HashDetailsHandler extends BasicHttpHandler {
|
|||||||
for (HashingConfig.Algorithm algorithm : config.getAlgorithms()) {
|
for (HashingConfig.Algorithm algorithm : config.getAlgorithms()) {
|
||||||
algorithms.add(algorithm.name().toLowerCase());
|
algorithms.add(algorithm.name().toLowerCase());
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
algorithms.add(HashingConfig.Algorithm.none.name().toLowerCase());
|
||||||
}
|
}
|
||||||
response.add("algorithms", algorithms);
|
response.add("algorithms", algorithms);
|
||||||
return response;
|
return response;
|
||||||
|
@@ -67,10 +67,6 @@ public class HashLookupHandler extends LookupHandler implements ApiHandler {
|
|||||||
log.info("Got bulk lookup request from {} with client {} - Is recursive? {}",
|
log.info("Got bulk lookup request from {} with client {} - Is recursive? {}",
|
||||||
lookupRequest.getRequester(), lookupRequest.getUserAgent(), lookupRequest.isRecursive());
|
lookupRequest.getRequester(), lookupRequest.getUserAgent(), lookupRequest.isRecursive());
|
||||||
|
|
||||||
if (!hashManager.getConfig().isEnabled()) {
|
|
||||||
throw new InvalidParamException();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!hashManager.getHashEngine().getPepper().equals(input.getPepper())) {
|
if (!hashManager.getHashEngine().getPepper().equals(input.getPepper())) {
|
||||||
throw new InvalidPepperException();
|
throw new InvalidPepperException();
|
||||||
}
|
}
|
||||||
@@ -89,7 +85,7 @@ public class HashLookupHandler extends LookupHandler implements ApiHandler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private void noneAlgorithm(HttpServerExchange exchange, HashLookupRequest request, ClientHashLookupRequest input) throws Exception {
|
private void noneAlgorithm(HttpServerExchange exchange, HashLookupRequest request, ClientHashLookupRequest input) throws Exception {
|
||||||
if (!hashManager.getConfig().getAlgorithms().contains(HashingConfig.Algorithm.none)) {
|
if (hashManager.getConfig().isEnabled() && !hashManager.getConfig().getAlgorithms().contains(HashingConfig.Algorithm.none)) {
|
||||||
throw new InvalidParamException();
|
throw new InvalidParamException();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -0,0 +1,30 @@
|
|||||||
|
package io.kamax.mxisd.http.undertow.handler.internal;
|
||||||
|
|
||||||
|
import com.google.gson.JsonObject;
|
||||||
|
import io.kamax.mxisd.http.undertow.handler.BasicHttpHandler;
|
||||||
|
import io.kamax.mxisd.invitation.InvitationManager;
|
||||||
|
import io.undertow.server.HttpServerExchange;
|
||||||
|
|
||||||
|
import java.util.concurrent.ExecutorService;
|
||||||
|
import java.util.concurrent.Executors;
|
||||||
|
|
||||||
|
public class InternalInviteManagerHandler extends BasicHttpHandler {
|
||||||
|
|
||||||
|
public static final String PATH = "/_ma1sd/internal/admin/inv_manager";
|
||||||
|
|
||||||
|
private final InvitationManager invitationManager;
|
||||||
|
private final ExecutorService executors = Executors.newFixedThreadPool(1);
|
||||||
|
|
||||||
|
public InternalInviteManagerHandler(InvitationManager invitationManager) {
|
||||||
|
this.invitationManager = invitationManager;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void handleRequest(HttpServerExchange exchange) throws Exception {
|
||||||
|
executors.submit(invitationManager::doMaintenance);
|
||||||
|
|
||||||
|
JsonObject obj = new JsonObject();
|
||||||
|
obj.addProperty("result", "ok");
|
||||||
|
respond(exchange, obj);
|
||||||
|
}
|
||||||
|
}
|
@@ -71,7 +71,7 @@ public class Register3pidRequestTokenHandler extends BasicHttpHandler {
|
|||||||
throw new NotAllowedException("Your " + medium + " address cannot be used for registration");
|
throw new NotAllowedException("Your " + medium + " address cannot be used for registration");
|
||||||
}
|
}
|
||||||
|
|
||||||
proxyPost(exchange, body, client, dns);
|
proxyPost(exchange, body, client, dns, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
@@ -155,7 +155,17 @@ public class InvitationManager {
|
|||||||
log.error("Error when running background maintenance", t);
|
log.error("Error when running background maintenance", t);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}, 5000L, TimeUnit.MILLISECONDS.convert(cfg.getResolution().getTimer(), TimeUnit.MINUTES));
|
}, 5000L, TimeUnit.MILLISECONDS.convert(cfg.getResolution().getTimer(), getTimeUnit()));
|
||||||
|
}
|
||||||
|
|
||||||
|
private TimeUnit getTimeUnit() {
|
||||||
|
switch (cfg.getResolution().getPeriod()) {
|
||||||
|
case seconds:
|
||||||
|
return TimeUnit.SECONDS;
|
||||||
|
case minutes:
|
||||||
|
default:
|
||||||
|
return TimeUnit.MINUTES;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private InvitationConfig requireValid(MxisdConfig cfg) {
|
private InvitationConfig requireValid(MxisdConfig cfg) {
|
||||||
@@ -176,7 +186,8 @@ public class InvitationManager {
|
|||||||
if (StringUtils.isBlank(cfg.getInvite().getExpiration().getResolveTo())) {
|
if (StringUtils.isBlank(cfg.getInvite().getExpiration().getResolveTo())) {
|
||||||
String localpart = cfg.getAppsvc().getUser().getInviteExpired();
|
String localpart = cfg.getAppsvc().getUser().getInviteExpired();
|
||||||
if (StringUtils.isBlank(localpart)) {
|
if (StringUtils.isBlank(localpart)) {
|
||||||
throw new ConfigurationException("Could not compute the Invitation expiration resolution target from App service user: not set");
|
throw new ConfigurationException(
|
||||||
|
"Could not compute the Invitation expiration resolution target from App service user: not set");
|
||||||
}
|
}
|
||||||
|
|
||||||
cfg.getInvite().getExpiration().setResolveTo(MatrixID.asAcceptable(localpart, cfg.getMatrix().getDomain()).getId());
|
cfg.getInvite().getExpiration().setResolveTo(MatrixID.asAcceptable(localpart, cfg.getMatrix().getDomain()).getId());
|
||||||
@@ -198,7 +209,8 @@ public class InvitationManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
private String getIdForLog(IThreePidInviteReply reply) {
|
private String getIdForLog(IThreePidInviteReply reply) {
|
||||||
return reply.getInvite().getSender().getId() + ":" + reply.getInvite().getRoomId() + ":" + reply.getInvite().getMedium() + ":" + reply.getInvite().getAddress();
|
return reply.getInvite().getSender().getId() + ":" + reply.getInvite().getRoomId() + ":" + reply.getInvite()
|
||||||
|
.getMedium() + ":" + reply.getInvite().getAddress();
|
||||||
}
|
}
|
||||||
|
|
||||||
private Optional<SingleLookupReply> lookup3pid(String medium, String address) {
|
private Optional<SingleLookupReply> lookup3pid(String medium, String address) {
|
||||||
@@ -252,13 +264,16 @@ public class InvitationManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
String invId = computeId(invitation);
|
String invId = computeId(invitation);
|
||||||
log.info("Handling invite for {}:{} from {} in room {}", invitation.getMedium(), invitation.getAddress(), invitation.getSender(), invitation.getRoomId());
|
log.info("Handling invite for {}:{} from {} in room {}", invitation.getMedium(), invitation.getAddress(), invitation.getSender(),
|
||||||
|
invitation.getRoomId());
|
||||||
IThreePidInviteReply reply = invitations.get(invId);
|
IThreePidInviteReply reply = invitations.get(invId);
|
||||||
if (reply != null) {
|
if (reply != null) {
|
||||||
log.info("Invite is already pending for {}:{}, returning data", invitation.getMedium(), invitation.getAddress());
|
log.info("Invite is already pending for {}:{}, returning data", invitation.getMedium(), invitation.getAddress());
|
||||||
if (!StringUtils.equals(invitation.getRoomId(), reply.getInvite().getRoomId())) {
|
if (!StringUtils.equals(invitation.getRoomId(), reply.getInvite().getRoomId())) {
|
||||||
log.info("Sending new notification as new invite room {} is different from the original {}", invitation.getRoomId(), reply.getInvite().getRoomId());
|
log.info("Sending new notification as new invite room {} is different from the original {}", invitation.getRoomId(),
|
||||||
notifMgr.sendForReply(new ThreePidInviteReply(reply.getId(), invitation, reply.getToken(), reply.getDisplayName(), reply.getPublicKeys()));
|
reply.getInvite().getRoomId());
|
||||||
|
notifMgr.sendForReply(
|
||||||
|
new ThreePidInviteReply(reply.getId(), invitation, reply.getToken(), reply.getDisplayName(), reply.getPublicKeys()));
|
||||||
} else {
|
} else {
|
||||||
// FIXME we should check attempt and send if bigger
|
// FIXME we should check attempt and send if bigger
|
||||||
}
|
}
|
||||||
@@ -272,7 +287,7 @@ public class InvitationManager {
|
|||||||
}
|
}
|
||||||
|
|
||||||
String token = RandomStringUtils.randomAlphanumeric(64);
|
String token = RandomStringUtils.randomAlphanumeric(64);
|
||||||
String displayName = invitation.getAddress().substring(0, 3) + "...";
|
String displayName = getInvitedDisplayName(invitation.getAddress());
|
||||||
KeyIdentifier pKeyId = keyMgr.getServerSigningKey().getId();
|
KeyIdentifier pKeyId = keyMgr.getServerSigningKey().getId();
|
||||||
KeyIdentifier eKeyId = keyMgr.generateKey(KeyType.Ephemeral);
|
KeyIdentifier eKeyId = keyMgr.generateKey(KeyType.Ephemeral);
|
||||||
|
|
||||||
@@ -295,11 +310,20 @@ public class InvitationManager {
|
|||||||
log.info("Storing invite under ID {}", invId);
|
log.info("Storing invite under ID {}", invId);
|
||||||
storage.insertInvite(reply);
|
storage.insertInvite(reply);
|
||||||
invitations.put(invId, reply);
|
invitations.put(invId, reply);
|
||||||
log.info("A new invite has been created for {}:{} on HS {}", invitation.getMedium(), invitation.getAddress(), invitation.getSender().getDomain());
|
log.info("A new invite has been created for {}:{} on HS {}", invitation.getMedium(), invitation.getAddress(),
|
||||||
|
invitation.getSender().getDomain());
|
||||||
|
|
||||||
return reply;
|
return reply;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private String getInvitedDisplayName(String origin) {
|
||||||
|
if (cfg.isFullDisplayName()) {
|
||||||
|
return origin;
|
||||||
|
} else {
|
||||||
|
return origin.substring(0, 3) + "...";
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
public boolean hasInvite(ThreePid tpid) {
|
public boolean hasInvite(ThreePid tpid) {
|
||||||
for (IThreePidInviteReply reply : invitations.values()) {
|
for (IThreePidInviteReply reply : invitations.values()) {
|
||||||
if (!StringUtils.equals(tpid.getMedium(), reply.getInvite().getMedium())) {
|
if (!StringUtils.equals(tpid.getMedium(), reply.getInvite().getMedium())) {
|
||||||
@@ -385,8 +409,10 @@ public class InvitationManager {
|
|||||||
public void publishMappingIfInvited(ThreePidMapping threePid) {
|
public void publishMappingIfInvited(ThreePidMapping threePid) {
|
||||||
log.info("Looking up possible pending invites for {}:{}", threePid.getMedium(), threePid.getValue());
|
log.info("Looking up possible pending invites for {}:{}", threePid.getMedium(), threePid.getValue());
|
||||||
for (IThreePidInviteReply reply : invitations.values()) {
|
for (IThreePidInviteReply reply : invitations.values()) {
|
||||||
if (StringUtils.equalsIgnoreCase(reply.getInvite().getMedium(), threePid.getMedium()) && StringUtils.equalsIgnoreCase(reply.getInvite().getAddress(), threePid.getValue())) {
|
if (StringUtils.equalsIgnoreCase(reply.getInvite().getMedium(), threePid.getMedium()) && StringUtils
|
||||||
log.info("{}:{} has an invite pending on HS {}, publishing mapping", threePid.getMedium(), threePid.getValue(), reply.getInvite().getSender().getDomain());
|
.equalsIgnoreCase(reply.getInvite().getAddress(), threePid.getValue())) {
|
||||||
|
log.info("{}:{} has an invite pending on HS {}, publishing mapping", threePid.getMedium(), threePid.getValue(),
|
||||||
|
reply.getInvite().getSender().getDomain());
|
||||||
publishMapping(reply, threePid.getMxid());
|
publishMapping(reply, threePid.getMxid());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -130,7 +130,9 @@ public class HomeserverFederationResolver {
|
|||||||
|
|
||||||
return Optional.empty();
|
return Optional.empty();
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new RuntimeException("Error while trying to lookup well-known for " + domain, e);
|
log.info("Error while trying to lookup well-known for " + domain);
|
||||||
|
log.trace("Error while trying to lookup well-known for " + domain, e);
|
||||||
|
return Optional.empty();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -24,12 +24,17 @@ import com.j256.ormlite.dao.CloseableWrappedIterable;
|
|||||||
import com.j256.ormlite.dao.Dao;
|
import com.j256.ormlite.dao.Dao;
|
||||||
import com.j256.ormlite.dao.DaoManager;
|
import com.j256.ormlite.dao.DaoManager;
|
||||||
import com.j256.ormlite.jdbc.JdbcConnectionSource;
|
import com.j256.ormlite.jdbc.JdbcConnectionSource;
|
||||||
|
import com.j256.ormlite.jdbc.JdbcPooledConnectionSource;
|
||||||
|
import com.j256.ormlite.jdbc.db.PostgresDatabaseType;
|
||||||
|
import com.j256.ormlite.jdbc.db.SqliteDatabaseType;
|
||||||
import com.j256.ormlite.stmt.QueryBuilder;
|
import com.j256.ormlite.stmt.QueryBuilder;
|
||||||
import com.j256.ormlite.support.ConnectionSource;
|
import com.j256.ormlite.support.ConnectionSource;
|
||||||
import com.j256.ormlite.table.TableUtils;
|
import com.j256.ormlite.table.TableUtils;
|
||||||
import io.kamax.matrix.ThreePid;
|
import io.kamax.matrix.ThreePid;
|
||||||
import io.kamax.mxisd.config.MxisdConfig;
|
|
||||||
import io.kamax.mxisd.config.PolicyConfig;
|
import io.kamax.mxisd.config.PolicyConfig;
|
||||||
|
import io.kamax.mxisd.config.PostgresqlStorageConfig;
|
||||||
|
import io.kamax.mxisd.config.SQLiteStorageConfig;
|
||||||
|
import io.kamax.mxisd.config.StorageConfig;
|
||||||
import io.kamax.mxisd.exception.ConfigurationException;
|
import io.kamax.mxisd.exception.ConfigurationException;
|
||||||
import io.kamax.mxisd.exception.InternalServerError;
|
import io.kamax.mxisd.exception.InternalServerError;
|
||||||
import io.kamax.mxisd.exception.InvalidCredentialsException;
|
import io.kamax.mxisd.exception.InvalidCredentialsException;
|
||||||
@@ -81,6 +86,9 @@ public class OrmLiteSqlStorage implements IStorage {
|
|||||||
|
|
||||||
public static class Migrations {
|
public static class Migrations {
|
||||||
public static final String FIX_ACCEPTED_DAO = "2019_12_09__2254__fix_accepted_dao";
|
public static final String FIX_ACCEPTED_DAO = "2019_12_09__2254__fix_accepted_dao";
|
||||||
|
public static final String FIX_HASH_DAO_UNIQUE_INDEX = "2020_03_22__1153__fix_hash_dao_unique_index";
|
||||||
|
public static final String CHANGE_TYPE_TO_TEXT_INVITE = "2020_04_21__2338__change_type_table_invites";
|
||||||
|
public static final String CHANGE_TYPE_TO_TEXT_INVITE_HISTORY = "2020_10_26__2200__change_type_table_invite_history";
|
||||||
}
|
}
|
||||||
|
|
||||||
private Dao<ThreePidInviteIO, String> invDao;
|
private Dao<ThreePidInviteIO, String> invDao;
|
||||||
@@ -91,40 +99,90 @@ public class OrmLiteSqlStorage implements IStorage {
|
|||||||
private Dao<AcceptedDao, Long> acceptedDao;
|
private Dao<AcceptedDao, Long> acceptedDao;
|
||||||
private Dao<HashDao, String> hashDao;
|
private Dao<HashDao, String> hashDao;
|
||||||
private Dao<ChangelogDao, String> changelogDao;
|
private Dao<ChangelogDao, String> changelogDao;
|
||||||
|
private StorageConfig.BackendEnum backend;
|
||||||
|
|
||||||
public OrmLiteSqlStorage(MxisdConfig cfg) {
|
public OrmLiteSqlStorage(StorageConfig.BackendEnum backend, StorageConfig.Provider provider) {
|
||||||
this(cfg.getStorage().getBackend(), cfg.getStorage().getProvider().getSqlite().getDatabase());
|
if (backend == null) {
|
||||||
|
throw new ConfigurationException("storage.backend");
|
||||||
}
|
}
|
||||||
|
this.backend = backend;
|
||||||
|
|
||||||
public OrmLiteSqlStorage(String backend, String path) {
|
withCatcher(() -> {
|
||||||
if (StringUtils.isBlank(backend)) {
|
ConnectionSource connPool;
|
||||||
|
switch (backend) {
|
||||||
|
case postgresql:
|
||||||
|
connPool = createPostgresqlConnection(provider.getPostgresql());
|
||||||
|
break;
|
||||||
|
case sqlite:
|
||||||
|
connPool = createSqliteConnection(provider.getSqlite());
|
||||||
|
break;
|
||||||
|
default:
|
||||||
throw new ConfigurationException("storage.backend");
|
throw new ConfigurationException("storage.backend");
|
||||||
}
|
}
|
||||||
|
|
||||||
if (StringUtils.isBlank(path)) {
|
|
||||||
throw new ConfigurationException("Storage destination cannot be empty");
|
|
||||||
}
|
|
||||||
|
|
||||||
withCatcher(() -> {
|
|
||||||
ConnectionSource connPool = new JdbcConnectionSource("jdbc:" + backend + ":" + path);
|
|
||||||
changelogDao = createDaoAndTable(connPool, ChangelogDao.class);
|
changelogDao = createDaoAndTable(connPool, ChangelogDao.class);
|
||||||
invDao = createDaoAndTable(connPool, ThreePidInviteIO.class);
|
invDao = createDaoAndTable(connPool, ThreePidInviteIO.class);
|
||||||
expInvDao = createDaoAndTable(connPool, HistoricalThreePidInviteIO.class);
|
expInvDao = createDaoAndTable(connPool, HistoricalThreePidInviteIO.class);
|
||||||
sessionDao = createDaoAndTable(connPool, ThreePidSessionDao.class);
|
sessionDao = createDaoAndTable(connPool, ThreePidSessionDao.class);
|
||||||
asTxnDao = createDaoAndTable(connPool, ASTransactionDao.class);
|
asTxnDao = createDaoAndTable(connPool, ASTransactionDao.class);
|
||||||
accountDao = createDaoAndTable(connPool, AccountDao.class);
|
accountDao = createDaoAndTable(connPool, AccountDao.class);
|
||||||
acceptedDao = createDaoAndTable(connPool, AcceptedDao.class);
|
acceptedDao = createDaoAndTable(connPool, AcceptedDao.class, true);
|
||||||
hashDao = createDaoAndTable(connPool, HashDao.class);
|
hashDao = createDaoAndTable(connPool, HashDao.class, true);
|
||||||
runMigration(connPool);
|
runMigration(connPool);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private ConnectionSource createSqliteConnection(SQLiteStorageConfig config) throws SQLException {
|
||||||
|
if (StringUtils.isBlank(config.getDatabase())) {
|
||||||
|
throw new ConfigurationException("Storage destination cannot be empty");
|
||||||
|
}
|
||||||
|
|
||||||
|
return new JdbcConnectionSource("jdbc:" + backend + ":" + config.getDatabase(), null, null, new SqliteDatabaseType());
|
||||||
|
}
|
||||||
|
|
||||||
|
private ConnectionSource createPostgresqlConnection(PostgresqlStorageConfig config) throws SQLException {
|
||||||
|
if (StringUtils.isBlank(config.getDatabase())) {
|
||||||
|
throw new ConfigurationException("Storage destination cannot be empty");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (config.isPool()) {
|
||||||
|
LOGGER.info("Enable pooling");
|
||||||
|
JdbcPooledConnectionSource source = new JdbcPooledConnectionSource(
|
||||||
|
"jdbc:" + backend + ":" + config.getDatabase(), config.getUsername(), config.getPassword(),
|
||||||
|
new PostgresDatabaseType());
|
||||||
|
source.setMaxConnectionsFree(config.getMaxConnectionsFree());
|
||||||
|
source.setMaxConnectionAgeMillis(config.getMaxConnectionAgeMillis());
|
||||||
|
source.setCheckConnectionsEveryMillis(config.getCheckConnectionsEveryMillis());
|
||||||
|
source.setTestBeforeGet(config.isTestBeforeGetFromPool());
|
||||||
|
return source;
|
||||||
|
} else {
|
||||||
|
return new JdbcConnectionSource("jdbc:" + backend + ":" + config.getDatabase(), config.getUsername(), config.getPassword(),
|
||||||
|
new PostgresDatabaseType());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private void runMigration(ConnectionSource connPol) throws SQLException {
|
private void runMigration(ConnectionSource connPol) throws SQLException {
|
||||||
ChangelogDao fixAcceptedDao = changelogDao.queryForId(Migrations.FIX_ACCEPTED_DAO);
|
ChangelogDao fixAcceptedDao = changelogDao.queryForId(Migrations.FIX_ACCEPTED_DAO);
|
||||||
if (fixAcceptedDao == null) {
|
if (fixAcceptedDao == null) {
|
||||||
fixAcceptedDao(connPol);
|
fixAcceptedDao(connPol);
|
||||||
changelogDao.create(new ChangelogDao(Migrations.FIX_ACCEPTED_DAO, new Date(), "Recreate the accepted table."));
|
changelogDao.create(new ChangelogDao(Migrations.FIX_ACCEPTED_DAO, new Date(), "Recreate the accepted table."));
|
||||||
}
|
}
|
||||||
|
ChangelogDao fixHashDaoUniqueIndex = changelogDao.queryForId(Migrations.FIX_HASH_DAO_UNIQUE_INDEX);
|
||||||
|
if (fixHashDaoUniqueIndex == null) {
|
||||||
|
fixHashDaoUniqueIndex(connPol);
|
||||||
|
changelogDao
|
||||||
|
.create(new ChangelogDao(Migrations.FIX_HASH_DAO_UNIQUE_INDEX, new Date(), "Add the id and migrate the unique index."));
|
||||||
|
}
|
||||||
|
ChangelogDao fixInviteTableColumnType = changelogDao.queryForId(Migrations.CHANGE_TYPE_TO_TEXT_INVITE);
|
||||||
|
if (fixInviteTableColumnType == null) {
|
||||||
|
fixInviteTableColumnType(connPol);
|
||||||
|
changelogDao.create(new ChangelogDao(Migrations.CHANGE_TYPE_TO_TEXT_INVITE, new Date(), "Modify column type to text."));
|
||||||
|
}
|
||||||
|
ChangelogDao fixInviteHistoryTableColumnType = changelogDao.queryForId(Migrations.CHANGE_TYPE_TO_TEXT_INVITE_HISTORY);
|
||||||
|
if (fixInviteHistoryTableColumnType == null) {
|
||||||
|
fixInviteHistoryTableColumnType(connPol);
|
||||||
|
changelogDao.create(new ChangelogDao(Migrations.CHANGE_TYPE_TO_TEXT_INVITE_HISTORY, new Date(), "Modify column type to text."));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private void fixAcceptedDao(ConnectionSource connPool) throws SQLException {
|
private void fixAcceptedDao(ConnectionSource connPool) throws SQLException {
|
||||||
@@ -133,10 +191,61 @@ public class OrmLiteSqlStorage implements IStorage {
|
|||||||
TableUtils.createTableIfNotExists(connPool, AcceptedDao.class);
|
TableUtils.createTableIfNotExists(connPool, AcceptedDao.class);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private void fixHashDaoUniqueIndex(ConnectionSource connPool) throws SQLException {
|
||||||
|
LOGGER.info("Migration: {}", Migrations.FIX_HASH_DAO_UNIQUE_INDEX);
|
||||||
|
TableUtils.dropTable(hashDao, true);
|
||||||
|
TableUtils.createTableIfNotExists(connPool, HashDao.class);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void fixInviteTableColumnType(ConnectionSource connPool) throws SQLException {
|
||||||
|
LOGGER.info("Migration: {}", Migrations.CHANGE_TYPE_TO_TEXT_INVITE);
|
||||||
|
if (StorageConfig.BackendEnum.postgresql == backend) {
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column \"roomId\" type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column id type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column token type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column sender type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column medium type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column address type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid alter column properties type text");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void fixInviteHistoryTableColumnType(ConnectionSource connPool) throws SQLException {
|
||||||
|
LOGGER.info("Migration: {}", Migrations.CHANGE_TYPE_TO_TEXT_INVITE_HISTORY);
|
||||||
|
if (StorageConfig.BackendEnum.postgresql == backend) {
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column \"resolvedTo\" type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column id type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column token type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column sender type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column medium type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column address type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column \"roomId\" type text");
|
||||||
|
invDao.executeRawNoArgs("alter table invite_3pid_history alter column properties type text");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private <V, K> Dao<V, K> createDaoAndTable(ConnectionSource connPool, Class<V> c) throws SQLException {
|
private <V, K> Dao<V, K> createDaoAndTable(ConnectionSource connPool, Class<V> c) throws SQLException {
|
||||||
|
return createDaoAndTable(connPool, c, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Workaround for https://github.com/j256/ormlite-core/issues/20.
|
||||||
|
*/
|
||||||
|
private <V, K> Dao<V, K> createDaoAndTable(ConnectionSource connPool, Class<V> c, boolean workaround) throws SQLException {
|
||||||
LOGGER.info("Create the dao: {}", c.getSimpleName());
|
LOGGER.info("Create the dao: {}", c.getSimpleName());
|
||||||
Dao<V, K> dao = DaoManager.createDao(connPool, c);
|
Dao<V, K> dao = DaoManager.createDao(connPool, c);
|
||||||
|
if (workaround && StorageConfig.BackendEnum.postgresql.equals(backend)) {
|
||||||
|
LOGGER.info("Workaround for postgresql on dao: {}", c.getSimpleName());
|
||||||
|
try {
|
||||||
|
dao.countOf();
|
||||||
|
LOGGER.info("Table exists, do nothing");
|
||||||
|
} catch (SQLException e) {
|
||||||
|
LOGGER.info("Table doesn't exist, create");
|
||||||
TableUtils.createTableIfNotExists(connPool, c);
|
TableUtils.createTableIfNotExists(connPool, c);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
TableUtils.createTableIfNotExists(connPool, c);
|
||||||
|
}
|
||||||
return dao;
|
return dao;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -6,13 +6,16 @@ import com.j256.ormlite.table.DatabaseTable;
|
|||||||
@DatabaseTable(tableName = "hashes")
|
@DatabaseTable(tableName = "hashes")
|
||||||
public class HashDao {
|
public class HashDao {
|
||||||
|
|
||||||
@DatabaseField(canBeNull = false, id = true)
|
@DatabaseField(generatedId = true)
|
||||||
|
private Long id;
|
||||||
|
|
||||||
|
@DatabaseField(canBeNull = false, uniqueCombo = true)
|
||||||
private String mxid;
|
private String mxid;
|
||||||
|
|
||||||
@DatabaseField(canBeNull = false)
|
@DatabaseField(canBeNull = false, uniqueCombo = true)
|
||||||
private String medium;
|
private String medium;
|
||||||
|
|
||||||
@DatabaseField(canBeNull = false)
|
@DatabaseField(canBeNull = false, uniqueCombo = true)
|
||||||
private String address;
|
private String address;
|
||||||
|
|
||||||
@DatabaseField(canBeNull = false)
|
@DatabaseField(canBeNull = false)
|
||||||
@@ -28,6 +31,14 @@ public class HashDao {
|
|||||||
this.hash = hash;
|
this.hash = hash;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public Long getId() {
|
||||||
|
return id;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setId(Long id) {
|
||||||
|
this.id = id;
|
||||||
|
}
|
||||||
|
|
||||||
public String getMxid() {
|
public String getMxid() {
|
||||||
return mxid;
|
return mxid;
|
||||||
}
|
}
|
||||||
|
@@ -20,6 +20,8 @@
|
|||||||
|
|
||||||
package io.kamax.mxisd.test.storage;
|
package io.kamax.mxisd.test.storage;
|
||||||
|
|
||||||
|
import io.kamax.mxisd.config.SQLiteStorageConfig;
|
||||||
|
import io.kamax.mxisd.config.StorageConfig;
|
||||||
import io.kamax.mxisd.storage.ormlite.OrmLiteSqlStorage;
|
import io.kamax.mxisd.storage.ormlite.OrmLiteSqlStorage;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
@@ -29,14 +31,22 @@ public class OrmLiteSqlStorageTest {
|
|||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void insertAsTxnDuplicate() {
|
public void insertAsTxnDuplicate() {
|
||||||
OrmLiteSqlStorage store = new OrmLiteSqlStorage("sqlite", ":memory:");
|
StorageConfig.Provider provider = new StorageConfig.Provider();
|
||||||
|
SQLiteStorageConfig config = new SQLiteStorageConfig();
|
||||||
|
config.setDatabase(":memory:");
|
||||||
|
provider.setSqlite(config);
|
||||||
|
OrmLiteSqlStorage store = new OrmLiteSqlStorage(StorageConfig.BackendEnum.sqlite, provider);
|
||||||
store.insertTransactionResult("mxisd", "1", Instant.now(), "{}");
|
store.insertTransactionResult("mxisd", "1", Instant.now(), "{}");
|
||||||
store.insertTransactionResult("mxisd", "2", Instant.now(), "{}");
|
store.insertTransactionResult("mxisd", "2", Instant.now(), "{}");
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test(expected = RuntimeException.class)
|
@Test(expected = RuntimeException.class)
|
||||||
public void insertAsTxnSame() {
|
public void insertAsTxnSame() {
|
||||||
OrmLiteSqlStorage store = new OrmLiteSqlStorage("sqlite", ":memory:");
|
StorageConfig.Provider provider = new StorageConfig.Provider();
|
||||||
|
SQLiteStorageConfig config = new SQLiteStorageConfig();
|
||||||
|
config.setDatabase(":memory:");
|
||||||
|
provider.setSqlite(config);
|
||||||
|
OrmLiteSqlStorage store = new OrmLiteSqlStorage(StorageConfig.BackendEnum.sqlite, provider);
|
||||||
store.insertTransactionResult("mxisd", "1", Instant.now(), "{}");
|
store.insertTransactionResult("mxisd", "1", Instant.now(), "{}");
|
||||||
store.insertTransactionResult("mxisd", "1", Instant.now(), "{}");
|
store.insertTransactionResult("mxisd", "1", Instant.now(), "{}");
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user