mirror of
https://git.pleroma.social/pleroma/relay.git
synced 2024-11-09 18:08:00 +00:00
Merge branch 'sqldatabase' into 'master'
Draft: Use SQL database backend See merge request pleroma/relay!47
This commit is contained in:
commit
172da058d1
|
@ -3,9 +3,9 @@
|
||||||
There are a number of commands to manage your relay's database and config. You can add `--help` to
|
There are a number of commands to manage your relay's database and config. You can add `--help` to
|
||||||
any category or command to get help on that specific option (ex. `activityrelay inbox --help`).
|
any category or command to get help on that specific option (ex. `activityrelay inbox --help`).
|
||||||
|
|
||||||
Note: Unless specified, it is recommended to run any commands while the relay is shutdown.
|
A config file can be specified by adding `--config [path/to/config.yaml]`.
|
||||||
|
|
||||||
Note 2: `activityrelay` is only available via pip or pipx if `~/.local/bin` is in `$PATH`. If it
|
Note: `activityrelay` is only available via pip or pipx if `~/.local/bin` is in `$PATH`. If it
|
||||||
isn't, use `python3 -m relay` if installed via pip or `~/.local/bin/activityrelay` if installed
|
isn't, use `python3 -m relay` if installed via pip or `~/.local/bin/activityrelay` if installed
|
||||||
via pipx
|
via pipx
|
||||||
|
|
||||||
|
@ -24,26 +24,35 @@ Run the setup wizard to configure your relay.
|
||||||
activityrelay setup
|
activityrelay setup
|
||||||
|
|
||||||
|
|
||||||
|
## Convert
|
||||||
|
|
||||||
|
Convert an old `relay.yaml` and `relay.jsonld` to the newer formats.
|
||||||
|
|
||||||
|
activityrelay convert [--old-config relay.yaml]
|
||||||
|
|
||||||
|
|
||||||
## Config
|
## Config
|
||||||
|
|
||||||
Manage the relay config
|
Manage the relay config.
|
||||||
|
|
||||||
activityrelay config
|
activityrelay config
|
||||||
|
|
||||||
|
|
||||||
### List
|
### List
|
||||||
|
|
||||||
List the current config key/value pairs
|
List the current config key/value pairs.
|
||||||
|
|
||||||
activityrelay config list
|
activityrelay config list
|
||||||
|
|
||||||
|
|
||||||
### Set
|
### Set
|
||||||
|
|
||||||
Set a value for a config option
|
Set a value for a config option.
|
||||||
|
|
||||||
activityrelay config set <key> <value>
|
activityrelay config set <key> <value>
|
||||||
|
|
||||||
|
note: The relay must be restarted if setting `log_level`, `workers`, `push_limit`, or `http_timeout`
|
||||||
|
|
||||||
|
|
||||||
## Inbox
|
## Inbox
|
||||||
|
|
||||||
|
@ -92,6 +101,32 @@ not exist anymore, use the `inbox remove` command instead.
|
||||||
Note: The relay must be running for this command to work.
|
Note: The relay must be running for this command to work.
|
||||||
|
|
||||||
|
|
||||||
|
## Request
|
||||||
|
|
||||||
|
Manage instance follow requests.
|
||||||
|
|
||||||
|
|
||||||
|
### List
|
||||||
|
|
||||||
|
List all instances asking to follow the relay.
|
||||||
|
|
||||||
|
activityrelay request list
|
||||||
|
|
||||||
|
|
||||||
|
### Approve
|
||||||
|
|
||||||
|
Allow an instance to join the relay.
|
||||||
|
|
||||||
|
activityrelay request approve <domain>
|
||||||
|
|
||||||
|
|
||||||
|
### Deny
|
||||||
|
|
||||||
|
Disallow an instance to join the relay.
|
||||||
|
|
||||||
|
activityrelay request deny <domain>
|
||||||
|
|
||||||
|
|
||||||
## Whitelist
|
## Whitelist
|
||||||
|
|
||||||
Manage the whitelisted domains.
|
Manage the whitelisted domains.
|
||||||
|
@ -120,7 +155,7 @@ Remove a domain from the whitelist.
|
||||||
|
|
||||||
### Import
|
### Import
|
||||||
|
|
||||||
Add all current inboxes to the whitelist
|
Add all current inboxes to the whitelist.
|
||||||
|
|
||||||
activityrelay whitelist import
|
activityrelay whitelist import
|
||||||
|
|
||||||
|
@ -132,7 +167,7 @@ Manage the instance ban list.
|
||||||
|
|
||||||
### List
|
### List
|
||||||
|
|
||||||
List the currently banned instances
|
List the currently banned instances.
|
||||||
|
|
||||||
activityrelay instance list
|
activityrelay instance list
|
||||||
|
|
||||||
|
|
|
@ -2,14 +2,6 @@
|
||||||
|
|
||||||
## General
|
## General
|
||||||
|
|
||||||
### DB
|
|
||||||
|
|
||||||
The path to the database. It contains the relay actor private key and all subscribed
|
|
||||||
instances. If the path is not absolute, it is relative to the working directory.
|
|
||||||
|
|
||||||
db: relay.jsonld
|
|
||||||
|
|
||||||
|
|
||||||
### Listener
|
### Listener
|
||||||
|
|
||||||
The address and port the relay will listen on. If the reverse proxy (nginx, apache, caddy, etc)
|
The address and port the relay will listen on. If the reverse proxy (nginx, apache, caddy, etc)
|
||||||
|
@ -19,46 +11,6 @@ is running on the same host, it is recommended to change `listen` to `localhost`
|
||||||
port: 8080
|
port: 8080
|
||||||
|
|
||||||
|
|
||||||
### Note
|
|
||||||
|
|
||||||
A small blurb to describe your relay instance. This will show up on the relay's home page.
|
|
||||||
|
|
||||||
note: "Make a note about your instance here."
|
|
||||||
|
|
||||||
|
|
||||||
### Post Limit
|
|
||||||
|
|
||||||
The maximum number of messages to send out at once. For each incoming message, a message will be
|
|
||||||
sent out to every subscribed instance minus the instance which sent the message. This limit
|
|
||||||
is to prevent too many outgoing connections from being made, so adjust if necessary.
|
|
||||||
|
|
||||||
Note: If the `workers` option is set to anything above 0, this limit will be per worker.
|
|
||||||
|
|
||||||
push_limit: 512
|
|
||||||
|
|
||||||
|
|
||||||
### Push Workers
|
|
||||||
|
|
||||||
The relay can be configured to use threads to push messages out. For smaller relays, this isn't
|
|
||||||
necessary, but bigger ones (>100 instances) will want to set this to the number of available cpu
|
|
||||||
threads.
|
|
||||||
|
|
||||||
workers: 0
|
|
||||||
|
|
||||||
|
|
||||||
### JSON GET cache limit
|
|
||||||
|
|
||||||
JSON objects (actors, nodeinfo, etc) will get cached when fetched. This will set the max number of
|
|
||||||
objects to keep in the cache.
|
|
||||||
|
|
||||||
json_cache: 1024
|
|
||||||
|
|
||||||
|
|
||||||
## AP
|
|
||||||
|
|
||||||
Various ActivityPub-related settings
|
|
||||||
|
|
||||||
|
|
||||||
### Host
|
### Host
|
||||||
|
|
||||||
The domain your relay will use to identify itself.
|
The domain your relay will use to identify itself.
|
||||||
|
@ -66,40 +18,123 @@ The domain your relay will use to identify itself.
|
||||||
host: relay.example.com
|
host: relay.example.com
|
||||||
|
|
||||||
|
|
||||||
### Whitelist Enabled
|
## Database
|
||||||
|
|
||||||
If set to `true`, only instances in the whitelist can follow the relay. Any subscribed instances
|
### Type
|
||||||
not in the whitelist will be removed from the inbox list on startup.
|
|
||||||
|
|
||||||
whitelist_enabled: false
|
The type of SQL database to use. Options:
|
||||||
|
|
||||||
|
* sqlite (default)
|
||||||
|
* postgresql
|
||||||
|
* mysql
|
||||||
|
|
||||||
|
type: sqlite
|
||||||
|
|
||||||
|
|
||||||
### Whitelist
|
### Minimum Connections
|
||||||
|
|
||||||
A list of domains of instances which are allowed to subscribe to your relay.
|
The minimum number of database connections to keep open (does nothing at the moment)
|
||||||
|
|
||||||
whitelist:
|
min_connections: 0
|
||||||
- bad-instance.example.com
|
|
||||||
- another-bad-instance.example.com
|
|
||||||
|
|
||||||
|
|
||||||
### Blocked Instances
|
### Maximum Connections
|
||||||
|
|
||||||
A list of instances which are unable to follow the instance. If a subscribed instance is added to
|
The maximum number of database connections to open (does nothing at the moment)
|
||||||
the block list, it will be removed from the inbox list on startup.
|
|
||||||
|
|
||||||
blocked_instances:
|
max_connections: 10
|
||||||
- bad-instance.example.com
|
|
||||||
- another-bad-instance.example.com
|
|
||||||
|
|
||||||
|
|
||||||
### Blocked Software
|
## Sqlite
|
||||||
|
|
||||||
A list of ActivityPub software which cannot follow your relay. This list is empty by default, but
|
### Database
|
||||||
setting this to the below list will block all other relays and prevent relay chains
|
|
||||||
|
|
||||||
blocked_software:
|
The path to the database file.
|
||||||
- activityrelay
|
|
||||||
- aoderelay
|
database: relay.sqlite3
|
||||||
- social.seattle.wa.us-relay
|
|
||||||
- unciarelay
|
If the path is relative, it will be relative to the directory the config file is located. For
|
||||||
|
instance, if the config is located at `/home/izalia/.config/activityrelay/config.yaml`, the
|
||||||
|
following:
|
||||||
|
|
||||||
|
relay.sqlite3
|
||||||
|
|
||||||
|
will resolve to:
|
||||||
|
|
||||||
|
/home/izalia/.config/activityrelay/relay.sqlite3
|
||||||
|
|
||||||
|
|
||||||
|
## PostgreSQL
|
||||||
|
|
||||||
|
### Database
|
||||||
|
|
||||||
|
Name of the database to use.
|
||||||
|
|
||||||
|
database: activityrelay
|
||||||
|
|
||||||
|
|
||||||
|
### Hostname
|
||||||
|
|
||||||
|
The address to use when connecting to the database. A value of `null` will use the default of
|
||||||
|
`/var/run/postgresql`
|
||||||
|
|
||||||
|
hostname: null
|
||||||
|
|
||||||
|
|
||||||
|
### Port
|
||||||
|
|
||||||
|
The port to use when connecting to the database. A value of `null` will use the default of `5432`
|
||||||
|
|
||||||
|
port: null
|
||||||
|
|
||||||
|
|
||||||
|
### Username
|
||||||
|
|
||||||
|
The user to use when connecting to the database. A value of `null` will use the current system
|
||||||
|
username.
|
||||||
|
|
||||||
|
username: null
|
||||||
|
|
||||||
|
|
||||||
|
### Password
|
||||||
|
|
||||||
|
The password for the database user.
|
||||||
|
|
||||||
|
password: null
|
||||||
|
|
||||||
|
|
||||||
|
## MySQL
|
||||||
|
|
||||||
|
### Database
|
||||||
|
|
||||||
|
Name of the database to use.
|
||||||
|
|
||||||
|
database: activityrelay
|
||||||
|
|
||||||
|
|
||||||
|
### Hostname
|
||||||
|
|
||||||
|
The address to use when connecting to the database. A value of `null` will use the default of
|
||||||
|
`/var/run/mysqld/mysqld.sock`
|
||||||
|
|
||||||
|
|
||||||
|
### Port
|
||||||
|
|
||||||
|
The port to use when connecting to the database. A value of `null` will use the default of `3306`
|
||||||
|
|
||||||
|
port: null
|
||||||
|
|
||||||
|
|
||||||
|
### Username
|
||||||
|
|
||||||
|
The user to use when connecting to the database. A value of `null` will use the current system
|
||||||
|
username.
|
||||||
|
|
||||||
|
username: null
|
||||||
|
|
||||||
|
|
||||||
|
### Password
|
||||||
|
|
||||||
|
The password for the database user.
|
||||||
|
|
||||||
|
password: null
|
||||||
|
|
|
@ -5,6 +5,19 @@ proxy, and setup the relay to run via a supervisor. Example configs for caddy, n
|
||||||
in `installation/`
|
in `installation/`
|
||||||
|
|
||||||
|
|
||||||
|
## Pre-build Executables
|
||||||
|
|
||||||
|
All in one executables can be downloaded from `https://git.pleroma.social/pleroma/relay/-/releases`
|
||||||
|
under the `Other` section of `Assets`. They don't require any extra setup and can be placed
|
||||||
|
anywhere. Run the setup wizard
|
||||||
|
|
||||||
|
./activityrelay setup
|
||||||
|
|
||||||
|
and start it up when done
|
||||||
|
|
||||||
|
./activityrelay run
|
||||||
|
|
||||||
|
|
||||||
## Pipx
|
## Pipx
|
||||||
|
|
||||||
Pipx uses pip and a custom venv implementation to automatically install modules into a Python
|
Pipx uses pip and a custom venv implementation to automatically install modules into a Python
|
||||||
|
|
10
relay.spec
10
relay.spec
|
@ -14,7 +14,15 @@ a = Analysis(
|
||||||
'aputils.errors',
|
'aputils.errors',
|
||||||
'aputils.misc',
|
'aputils.misc',
|
||||||
'aputils.objects',
|
'aputils.objects',
|
||||||
'aputils.signer'
|
'aputils.signer',
|
||||||
|
|
||||||
|
'tinysql.base',
|
||||||
|
'tinysql.database',
|
||||||
|
'tinysql.error',
|
||||||
|
'tinysql.mysql',
|
||||||
|
'tinysql.postgresql',
|
||||||
|
'tinysql.sqlite',
|
||||||
|
'tinysql.statement'
|
||||||
],
|
],
|
||||||
hookspath=[],
|
hookspath=[],
|
||||||
hooksconfig={},
|
hooksconfig={},
|
||||||
|
|
|
@ -1,43 +1,32 @@
|
||||||
# this is the path that the object graph will get dumped to (in JSON-LD format),
|
general:
|
||||||
# you probably shouldn't change it, but you can if you want.
|
# Address the relay will listen on. Set to "0.0.0.0" for any address
|
||||||
db: relay.jsonld
|
listen: 0.0.0.0
|
||||||
|
# TCP port the relay will listen on
|
||||||
|
port: 3621
|
||||||
|
# Domain the relay will advertise itself as
|
||||||
|
host: relay.example.com
|
||||||
|
|
||||||
# Listener
|
database:
|
||||||
listen: 0.0.0.0
|
# SQL backend to use. Available options: "sqlite", "postgresql", "mysql".
|
||||||
port: 8080
|
type: sqlite
|
||||||
|
# Minimum number of database connections to keep open
|
||||||
|
min_connections: 0
|
||||||
|
# Maximum number of database connections to open
|
||||||
|
max_connections: 10
|
||||||
|
|
||||||
# Note
|
sqlite:
|
||||||
note: "Make a note about your instance here."
|
database: relay.sqlite3
|
||||||
|
|
||||||
# Number of worker threads to start. If 0, use asyncio futures instead of threads.
|
postgres:
|
||||||
workers: 0
|
database: activityrelay
|
||||||
|
hostname: null
|
||||||
|
port: null
|
||||||
|
username: null
|
||||||
|
password: null
|
||||||
|
|
||||||
# Maximum number of inbox posts to do at once
|
mysql:
|
||||||
# If workers is set to 1 or above, this is the max for each worker
|
database: activityrelay
|
||||||
push_limit: 512
|
hostname: null
|
||||||
|
port: null
|
||||||
# The amount of json objects to cache from GET requests
|
username: null
|
||||||
json_cache: 1024
|
password: null
|
||||||
|
|
||||||
ap:
|
|
||||||
# This is used for generating activitypub messages, as well as instructions for
|
|
||||||
# linking AP identities. It should be an SSL-enabled domain reachable by https.
|
|
||||||
host: 'relay.example.com'
|
|
||||||
|
|
||||||
blocked_instances:
|
|
||||||
- 'bad-instance.example.com'
|
|
||||||
- 'another-bad-instance.example.com'
|
|
||||||
|
|
||||||
whitelist_enabled: false
|
|
||||||
|
|
||||||
whitelist:
|
|
||||||
- 'good-instance.example.com'
|
|
||||||
- 'another.good-instance.example.com'
|
|
||||||
|
|
||||||
# uncomment the lines below to prevent certain activitypub software from posting
|
|
||||||
# to the relay (all known relays by default). this uses the software name in nodeinfo
|
|
||||||
#blocked_software:
|
|
||||||
#- 'activityrelay'
|
|
||||||
#- 'aoderelay'
|
|
||||||
#- 'social.seattle.wa.us-relay'
|
|
||||||
#- 'unciarelay'
|
|
||||||
|
|
|
@ -1,4 +1,5 @@
|
||||||
import asyncio
|
import asyncio
|
||||||
|
import inspect
|
||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import queue
|
import queue
|
||||||
|
@ -7,49 +8,41 @@ import threading
|
||||||
import traceback
|
import traceback
|
||||||
|
|
||||||
from aiohttp import web
|
from aiohttp import web
|
||||||
|
from aputils import Signer
|
||||||
from datetime import datetime, timedelta
|
from datetime import datetime, timedelta
|
||||||
|
|
||||||
from .config import RelayConfig
|
from .config import Config
|
||||||
from .database import RelayDatabase
|
from .database import Database
|
||||||
from .http_client import HttpClient
|
from .http_client import HttpClient
|
||||||
|
from .logger import set_level
|
||||||
from .misc import DotDict, check_open_port, set_app
|
from .misc import DotDict, check_open_port, set_app
|
||||||
from .views import routes
|
from .views import routes
|
||||||
|
|
||||||
|
|
||||||
class Application(web.Application):
|
class Application(web.Application):
|
||||||
def __init__(self, cfgpath):
|
def __init__(self, cfgpath):
|
||||||
web.Application.__init__(self)
|
web.Application.__init__(self,
|
||||||
|
middlewares = [
|
||||||
self['starttime'] = None
|
server_middleware
|
||||||
self['running'] = False
|
]
|
||||||
self['config'] = RelayConfig(cfgpath)
|
)
|
||||||
|
|
||||||
if not self['config'].load():
|
|
||||||
self['config'].save()
|
|
||||||
|
|
||||||
if self.config.is_docker:
|
|
||||||
self.config.update({
|
|
||||||
'db': '/data/relay.jsonld',
|
|
||||||
'listen': '0.0.0.0',
|
|
||||||
'port': 8080
|
|
||||||
})
|
|
||||||
|
|
||||||
self['workers'] = []
|
|
||||||
self['last_worker'] = 0
|
|
||||||
|
|
||||||
set_app(self)
|
set_app(self)
|
||||||
|
|
||||||
self['database'] = RelayDatabase(self['config'])
|
self['config'] = Config(cfgpath)
|
||||||
self['database'].load()
|
self['database'] = Database(**self.config.dbconfig)
|
||||||
|
self['client'] = HttpClient()
|
||||||
|
|
||||||
self['client'] = HttpClient(
|
self['starttime'] = None
|
||||||
database = self.database,
|
self['signer'] = None
|
||||||
limit = self.config.push_limit,
|
self['running'] = False
|
||||||
timeout = self.config.timeout,
|
self['workers'] = []
|
||||||
cache_size = self.config.json_cache
|
self['last_worker'] = 0
|
||||||
)
|
|
||||||
|
|
||||||
self.set_signal_handler()
|
self.database.create()
|
||||||
|
|
||||||
|
with self.database.session as s:
|
||||||
|
set_level(s.get_config('log_level'))
|
||||||
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
|
@ -67,18 +60,32 @@ class Application(web.Application):
|
||||||
return self['database']
|
return self['database']
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def signer(self):
|
||||||
|
if not self['signer']:
|
||||||
|
with self.database.session as s:
|
||||||
|
privkey = s.get_config('privkey')
|
||||||
|
|
||||||
|
if not privkey:
|
||||||
|
self['signer'] = Signer.new(self.config.keyid)
|
||||||
|
s.put_config('privkey', self['signer'].export())
|
||||||
|
|
||||||
|
else:
|
||||||
|
self['signer'] = Signer(privkey, self.config.keyid)
|
||||||
|
|
||||||
|
return self['signer']
|
||||||
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def uptime(self):
|
def uptime(self):
|
||||||
if not self['starttime']:
|
if not self['starttime']:
|
||||||
return timedelta(seconds=0)
|
return timedelta(seconds=0)
|
||||||
|
|
||||||
uptime = datetime.now() - self['starttime']
|
return datetime.now() - self['starttime']
|
||||||
|
|
||||||
return timedelta(seconds=uptime.seconds)
|
|
||||||
|
|
||||||
|
|
||||||
def push_message(self, inbox, message):
|
def push_message(self, inbox, message):
|
||||||
if self.config.workers <= 0:
|
if len(self['workers']) <= 0:
|
||||||
return asyncio.ensure_future(self.client.post(inbox, message))
|
return asyncio.ensure_future(self.client.post(inbox, message))
|
||||||
|
|
||||||
worker = self['workers'][self['last_worker']]
|
worker = self['workers'][self['last_worker']]
|
||||||
|
@ -90,10 +97,10 @@ class Application(web.Application):
|
||||||
self['last_worker'] = 0
|
self['last_worker'] = 0
|
||||||
|
|
||||||
|
|
||||||
def set_signal_handler(self):
|
def set_signal_handler(self, enable=True):
|
||||||
for sig in {'SIGHUP', 'SIGINT', 'SIGQUIT', 'SIGTERM'}:
|
for sig in {'SIGHUP', 'SIGINT', 'SIGQUIT', 'SIGTERM'}:
|
||||||
try:
|
try:
|
||||||
signal.signal(getattr(signal, sig), self.stop)
|
signal.signal(getattr(signal, sig), self.stop if enable else signal.SIG_DFL)
|
||||||
|
|
||||||
# some signals don't exist in windows, so skip them
|
# some signals don't exist in windows, so skip them
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
|
@ -109,21 +116,30 @@ class Application(web.Application):
|
||||||
|
|
||||||
logging.info(f'Starting webserver at {self.config.host} ({self.config.listen}:{self.config.port})')
|
logging.info(f'Starting webserver at {self.config.host} ({self.config.listen}:{self.config.port})')
|
||||||
asyncio.run(self.handle_run())
|
asyncio.run(self.handle_run())
|
||||||
|
self.database.disconnect()
|
||||||
|
|
||||||
|
|
||||||
def stop(self, *_):
|
def stop(self, *_):
|
||||||
self['running'] = False
|
self['running'] = False
|
||||||
|
|
||||||
|
|
||||||
|
def setup(self):
|
||||||
|
self.client.setup()
|
||||||
|
|
||||||
|
|
||||||
async def handle_run(self):
|
async def handle_run(self):
|
||||||
|
self.set_signal_handler(True)
|
||||||
self['running'] = True
|
self['running'] = True
|
||||||
|
|
||||||
if self.config.workers > 0:
|
with self.database.session as s:
|
||||||
for i in range(self.config.workers):
|
workers = s.get_config('workers')
|
||||||
worker = PushWorker(self)
|
|
||||||
worker.start()
|
|
||||||
|
|
||||||
self['workers'].append(worker)
|
if workers > 0:
|
||||||
|
for i in range(workers):
|
||||||
|
worker = PushWorker(self)
|
||||||
|
worker.start()
|
||||||
|
|
||||||
|
self['workers'].append(worker)
|
||||||
|
|
||||||
runner = web.AppRunner(self, access_log_format='%{X-Forwarded-For}i "%r" %s %b "%{User-Agent}i"')
|
runner = web.AppRunner(self, access_log_format='%{X-Forwarded-For}i "%r" %s %b "%{User-Agent}i"')
|
||||||
await runner.setup()
|
await runner.setup()
|
||||||
|
@ -145,6 +161,7 @@ class Application(web.Application):
|
||||||
self['starttime'] = None
|
self['starttime'] = None
|
||||||
self['running'] = False
|
self['running'] = False
|
||||||
self['workers'].clear()
|
self['workers'].clear()
|
||||||
|
self.set_signal_handler(False)
|
||||||
|
|
||||||
|
|
||||||
class PushWorker(threading.Thread):
|
class PushWorker(threading.Thread):
|
||||||
|
@ -155,12 +172,8 @@ class PushWorker(threading.Thread):
|
||||||
|
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
self.client = HttpClient(
|
self.client = HttpClient()
|
||||||
database = self.app.database,
|
self.client.setup()
|
||||||
limit = self.app.config.push_limit,
|
|
||||||
timeout = self.app.config.timeout,
|
|
||||||
cache_size = self.app.config.json_cache
|
|
||||||
)
|
|
||||||
|
|
||||||
asyncio.run(self.handle_queue())
|
asyncio.run(self.handle_queue())
|
||||||
|
|
||||||
|
@ -183,6 +196,24 @@ class PushWorker(threading.Thread):
|
||||||
await self.client.close()
|
await self.client.close()
|
||||||
|
|
||||||
|
|
||||||
|
@web.middleware
|
||||||
|
async def server_middleware(request, handler):
|
||||||
|
if len(inspect.signature(handler).parameters) == 1:
|
||||||
|
response = await handler(request)
|
||||||
|
|
||||||
|
else:
|
||||||
|
with request.database.session as s:
|
||||||
|
response = await handler(request, s)
|
||||||
|
|
||||||
|
## make sure there's some sort of response
|
||||||
|
if response == None:
|
||||||
|
logging.error(f'No response for handler: {handler}')
|
||||||
|
response = Response.new_error(500, 'No response')
|
||||||
|
|
||||||
|
response.headers['Server'] = 'ActivityRelay'
|
||||||
|
return response
|
||||||
|
|
||||||
|
|
||||||
## Can't sub-class web.Request, so let's just add some properties
|
## Can't sub-class web.Request, so let's just add some properties
|
||||||
def request_actor(self):
|
def request_actor(self):
|
||||||
try: return self['actor']
|
try: return self['actor']
|
||||||
|
|
319
relay/config.py
319
relay/config.py
|
@ -1,58 +1,135 @@
|
||||||
import json
|
|
||||||
import os
|
import os
|
||||||
|
import sys
|
||||||
import yaml
|
import yaml
|
||||||
|
|
||||||
from functools import cached_property
|
from functools import cached_property
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from urllib.parse import urlparse
|
from platform import system
|
||||||
|
|
||||||
from .misc import DotDict, boolean
|
from .misc import AppBase, DotDict
|
||||||
|
|
||||||
|
|
||||||
RELAY_SOFTWARE = [
|
DEFAULTS = {
|
||||||
'activityrelay', # https://git.pleroma.social/pleroma/relay
|
'general_listen': '0.0.0.0',
|
||||||
'aoderelay', # https://git.asonix.dog/asonix/relay
|
'general_port': 8080,
|
||||||
'feditools-relay' # https://git.ptzo.gdn/feditools/relay
|
'general_host': 'relay.example.com',
|
||||||
]
|
'database_type': 'sqlite',
|
||||||
|
'database_min_connections': 0,
|
||||||
|
'database_max_connections': 10,
|
||||||
|
'sqlite_database': Path('relay.sqlite3'),
|
||||||
|
'postgres_database': 'activityrelay',
|
||||||
|
'postgres_hostname': None,
|
||||||
|
'postgres_port': None,
|
||||||
|
'postgres_username': None,
|
||||||
|
'postgres_password': None,
|
||||||
|
'mysql_database': 'activityrelay',
|
||||||
|
'mysql_hostname': None,
|
||||||
|
'mysql_port': None,
|
||||||
|
'mysql_username': None,
|
||||||
|
'mysql_password': None
|
||||||
|
}
|
||||||
|
|
||||||
APKEYS = [
|
CATEGORY_NAMES = [
|
||||||
'host',
|
'general',
|
||||||
'whitelist_enabled',
|
'database',
|
||||||
'blocked_software',
|
'sqlite',
|
||||||
'blocked_instances',
|
'postgres',
|
||||||
'whitelist'
|
'mysql'
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
class RelayConfig(DotDict):
|
def get_config_dir():
|
||||||
def __init__(self, path):
|
cwd = Path.cwd().joinpath('config.yaml')
|
||||||
DotDict.__init__(self, {})
|
plat = system()
|
||||||
|
|
||||||
|
if cwd.exists():
|
||||||
|
return cwd
|
||||||
|
|
||||||
|
elif plat == 'Linux':
|
||||||
|
cfgpath = Path('~/.config/activityrelay/config.yaml').expanduser()
|
||||||
|
|
||||||
|
if cfgpath.exists():
|
||||||
|
return cfgpath
|
||||||
|
|
||||||
|
etcpath = Path('/etc/activityrelay/config.yaml')
|
||||||
|
|
||||||
|
if etcpath.exists() and os.getuid() == etcpath.stat().st_uid:
|
||||||
|
return etcpath
|
||||||
|
|
||||||
|
elif plat == 'Windows':
|
||||||
|
cfgpath = Path('~/AppData/Roaming/activityrelay/config.yaml').expanduer()
|
||||||
|
|
||||||
|
if cfgpath.exists():
|
||||||
|
return cfgpath
|
||||||
|
|
||||||
|
elif plat == 'Darwin':
|
||||||
|
cfgpath = Path('~/Library/Application Support/activityaelay/config.yaml')
|
||||||
|
|
||||||
|
return cwd
|
||||||
|
|
||||||
|
|
||||||
|
class Config(AppBase, dict):
|
||||||
|
def __init__(self, path=None):
|
||||||
|
DotDict.__init__(self, DEFAULTS)
|
||||||
|
|
||||||
if self.is_docker:
|
if self.is_docker:
|
||||||
path = '/data/config.yaml'
|
path = Path('/data/config.yaml')
|
||||||
|
|
||||||
self._path = Path(path).expanduser()
|
elif not path:
|
||||||
self.reset()
|
path = get_config_dir()
|
||||||
|
|
||||||
|
else:
|
||||||
|
path = Path(path).expanduser()
|
||||||
|
|
||||||
|
self._path = path
|
||||||
|
self.load()
|
||||||
|
|
||||||
|
|
||||||
def __setitem__(self, key, value):
|
def __setitem__(self, key, value):
|
||||||
if key in ['blocked_instances', 'blocked_software', 'whitelist']:
|
if key in {'database', 'hostname', 'port', 'username', 'password'}:
|
||||||
assert isinstance(value, (list, set, tuple))
|
key = f'{self.dbtype}_{key}'
|
||||||
|
|
||||||
elif key in ['port', 'workers', 'json_cache', 'timeout']:
|
if (self.is_docker and key in {'general_host', 'general_port'}) or value == '__DEFAULT__':
|
||||||
if not isinstance(value, int):
|
value = DEFAULTS[key]
|
||||||
value = int(value)
|
|
||||||
|
|
||||||
elif key == 'whitelist_enabled':
|
elif key in {'general_port', 'database_min_connections', 'database_max_connections'}:
|
||||||
if not isinstance(value, bool):
|
value = int(value)
|
||||||
value = boolean(value)
|
|
||||||
|
|
||||||
super().__setitem__(key, value)
|
elif key == 'sqlite_database':
|
||||||
|
if not isinstance(value, Path):
|
||||||
|
value = Path(value)
|
||||||
|
|
||||||
|
dict.__setitem__(self, key, value)
|
||||||
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def db(self):
|
def dbconfig(self):
|
||||||
return Path(self['db']).expanduser().resolve()
|
config = {
|
||||||
|
'type': self['database_type'],
|
||||||
|
'min_conn': self['database_min_connections'],
|
||||||
|
'max_conn': self['database_max_connections']
|
||||||
|
}
|
||||||
|
|
||||||
|
if self.dbtype == 'sqlite':
|
||||||
|
if not self['sqlite_database'].is_absolute():
|
||||||
|
config['database'] = self.path.with_name(str(self['sqlite_database'])).resolve()
|
||||||
|
|
||||||
|
else:
|
||||||
|
config['database'] = self['sqlite_database'].resolve()
|
||||||
|
|
||||||
|
else:
|
||||||
|
for key, value in self.items():
|
||||||
|
cat, name = key.split('_', 1)
|
||||||
|
|
||||||
|
if self.dbtype == cat:
|
||||||
|
config[name] = value
|
||||||
|
|
||||||
|
return config
|
||||||
|
|
||||||
|
|
||||||
|
@cached_property
|
||||||
|
def is_docker(self):
|
||||||
|
return bool(os.getenv('DOCKER_RUNNING'))
|
||||||
|
|
||||||
|
|
||||||
@property
|
@property
|
||||||
|
@ -60,6 +137,29 @@ class RelayConfig(DotDict):
|
||||||
return self._path
|
return self._path
|
||||||
|
|
||||||
|
|
||||||
|
## General config
|
||||||
|
@property
|
||||||
|
def host(self):
|
||||||
|
return self['general_host']
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def listen(self):
|
||||||
|
return self['general_listen']
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def port(self):
|
||||||
|
return self['general_port']
|
||||||
|
|
||||||
|
|
||||||
|
## Database config
|
||||||
|
@property
|
||||||
|
def dbtype(self):
|
||||||
|
return self['database_type']
|
||||||
|
|
||||||
|
|
||||||
|
## AP URLs
|
||||||
@property
|
@property
|
||||||
def actor(self):
|
def actor(self):
|
||||||
return f'https://{self.host}/actor'
|
return f'https://{self.host}/actor'
|
||||||
|
@ -75,117 +175,12 @@ class RelayConfig(DotDict):
|
||||||
return f'{self.actor}#main-key'
|
return f'{self.actor}#main-key'
|
||||||
|
|
||||||
|
|
||||||
@cached_property
|
|
||||||
def is_docker(self):
|
|
||||||
return bool(os.environ.get('DOCKER_RUNNING'))
|
|
||||||
|
|
||||||
|
|
||||||
def reset(self):
|
def reset(self):
|
||||||
self.clear()
|
self.clear()
|
||||||
self.update({
|
self.update(DEFAULTS)
|
||||||
'db': str(self._path.parent.joinpath(f'{self._path.stem}.jsonld')),
|
|
||||||
'listen': '0.0.0.0',
|
|
||||||
'port': 8080,
|
|
||||||
'note': 'Make a note about your instance here.',
|
|
||||||
'push_limit': 512,
|
|
||||||
'json_cache': 1024,
|
|
||||||
'timeout': 10,
|
|
||||||
'workers': 0,
|
|
||||||
'host': 'relay.example.com',
|
|
||||||
'whitelist_enabled': False,
|
|
||||||
'blocked_software': [],
|
|
||||||
'blocked_instances': [],
|
|
||||||
'whitelist': []
|
|
||||||
})
|
|
||||||
|
|
||||||
|
|
||||||
def ban_instance(self, instance):
|
|
||||||
if instance.startswith('http'):
|
|
||||||
instance = urlparse(instance).hostname
|
|
||||||
|
|
||||||
if self.is_banned(instance):
|
|
||||||
return False
|
|
||||||
|
|
||||||
self.blocked_instances.append(instance)
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def unban_instance(self, instance):
|
|
||||||
if instance.startswith('http'):
|
|
||||||
instance = urlparse(instance).hostname
|
|
||||||
|
|
||||||
try:
|
|
||||||
self.blocked_instances.remove(instance)
|
|
||||||
return True
|
|
||||||
|
|
||||||
except:
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def ban_software(self, software):
|
|
||||||
if self.is_banned_software(software):
|
|
||||||
return False
|
|
||||||
|
|
||||||
self.blocked_software.append(software)
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def unban_software(self, software):
|
|
||||||
try:
|
|
||||||
self.blocked_software.remove(software)
|
|
||||||
return True
|
|
||||||
|
|
||||||
except:
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def add_whitelist(self, instance):
|
|
||||||
if instance.startswith('http'):
|
|
||||||
instance = urlparse(instance).hostname
|
|
||||||
|
|
||||||
if self.is_whitelisted(instance):
|
|
||||||
return False
|
|
||||||
|
|
||||||
self.whitelist.append(instance)
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def del_whitelist(self, instance):
|
|
||||||
if instance.startswith('http'):
|
|
||||||
instance = urlparse(instance).hostname
|
|
||||||
|
|
||||||
try:
|
|
||||||
self.whitelist.remove(instance)
|
|
||||||
return True
|
|
||||||
|
|
||||||
except:
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def is_banned(self, instance):
|
|
||||||
if instance.startswith('http'):
|
|
||||||
instance = urlparse(instance).hostname
|
|
||||||
|
|
||||||
return instance in self.blocked_instances
|
|
||||||
|
|
||||||
|
|
||||||
def is_banned_software(self, software):
|
|
||||||
if not software:
|
|
||||||
return False
|
|
||||||
|
|
||||||
return software.lower() in self.blocked_software
|
|
||||||
|
|
||||||
|
|
||||||
def is_whitelisted(self, instance):
|
|
||||||
if instance.startswith('http'):
|
|
||||||
instance = urlparse(instance).hostname
|
|
||||||
|
|
||||||
return instance in self.whitelist
|
|
||||||
|
|
||||||
|
|
||||||
def load(self):
|
def load(self):
|
||||||
self.reset()
|
|
||||||
|
|
||||||
options = {}
|
options = {}
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
@ -201,45 +196,21 @@ class RelayConfig(DotDict):
|
||||||
except FileNotFoundError:
|
except FileNotFoundError:
|
||||||
return False
|
return False
|
||||||
|
|
||||||
if not config:
|
for key, value in DEFAULTS.items():
|
||||||
return False
|
cat, name = key.split('_', 1)
|
||||||
|
self[key] = config.get(cat, {}).get(name, DEFAULTS[key])
|
||||||
for key, value in config.items():
|
|
||||||
if key in ['ap']:
|
|
||||||
for k, v in value.items():
|
|
||||||
if k not in self:
|
|
||||||
continue
|
|
||||||
|
|
||||||
self[k] = v
|
|
||||||
|
|
||||||
continue
|
|
||||||
|
|
||||||
elif key not in self:
|
|
||||||
continue
|
|
||||||
|
|
||||||
self[key] = value
|
|
||||||
|
|
||||||
if self.host.endswith('example.com'):
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def save(self):
|
def save(self):
|
||||||
config = {
|
config = {key: {} for key in CATEGORY_NAMES}
|
||||||
# just turning config.db into a string is good enough for now
|
|
||||||
'db': str(self.db),
|
|
||||||
'listen': self.listen,
|
|
||||||
'port': self.port,
|
|
||||||
'note': self.note,
|
|
||||||
'push_limit': self.push_limit,
|
|
||||||
'workers': self.workers,
|
|
||||||
'json_cache': self.json_cache,
|
|
||||||
'timeout': self.timeout,
|
|
||||||
'ap': {key: self[key] for key in APKEYS}
|
|
||||||
}
|
|
||||||
|
|
||||||
with open(self._path, 'w') as fd:
|
for key, value in self.items():
|
||||||
|
cat, name = key.split('_', 1)
|
||||||
|
|
||||||
|
if isinstance(value, Path):
|
||||||
|
value = str(value)
|
||||||
|
|
||||||
|
config[cat][name] = value
|
||||||
|
|
||||||
|
with open(self.path, 'w') as fd:
|
||||||
yaml.dump(config, fd, sort_keys=False)
|
yaml.dump(config, fd, sort_keys=False)
|
||||||
|
|
||||||
return config
|
|
||||||
|
|
|
@ -1,197 +0,0 @@
|
||||||
import aputils
|
|
||||||
import asyncio
|
|
||||||
import json
|
|
||||||
import logging
|
|
||||||
import traceback
|
|
||||||
|
|
||||||
from urllib.parse import urlparse
|
|
||||||
|
|
||||||
|
|
||||||
class RelayDatabase(dict):
|
|
||||||
def __init__(self, config):
|
|
||||||
dict.__init__(self, {
|
|
||||||
'relay-list': {},
|
|
||||||
'private-key': None,
|
|
||||||
'follow-requests': {},
|
|
||||||
'version': 1
|
|
||||||
})
|
|
||||||
|
|
||||||
self.config = config
|
|
||||||
self.signer = None
|
|
||||||
|
|
||||||
|
|
||||||
@property
|
|
||||||
def hostnames(self):
|
|
||||||
return tuple(self['relay-list'].keys())
|
|
||||||
|
|
||||||
|
|
||||||
@property
|
|
||||||
def inboxes(self):
|
|
||||||
return tuple(data['inbox'] for data in self['relay-list'].values())
|
|
||||||
|
|
||||||
|
|
||||||
def load(self):
|
|
||||||
new_db = True
|
|
||||||
|
|
||||||
try:
|
|
||||||
with self.config.db.open() as fd:
|
|
||||||
data = json.load(fd)
|
|
||||||
|
|
||||||
self['version'] = data.get('version', None)
|
|
||||||
self['private-key'] = data.get('private-key')
|
|
||||||
|
|
||||||
if self['version'] == None:
|
|
||||||
self['version'] = 1
|
|
||||||
|
|
||||||
if 'actorKeys' in data:
|
|
||||||
self['private-key'] = data['actorKeys']['privateKey']
|
|
||||||
|
|
||||||
for item in data.get('relay-list', []):
|
|
||||||
domain = urlparse(item).hostname
|
|
||||||
self['relay-list'][domain] = {
|
|
||||||
'domain': domain,
|
|
||||||
'inbox': item,
|
|
||||||
'followid': None
|
|
||||||
}
|
|
||||||
|
|
||||||
else:
|
|
||||||
self['relay-list'] = data.get('relay-list', {})
|
|
||||||
|
|
||||||
for domain, instance in self['relay-list'].items():
|
|
||||||
if self.config.is_banned(domain) or (self.config.whitelist_enabled and not self.config.is_whitelisted(domain)):
|
|
||||||
self.del_inbox(domain)
|
|
||||||
continue
|
|
||||||
|
|
||||||
if not instance.get('domain'):
|
|
||||||
instance['domain'] = domain
|
|
||||||
|
|
||||||
new_db = False
|
|
||||||
|
|
||||||
except FileNotFoundError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
except json.decoder.JSONDecodeError as e:
|
|
||||||
if self.config.db.stat().st_size > 0:
|
|
||||||
raise e from None
|
|
||||||
|
|
||||||
if not self['private-key']:
|
|
||||||
logging.info("No actor keys present, generating 4096-bit RSA keypair.")
|
|
||||||
self.signer = aputils.Signer.new(self.config.keyid, size=4096)
|
|
||||||
self['private-key'] = self.signer.export()
|
|
||||||
|
|
||||||
else:
|
|
||||||
self.signer = aputils.Signer(self['private-key'], self.config.keyid)
|
|
||||||
|
|
||||||
self.save()
|
|
||||||
return not new_db
|
|
||||||
|
|
||||||
|
|
||||||
def save(self):
|
|
||||||
with self.config.db.open('w') as fd:
|
|
||||||
json.dump(self, fd, indent=4)
|
|
||||||
|
|
||||||
|
|
||||||
def get_inbox(self, domain, fail=False):
|
|
||||||
if domain.startswith('http'):
|
|
||||||
domain = urlparse(domain).hostname
|
|
||||||
|
|
||||||
inbox = self['relay-list'].get(domain)
|
|
||||||
|
|
||||||
if inbox:
|
|
||||||
return inbox
|
|
||||||
|
|
||||||
if fail:
|
|
||||||
raise KeyError(domain)
|
|
||||||
|
|
||||||
|
|
||||||
def add_inbox(self, inbox, followid=None, software=None):
|
|
||||||
assert inbox.startswith('https'), 'Inbox must be a url'
|
|
||||||
domain = urlparse(inbox).hostname
|
|
||||||
instance = self.get_inbox(domain)
|
|
||||||
|
|
||||||
if instance:
|
|
||||||
if followid:
|
|
||||||
instance['followid'] = followid
|
|
||||||
|
|
||||||
if software:
|
|
||||||
instance['software'] = software
|
|
||||||
|
|
||||||
return instance
|
|
||||||
|
|
||||||
self['relay-list'][domain] = {
|
|
||||||
'domain': domain,
|
|
||||||
'inbox': inbox,
|
|
||||||
'followid': followid,
|
|
||||||
'software': software
|
|
||||||
}
|
|
||||||
|
|
||||||
logging.verbose(f'Added inbox to database: {inbox}')
|
|
||||||
return self['relay-list'][domain]
|
|
||||||
|
|
||||||
|
|
||||||
def del_inbox(self, domain, followid=None, fail=False):
|
|
||||||
data = self.get_inbox(domain, fail=False)
|
|
||||||
|
|
||||||
if not data:
|
|
||||||
if fail:
|
|
||||||
raise KeyError(domain)
|
|
||||||
|
|
||||||
return False
|
|
||||||
|
|
||||||
if not data['followid'] or not followid or data['followid'] == followid:
|
|
||||||
del self['relay-list'][data['domain']]
|
|
||||||
logging.verbose(f'Removed inbox from database: {data["inbox"]}')
|
|
||||||
return True
|
|
||||||
|
|
||||||
if fail:
|
|
||||||
raise ValueError('Follow IDs do not match')
|
|
||||||
|
|
||||||
logging.debug(f'Follow ID does not match: db = {data["followid"]}, object = {followid}')
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def get_request(self, domain, fail=True):
|
|
||||||
if domain.startswith('http'):
|
|
||||||
domain = urlparse(domain).hostname
|
|
||||||
|
|
||||||
try:
|
|
||||||
return self['follow-requests'][domain]
|
|
||||||
|
|
||||||
except KeyError as e:
|
|
||||||
if fail:
|
|
||||||
raise e
|
|
||||||
|
|
||||||
|
|
||||||
def add_request(self, actor, inbox, followid):
|
|
||||||
domain = urlparse(inbox).hostname
|
|
||||||
|
|
||||||
try:
|
|
||||||
request = self.get_request(domain)
|
|
||||||
request['followid'] = followid
|
|
||||||
|
|
||||||
except KeyError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
self['follow-requests'][domain] = {
|
|
||||||
'actor': actor,
|
|
||||||
'inbox': inbox,
|
|
||||||
'followid': followid
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def del_request(self, domain):
|
|
||||||
if domain.startswith('http'):
|
|
||||||
domain = urlparse(inbox).hostname
|
|
||||||
|
|
||||||
del self['follow-requests'][domain]
|
|
||||||
|
|
||||||
|
|
||||||
def distill_inboxes(self, message):
|
|
||||||
src_domains = {
|
|
||||||
message.domain,
|
|
||||||
urlparse(message.objectid).netloc
|
|
||||||
}
|
|
||||||
|
|
||||||
for domain, instance in self['relay-list'].items():
|
|
||||||
if domain not in src_domains:
|
|
||||||
yield instance['inbox']
|
|
17
relay/database/__init__.py
Normal file
17
relay/database/__init__.py
Normal file
|
@ -0,0 +1,17 @@
|
||||||
|
import tinysql
|
||||||
|
|
||||||
|
from .base import DEFAULT_CONFIG, RELAY_SOFTWARE, TABLES
|
||||||
|
from .connection import Connection
|
||||||
|
from .rows import ROWS
|
||||||
|
|
||||||
|
|
||||||
|
class Database(tinysql.Database):
|
||||||
|
def __init__(self, **config):
|
||||||
|
tinysql.Database.__init__(self, **config,
|
||||||
|
connection_class = Connection,
|
||||||
|
row_classes = ROWS
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def create(self):
|
||||||
|
self.create_database(TABLES)
|
67
relay/database/base.py
Normal file
67
relay/database/base.py
Normal file
|
@ -0,0 +1,67 @@
|
||||||
|
from tinysql import Column, Table
|
||||||
|
|
||||||
|
|
||||||
|
DEFAULT_CONFIG = {
|
||||||
|
'description': ('str', 'Make a note about your relay here'),
|
||||||
|
'http_timeout': ('int', 10),
|
||||||
|
'json_cache': ('int', 1024),
|
||||||
|
'log_level': ('str', 'INFO'),
|
||||||
|
'name': ('str', 'ActivityRelay'),
|
||||||
|
'privkey': ('str', ''),
|
||||||
|
'push_limit': ('int', 512),
|
||||||
|
'require_approval': ('bool', False),
|
||||||
|
'version': ('int', 20221211),
|
||||||
|
'whitelist': ('bool', False),
|
||||||
|
'workers': ('int', 8)
|
||||||
|
}
|
||||||
|
|
||||||
|
RELAY_SOFTWARE = [
|
||||||
|
'activity-relay', # https://github.com/yukimochi/Activity-Relay
|
||||||
|
'activityrelay', # https://git.pleroma.social/pleroma/relay
|
||||||
|
'aoderelay', # https://git.asonix.dog/asonix/relay
|
||||||
|
'feditools-relay' # https://git.ptzo.gdn/feditools/relay
|
||||||
|
]
|
||||||
|
|
||||||
|
TABLES = [
|
||||||
|
Table('config',
|
||||||
|
Column('key', 'text', unique=True, nullable=False, primary_key=True),
|
||||||
|
Column('value', 'text')
|
||||||
|
),
|
||||||
|
Table('instances',
|
||||||
|
Column('id', 'serial'),
|
||||||
|
Column('domain', 'text', unique=True, nullable=False),
|
||||||
|
Column('actor', 'text'),
|
||||||
|
Column('inbox', 'text', nullable=False),
|
||||||
|
Column('followid', 'text'),
|
||||||
|
Column('software', 'text'),
|
||||||
|
Column('note', 'text'),
|
||||||
|
Column('joined', 'datetime', nullable=False),
|
||||||
|
Column('updated', 'datetime')
|
||||||
|
),
|
||||||
|
Table('whitelist',
|
||||||
|
Column('id', 'serial'),
|
||||||
|
Column('domain', 'text', unique=True),
|
||||||
|
Column('created', 'datetime', nullable=False)
|
||||||
|
),
|
||||||
|
Table('bans',
|
||||||
|
Column('id', 'serial'),
|
||||||
|
Column('name', 'text', unique=True),
|
||||||
|
Column('note', 'text'),
|
||||||
|
Column('type', 'text', nullable=False),
|
||||||
|
Column('created', 'datetime', nullable=False)
|
||||||
|
),
|
||||||
|
Table('users',
|
||||||
|
Column('id', 'serial'),
|
||||||
|
Column('handle', 'text', unique=True, nullable=False),
|
||||||
|
Column('domain', 'text', nullable=False),
|
||||||
|
Column('api_token', 'text'),
|
||||||
|
Column('created', 'datetime', nullable=False),
|
||||||
|
Column('updated', 'datetime')
|
||||||
|
),
|
||||||
|
Table('tokens',
|
||||||
|
Column('id', 'text', unique=True, nullable=False, primary_key=True),
|
||||||
|
Column('userid', 'integer', nullable=False),
|
||||||
|
Column('created', 'datetime', nullable=False),
|
||||||
|
Column('updated', 'datetime')
|
||||||
|
)
|
||||||
|
]
|
239
relay/database/connection.py
Normal file
239
relay/database/connection.py
Normal file
|
@ -0,0 +1,239 @@
|
||||||
|
import tinysql
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
|
from .base import DEFAULT_CONFIG
|
||||||
|
from ..misc import DotDict
|
||||||
|
|
||||||
|
|
||||||
|
class Connection(tinysql.ConnectionMixin):
|
||||||
|
## Misc methods
|
||||||
|
def accept_request(self, domain):
|
||||||
|
row = self.get_request(domain)
|
||||||
|
|
||||||
|
if not row:
|
||||||
|
raise KeyError(domain)
|
||||||
|
|
||||||
|
data = {'joined': datetime.now()}
|
||||||
|
self.update('instances', data, id=row.id)
|
||||||
|
|
||||||
|
|
||||||
|
def distill_inboxes(self, message):
|
||||||
|
src_domains = {
|
||||||
|
message.domain,
|
||||||
|
urlparse(message.objectid).netloc
|
||||||
|
}
|
||||||
|
|
||||||
|
for instance in self.get_instances():
|
||||||
|
if instance.domain not in src_domains:
|
||||||
|
yield instance.inbox
|
||||||
|
|
||||||
|
|
||||||
|
## Delete methods
|
||||||
|
def delete_ban(self, type, name):
|
||||||
|
row = self.get_ban(type, name)
|
||||||
|
|
||||||
|
if not row:
|
||||||
|
raise KeyError(name)
|
||||||
|
|
||||||
|
self.delete('bans', id=row.id)
|
||||||
|
|
||||||
|
|
||||||
|
def delete_instance(self, domain):
|
||||||
|
row = self.get_instance(domain)
|
||||||
|
|
||||||
|
if not row:
|
||||||
|
raise KeyError(domain)
|
||||||
|
|
||||||
|
self.delete('instances', id=row.id)
|
||||||
|
|
||||||
|
|
||||||
|
def delete_whitelist(self, domain):
|
||||||
|
row = self.get_whitelist_domain(domain)
|
||||||
|
|
||||||
|
if not row:
|
||||||
|
raise KeyError(domain)
|
||||||
|
|
||||||
|
self.delete('whitelist', id=row.id)
|
||||||
|
|
||||||
|
|
||||||
|
## Get methods
|
||||||
|
def get_ban(self, type, name):
|
||||||
|
if type not in {'software', 'domain'}:
|
||||||
|
raise ValueError('Ban type must be "software" or "domain"')
|
||||||
|
|
||||||
|
return self.select('bans', name=name, type=type).one()
|
||||||
|
|
||||||
|
|
||||||
|
def get_bans(self, type):
|
||||||
|
if type not in {'software', 'domain'}:
|
||||||
|
raise ValueError('Ban type must be "software" or "domain"')
|
||||||
|
|
||||||
|
return self.select('bans', type=type).all()
|
||||||
|
|
||||||
|
|
||||||
|
def get_config(self, key):
|
||||||
|
if key not in DEFAULT_CONFIG:
|
||||||
|
raise KeyError(key)
|
||||||
|
|
||||||
|
row = self.select('config', key=key).one()
|
||||||
|
|
||||||
|
if not row:
|
||||||
|
return DEFAULT_CONFIG[key][1]
|
||||||
|
|
||||||
|
return row.value
|
||||||
|
|
||||||
|
|
||||||
|
def get_config_all(self):
|
||||||
|
rows = self.select('config').all()
|
||||||
|
config = DotDict({row.key: row.value for row in rows})
|
||||||
|
|
||||||
|
for key, data in DEFAULT_CONFIG.items():
|
||||||
|
if key not in config:
|
||||||
|
config[key] = data[1]
|
||||||
|
|
||||||
|
return config
|
||||||
|
|
||||||
|
|
||||||
|
def get_hostnames(self):
|
||||||
|
return tuple(row.domain for row in self.get_instances())
|
||||||
|
|
||||||
|
|
||||||
|
def get_instance(self, data):
|
||||||
|
if data.startswith('http') and '#' in data:
|
||||||
|
data = data.split('#', 1)[0]
|
||||||
|
|
||||||
|
query = 'SELECT * FROM instances WHERE domain = :data OR actor = :data OR inbox = :data'
|
||||||
|
row = self.execute(query, dict(data=data), table='instances').one()
|
||||||
|
return row if row and row.joined else None
|
||||||
|
|
||||||
|
|
||||||
|
def get_instances(self):
|
||||||
|
query = 'SELECT * FROM instances WHERE joined IS NOT NULL'
|
||||||
|
query += ' ORDER BY domain ASC'
|
||||||
|
return self.execute(query, table='instances').all()
|
||||||
|
|
||||||
|
|
||||||
|
def get_request(self, domain):
|
||||||
|
for instance in self.get_requests():
|
||||||
|
if instance.domain == domain:
|
||||||
|
return instance
|
||||||
|
|
||||||
|
raise KeyError(domain)
|
||||||
|
|
||||||
|
|
||||||
|
def get_requests(self):
|
||||||
|
query = 'SELECT * FROM instances WHERE joined IS NULL ORDER BY domain ASC'
|
||||||
|
return self.execute(query, table='instances').all()
|
||||||
|
|
||||||
|
|
||||||
|
def get_whitelist(self):
|
||||||
|
return self.select('whitelist').all()
|
||||||
|
|
||||||
|
|
||||||
|
def get_whitelist_domain(self, domain):
|
||||||
|
return self.select('whitelist', domain=domain).one()
|
||||||
|
|
||||||
|
|
||||||
|
## Put methods
|
||||||
|
def put_ban(self, type, name, note=None):
|
||||||
|
if type not in {'software', 'domain'}:
|
||||||
|
raise ValueError('Ban type must be "software" or "domain"')
|
||||||
|
|
||||||
|
row = self.select('bans', name=name, type=type).one()
|
||||||
|
|
||||||
|
if row:
|
||||||
|
if note == None:
|
||||||
|
raise KeyError(name)
|
||||||
|
|
||||||
|
data = {'note': note}
|
||||||
|
self.update('bans', data, id=row.id)
|
||||||
|
return
|
||||||
|
|
||||||
|
self.insert('bans', {
|
||||||
|
'name': name,
|
||||||
|
'type': type,
|
||||||
|
'note': note,
|
||||||
|
'created': datetime.now()
|
||||||
|
})
|
||||||
|
|
||||||
|
|
||||||
|
def put_config(self, key, value='__DEFAULT__'):
|
||||||
|
if key not in DEFAULT_CONFIG:
|
||||||
|
raise KeyError(key)
|
||||||
|
|
||||||
|
if value == '__DEFAULT__':
|
||||||
|
value = DEFAULT_CONFIG[key][1]
|
||||||
|
|
||||||
|
elif key == 'log_level' and not getattr(logging, value.upper(), False):
|
||||||
|
raise KeyError(value)
|
||||||
|
|
||||||
|
row = self.select('config', key=key).one()
|
||||||
|
|
||||||
|
if row:
|
||||||
|
self.update('config', {'value': value}, key=key)
|
||||||
|
return
|
||||||
|
|
||||||
|
self.insert('config', {
|
||||||
|
'key': key,
|
||||||
|
'value': value
|
||||||
|
})
|
||||||
|
|
||||||
|
|
||||||
|
def put_instance(self, domain, actor=None, inbox=None, followid=None, software=None, actor_data=None, note=None, accept=True):
|
||||||
|
new_data = {
|
||||||
|
'actor': actor,
|
||||||
|
'inbox': inbox,
|
||||||
|
'followid': followid,
|
||||||
|
'software': software,
|
||||||
|
'note': note
|
||||||
|
}
|
||||||
|
|
||||||
|
if actor_data:
|
||||||
|
new_data['actor_data'] = dict(actor_data)
|
||||||
|
|
||||||
|
new_data = {key: value for key, value in new_data.items() if value != None}
|
||||||
|
instance = self.get_instance(domain)
|
||||||
|
|
||||||
|
if instance:
|
||||||
|
if not new_data:
|
||||||
|
raise KeyError(domain)
|
||||||
|
|
||||||
|
instance.update(new_data)
|
||||||
|
self.update('instances', new_data, id=instance.id)
|
||||||
|
return instance
|
||||||
|
|
||||||
|
if not inbox:
|
||||||
|
raise ValueError('Inbox must be included in instance data')
|
||||||
|
|
||||||
|
if accept:
|
||||||
|
new_data['joined'] = datetime.now()
|
||||||
|
|
||||||
|
new_data['domain'] = domain
|
||||||
|
|
||||||
|
self.insert('instances', new_data)
|
||||||
|
return self.get_instance(domain)
|
||||||
|
|
||||||
|
|
||||||
|
def put_instance_actor(self, actor, nodeinfo=None, accept=True):
|
||||||
|
data = {
|
||||||
|
'domain': actor.domain,
|
||||||
|
'actor': actor.id,
|
||||||
|
'inbox': actor.shared_inbox,
|
||||||
|
'actor_data': actor,
|
||||||
|
'accept': accept,
|
||||||
|
'software': nodeinfo.sw_name if nodeinfo else None
|
||||||
|
}
|
||||||
|
|
||||||
|
return self.put_instance(**data)
|
||||||
|
|
||||||
|
|
||||||
|
def put_whitelist(self, domain):
|
||||||
|
if self.get_whitelist_domain(domain):
|
||||||
|
raise KeyError(domain)
|
||||||
|
|
||||||
|
self.insert('whitelist', {
|
||||||
|
'domain': domain,
|
||||||
|
'created': datetime.now()
|
||||||
|
})
|
35
relay/database/rows.py
Normal file
35
relay/database/rows.py
Normal file
|
@ -0,0 +1,35 @@
|
||||||
|
import json
|
||||||
|
from tinysql import Row
|
||||||
|
from .base import DEFAULT_CONFIG
|
||||||
|
from ..misc import DotDict, boolean
|
||||||
|
|
||||||
|
|
||||||
|
ROWS = []
|
||||||
|
|
||||||
|
|
||||||
|
def register(cls):
|
||||||
|
ROWS.append(cls)
|
||||||
|
return cls
|
||||||
|
|
||||||
|
|
||||||
|
@register
|
||||||
|
class ConfigRow(Row):
|
||||||
|
__table__ = 'config'
|
||||||
|
|
||||||
|
@property
|
||||||
|
def value(self):
|
||||||
|
type = DEFAULT_CONFIG[self.key][0]
|
||||||
|
|
||||||
|
if type == 'int':
|
||||||
|
return int(self['value'])
|
||||||
|
|
||||||
|
elif type == 'bool':
|
||||||
|
return boolean(self['value'])
|
||||||
|
|
||||||
|
elif type == 'list':
|
||||||
|
return json.loads(self['value'])
|
||||||
|
|
||||||
|
elif type == 'json':
|
||||||
|
return DotDict.parse(self['value'])
|
||||||
|
|
||||||
|
return self['value']
|
|
@ -8,11 +8,14 @@ from aputils import Nodeinfo, WellKnownNodeinfo
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from cachetools import LRUCache
|
from cachetools import LRUCache
|
||||||
from json.decoder import JSONDecodeError
|
from json.decoder import JSONDecodeError
|
||||||
|
from urllib.error import HTTPError
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
|
from urllib.request import Request, urlopen
|
||||||
|
|
||||||
from . import __version__
|
from . import __version__
|
||||||
from .misc import (
|
from .misc import (
|
||||||
MIMETYPES,
|
MIMETYPES,
|
||||||
|
AppBase,
|
||||||
DotDict,
|
DotDict,
|
||||||
Message
|
Message
|
||||||
)
|
)
|
||||||
|
@ -29,9 +32,8 @@ class Cache(LRUCache):
|
||||||
self.__maxsize = int(value)
|
self.__maxsize = int(value)
|
||||||
|
|
||||||
|
|
||||||
class HttpClient:
|
class HttpClient(AppBase):
|
||||||
def __init__(self, database, limit=100, timeout=10, cache_size=1024):
|
def __init__(self, limit=100, timeout=10, cache_size=1024):
|
||||||
self.database = database
|
|
||||||
self.cache = Cache(cache_size)
|
self.cache = Cache(cache_size)
|
||||||
self.cfg = {'limit': limit, 'timeout': timeout}
|
self.cfg = {'limit': limit, 'timeout': timeout}
|
||||||
self._conn = None
|
self._conn = None
|
||||||
|
@ -57,8 +59,16 @@ class HttpClient:
|
||||||
return self.cfg['timeout']
|
return self.cfg['timeout']
|
||||||
|
|
||||||
|
|
||||||
|
def setup(self):
|
||||||
|
with self.database.session as s:
|
||||||
|
config = s.get_config_all()
|
||||||
|
self.client.cfg['limit'] = config.push_limit
|
||||||
|
self.client.cfg['timeout'] = config.http_timeout
|
||||||
|
self.client.cache.set_maxsize(config.json_cache)
|
||||||
|
|
||||||
|
|
||||||
async def open(self):
|
async def open(self):
|
||||||
if self._session:
|
if self._session and self._session._loop.is_running():
|
||||||
return
|
return
|
||||||
|
|
||||||
self._conn = TCPConnector(
|
self._conn = TCPConnector(
|
||||||
|
@ -97,7 +107,7 @@ class HttpClient:
|
||||||
headers = {}
|
headers = {}
|
||||||
|
|
||||||
if sign_headers:
|
if sign_headers:
|
||||||
headers.update(self.database.signer.sign_headers('GET', url, algorithm='original'))
|
headers.update(self.signer.sign_headers('GET', url, algorithm='original'))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
logging.verbose(f'Fetching resource: {url}')
|
logging.verbose(f'Fetching resource: {url}')
|
||||||
|
@ -144,37 +154,35 @@ class HttpClient:
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
|
|
||||||
|
|
||||||
async def post(self, url, message):
|
async def post(self, inbox, message):
|
||||||
await self.open()
|
await self.open()
|
||||||
|
|
||||||
instance = self.database.get_inbox(url)
|
with self.database.session as s:
|
||||||
|
instance = s.get_instance(inbox)
|
||||||
|
|
||||||
## Using the old algo by default is probably a better idea right now
|
## Using the old algo by default is probably a better idea right now
|
||||||
if instance and instance.get('software') in {'mastodon'}:
|
if instance and instance['software'] in {'mastodon'}:
|
||||||
algorithm = 'hs2019'
|
algorithm = 'hs2019'
|
||||||
|
|
||||||
else:
|
else:
|
||||||
algorithm = 'original'
|
algorithm = 'original'
|
||||||
|
|
||||||
headers = {'Content-Type': 'application/activity+json'}
|
headers = {'Content-Type': 'application/activity+json'}
|
||||||
headers.update(self.database.signer.sign_headers('POST', url, message, algorithm=algorithm))
|
headers.update(self.signer.sign_headers('POST', inbox, message, algorithm=algorithm))
|
||||||
|
|
||||||
try:
|
try:
|
||||||
logging.verbose(f'Sending "{message.type}" to {url}')
|
logging.verbose(f'Sending "{message.type}" to {inbox}')
|
||||||
|
|
||||||
async with self._session.post(url, headers=headers, data=message.to_json()) as resp:
|
async with self._session.post(inbox, headers=headers, data=message.to_json()) as resp:
|
||||||
## Not expecting a response, so just return
|
## Not expecting a response, so just return
|
||||||
if resp.status in {200, 202}:
|
if resp.status in {200, 202}:
|
||||||
return logging.verbose(f'Successfully sent "{message.type}" to {url}')
|
return logging.verbose(f'Successfully sent "{message.type}" to {inbox}')
|
||||||
|
|
||||||
logging.verbose(f'Received error when pushing to {url}: {resp.status}')
|
logging.verbose(f'Received error when pushing to {inbox}: {resp.status}')
|
||||||
return logging.verbose(await resp.read()) # change this to debug
|
return logging.verbose(await resp.read()) # change this to debug
|
||||||
|
|
||||||
except ClientSSLError:
|
except (ClientConnectorError, ServerTimeoutError):
|
||||||
logging.warning(f'SSL error when pushing to {urlparse(url).netloc}')
|
logging.verbose(f'Failed to connect to {inbox}')
|
||||||
|
|
||||||
except (AsyncTimeoutError, ClientConnectionError):
|
|
||||||
logging.warning(f'Failed to connect to {urlparse(url).netloc} for message push')
|
|
||||||
|
|
||||||
## prevent workers from being brought down
|
## prevent workers from being brought down
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
@ -207,16 +215,18 @@ class HttpClient:
|
||||||
return await self.get(nodeinfo_url, loads=Nodeinfo.new_from_json) or False
|
return await self.get(nodeinfo_url, loads=Nodeinfo.new_from_json) or False
|
||||||
|
|
||||||
|
|
||||||
async def get(database, *args, **kwargs):
|
## http client methods can't be called directly from manage.py,
|
||||||
async with HttpClient(database) as client:
|
## so here's some wrapper functions
|
||||||
|
async def get(*args, **kwargs):
|
||||||
|
async with HttpClient() as client:
|
||||||
return await client.get(*args, **kwargs)
|
return await client.get(*args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
async def post(database, *args, **kwargs):
|
async def post(*args, **kwargs):
|
||||||
async with HttpClient(database) as client:
|
async with HttpClient() as client:
|
||||||
return await client.post(*args, **kwargs)
|
return await client.post(*args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
async def fetch_nodeinfo(database, *args, **kwargs):
|
async def fetch_nodeinfo(*args, **kwargs):
|
||||||
async with HttpClient(database) as client:
|
async with HttpClient() as client:
|
||||||
return await client.fetch_nodeinfo(*args, **kwargs)
|
return await client.fetch_nodeinfo(*args, **kwargs)
|
||||||
|
|
|
@ -4,6 +4,16 @@ import os
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
|
|
||||||
|
LEVELS = {
|
||||||
|
'critical': logging.CRITICAL,
|
||||||
|
'error': logging.ERROR,
|
||||||
|
'warning': logging.WARNING,
|
||||||
|
'info': logging.INFO,
|
||||||
|
'verbose': 15,
|
||||||
|
'debug': logging.DEBUG
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
## Add the verbose logging level
|
## Add the verbose logging level
|
||||||
def verbose(message, *args, **kwargs):
|
def verbose(message, *args, **kwargs):
|
||||||
if not logging.root.isEnabledFor(logging.VERBOSE):
|
if not logging.root.isEnabledFor(logging.VERBOSE):
|
||||||
|
@ -15,10 +25,6 @@ setattr(logging, 'verbose', verbose)
|
||||||
setattr(logging, 'VERBOSE', 15)
|
setattr(logging, 'VERBOSE', 15)
|
||||||
logging.addLevelName(15, 'VERBOSE')
|
logging.addLevelName(15, 'VERBOSE')
|
||||||
|
|
||||||
|
|
||||||
## Get log level and file from environment if possible
|
|
||||||
env_log_level = os.environ.get('LOG_LEVEL', 'INFO').upper()
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
env_log_file = Path(os.environ.get('LOG_FILE')).expanduser().resolve()
|
env_log_file = Path(os.environ.get('LOG_FILE')).expanduser().resolve()
|
||||||
|
|
||||||
|
@ -26,14 +32,6 @@ except TypeError:
|
||||||
env_log_file = None
|
env_log_file = None
|
||||||
|
|
||||||
|
|
||||||
## Make sure the level from the environment is valid
|
|
||||||
try:
|
|
||||||
log_level = getattr(logging, env_log_level)
|
|
||||||
|
|
||||||
except AttributeError:
|
|
||||||
log_level = logging.INFO
|
|
||||||
|
|
||||||
|
|
||||||
## Set logging config
|
## Set logging config
|
||||||
handlers = [logging.StreamHandler()]
|
handlers = [logging.StreamHandler()]
|
||||||
|
|
||||||
|
@ -41,7 +39,11 @@ if env_log_file:
|
||||||
handlers.append(logging.FileHandler(env_log_file))
|
handlers.append(logging.FileHandler(env_log_file))
|
||||||
|
|
||||||
logging.basicConfig(
|
logging.basicConfig(
|
||||||
level = log_level,
|
level = logging.INFO,
|
||||||
format = "[%(asctime)s] %(levelname)s: %(message)s",
|
format = "[%(asctime)s] %(levelname)s: %(message)s",
|
||||||
handlers = handlers
|
handlers = handlers
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def set_level(level):
|
||||||
|
logging.getLogger().setLevel(LEVELS[level.lower()])
|
||||||
|
|
573
relay/manage.py
573
relay/manage.py
|
@ -1,29 +1,39 @@
|
||||||
import Crypto
|
import Crypto
|
||||||
import asyncio
|
import asyncio
|
||||||
import click
|
import click
|
||||||
|
import json
|
||||||
import logging
|
import logging
|
||||||
import platform
|
import platform
|
||||||
|
import yaml
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
from . import misc, __version__
|
from . import __version__
|
||||||
from . import http_client as http
|
|
||||||
from .application import Application
|
from .application import Application
|
||||||
from .config import RELAY_SOFTWARE
|
from .database import DEFAULT_CONFIG, RELAY_SOFTWARE
|
||||||
|
from .http_client import get, post, fetch_nodeinfo
|
||||||
|
from .misc import Message, boolean, check_open_port
|
||||||
|
|
||||||
|
|
||||||
app = None
|
app = None
|
||||||
CONFIG_IGNORE = {'blocked_software', 'blocked_instances', 'whitelist'}
|
CONFIG_IGNORE = {
|
||||||
|
'privkey',
|
||||||
|
'version'
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
@click.group('cli', context_settings={'show_default': True}, invoke_without_command=True)
|
@click.group('cli', context_settings={'show_default': True}, invoke_without_command=True)
|
||||||
@click.option('--config', '-c', default='relay.yaml', help='path to the relay\'s config')
|
@click.option('--config', '-c', help='path to the relay\'s config')
|
||||||
@click.version_option(version=__version__, prog_name='ActivityRelay')
|
@click.version_option(version=__version__, prog_name='ActivityRelay')
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
def cli(ctx, config):
|
def cli(ctx, config):
|
||||||
global app
|
global app
|
||||||
app = Application(config)
|
app = Application(config)
|
||||||
|
|
||||||
|
if ctx.invoked_subcommand != 'convert':
|
||||||
|
app.setup()
|
||||||
|
|
||||||
if not ctx.invoked_subcommand:
|
if not ctx.invoked_subcommand:
|
||||||
if app.config.host.endswith('example.com'):
|
if app.config.host.endswith('example.com'):
|
||||||
cli_setup.callback()
|
cli_setup.callback()
|
||||||
|
@ -32,12 +42,76 @@ def cli(ctx, config):
|
||||||
cli_run.callback()
|
cli_run.callback()
|
||||||
|
|
||||||
|
|
||||||
|
@cli.command('convert')
|
||||||
|
@click.option('--old-config', '-o', help='path to the old relay config')
|
||||||
|
def cli_convert(old_config):
|
||||||
|
'Convert an old relay.yaml and relay.jsonld to the the new formats'
|
||||||
|
|
||||||
|
with open(old_config or 'relay.yaml') as fd:
|
||||||
|
config = yaml.load(fd.read(), Loader=yaml.SafeLoader)
|
||||||
|
ap = config.get('ap', {})
|
||||||
|
|
||||||
|
with open(config.get('db', 'relay.jsonld')) as fd:
|
||||||
|
db = json.load(fd)
|
||||||
|
|
||||||
|
app.config['general_host'] = ap.get('host', '__DEFAULT__')
|
||||||
|
app.config['general_listen'] = config.get('listen', '__DEFAULT__')
|
||||||
|
app.config['general_port'] = config.get('port', '__DEFAULT__')
|
||||||
|
|
||||||
|
with app.database.session as s:
|
||||||
|
s.put_config('description', config.get('note', '__DEFAULT__'))
|
||||||
|
s.put_config('push_limit', config.get('push_limit', '__DEFAULT__'))
|
||||||
|
s.put_config('json_cache', config.get('json_cache', '__DEFAULT__'))
|
||||||
|
s.put_config('workers', config.get('workers', '__DEFAULT__'))
|
||||||
|
s.put_config('http_timeout', config.get('timeout', '__DEFAULT__'))
|
||||||
|
s.put_config('privkey', db.get('private-key'))
|
||||||
|
|
||||||
|
for name in ap.get('blocked_software', []):
|
||||||
|
try: s.put_ban('software', name)
|
||||||
|
except KeyError: print(f'Already banned software: {name}')
|
||||||
|
|
||||||
|
for name in ap.get('blocked_instances', []):
|
||||||
|
try: s.put_ban('domain', name)
|
||||||
|
except KeyError: print(f'Already banned instance: {name}')
|
||||||
|
|
||||||
|
for name in ap.get('whitelist', []):
|
||||||
|
try: s.put_whitelist(name)
|
||||||
|
except KeyError: print(f'Already whitelisted domain: {name}')
|
||||||
|
|
||||||
|
for instance in db.get('relay-list', {}).values():
|
||||||
|
domain = instance['domain']
|
||||||
|
software = instance.get('software')
|
||||||
|
actor = None
|
||||||
|
|
||||||
|
if software == 'mastodon':
|
||||||
|
actor = f'https://{domain}/actor'
|
||||||
|
|
||||||
|
elif software in {'pleroma', 'akkoma'}:
|
||||||
|
actor = f'https://{domain}/relay'
|
||||||
|
|
||||||
|
s.put_instance(
|
||||||
|
domain = domain,
|
||||||
|
inbox = instance.get('inbox'),
|
||||||
|
software = software,
|
||||||
|
actor = actor,
|
||||||
|
followid = instance.get('followid'),
|
||||||
|
accept = True
|
||||||
|
)
|
||||||
|
|
||||||
|
app.config.save()
|
||||||
|
|
||||||
|
print('Config and database converted :3')
|
||||||
|
|
||||||
|
|
||||||
@cli.command('setup')
|
@cli.command('setup')
|
||||||
def cli_setup():
|
def cli_setup():
|
||||||
'Generate a new config'
|
'Generate a new config'
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
app.config.host = click.prompt('What domain will the relay be hosted on?', default=app.config.host)
|
app.config['general_host'] = click.prompt(
|
||||||
|
'What domain will the relay be hosted on?',
|
||||||
|
default = app.config.host
|
||||||
|
)
|
||||||
|
|
||||||
if not app.config.host.endswith('example.com'):
|
if not app.config.host.endswith('example.com'):
|
||||||
break
|
break
|
||||||
|
@ -45,14 +119,85 @@ def cli_setup():
|
||||||
click.echo('The domain must not be example.com')
|
click.echo('The domain must not be example.com')
|
||||||
|
|
||||||
if not app.config.is_docker:
|
if not app.config.is_docker:
|
||||||
app.config.listen = click.prompt('Which address should the relay listen on?', default=app.config.listen)
|
app.config['general_listen'] = click.prompt(
|
||||||
|
'Which address should the relay listen on?',
|
||||||
|
default = app.config.listen
|
||||||
|
)
|
||||||
|
|
||||||
while True:
|
while True:
|
||||||
app.config.port = click.prompt('What TCP port should the relay listen on?', default=app.config.port, type=int)
|
app.config['general_port'] = click.prompt(
|
||||||
|
'What TCP port should the relay listen on?',
|
||||||
|
default = app.config.port,
|
||||||
|
type = int
|
||||||
|
)
|
||||||
|
|
||||||
break
|
break
|
||||||
|
|
||||||
|
app.config['database_type'] = click.prompt(
|
||||||
|
'What database backend would you like to use for the relay?',
|
||||||
|
default = app.config.dbtype,
|
||||||
|
type = click.Choice(['sqlite', 'postgresql', 'mysql']),
|
||||||
|
show_choices = True
|
||||||
|
)
|
||||||
|
|
||||||
|
if app.config.dbtype == 'sqlite':
|
||||||
|
app.config['sqlite_database'] = click.prompt(
|
||||||
|
'Where would you like to store your database file? Relative paths are relative to the config file location.',
|
||||||
|
default = app.config['sqlite_database']
|
||||||
|
)
|
||||||
|
|
||||||
|
else:
|
||||||
|
dbconfig = app.config.dbconfig
|
||||||
|
app.config.hostname = click.prompt(
|
||||||
|
'What address is your database listening on?',
|
||||||
|
default = dbconfig.hostname
|
||||||
|
) or None
|
||||||
|
|
||||||
|
app.config.port = click.prompt(
|
||||||
|
'What port is your database listening on?',
|
||||||
|
default = dbconfig.port
|
||||||
|
) or None
|
||||||
|
|
||||||
|
app.config.database = click.prompt(
|
||||||
|
'What would you like the name of the database be?',
|
||||||
|
default = dbconfig.database
|
||||||
|
) or None
|
||||||
|
|
||||||
|
app.config.username = click.prompt(
|
||||||
|
'Which user will be connecting to the database?',
|
||||||
|
default = dbconfig.username
|
||||||
|
) or None
|
||||||
|
|
||||||
|
app.config.password = click.prompt(
|
||||||
|
'What is the database user\'s password?',
|
||||||
|
default = dbconfig.password
|
||||||
|
) or None
|
||||||
|
|
||||||
app.config.save()
|
app.config.save()
|
||||||
|
|
||||||
|
with app.database.session as s:
|
||||||
|
s.put_config('name', click.prompt(
|
||||||
|
'What do you want to name your relay?',
|
||||||
|
default = s.get_config('name')
|
||||||
|
))
|
||||||
|
|
||||||
|
s.put_config('description', click.prompt(
|
||||||
|
'Provide a small description of your relay. This will be on the front page',
|
||||||
|
default = s.get_config('description')
|
||||||
|
))
|
||||||
|
|
||||||
|
s.put_config('whitelist', click.prompt(
|
||||||
|
'Enable the whitelist?',
|
||||||
|
default = s.get_config('whitelist'),
|
||||||
|
type = boolean
|
||||||
|
))
|
||||||
|
|
||||||
|
s.put_config('require_approval', click.prompt(
|
||||||
|
'Require instances to be approved when following?',
|
||||||
|
default = s.get_config('require_approval'),
|
||||||
|
type = boolean
|
||||||
|
))
|
||||||
|
|
||||||
if not app.config.is_docker and click.confirm('Relay all setup! Would you like to run it now?'):
|
if not app.config.is_docker and click.confirm('Relay all setup! Would you like to run it now?'):
|
||||||
cli_run.callback()
|
cli_run.callback()
|
||||||
|
|
||||||
|
@ -61,8 +206,9 @@ def cli_setup():
|
||||||
def cli_run():
|
def cli_run():
|
||||||
'Run the relay'
|
'Run the relay'
|
||||||
|
|
||||||
if app.config.host.endswith('example.com'):
|
with app.database.session as s:
|
||||||
return click.echo('Relay is not set up. Please edit your relay config or run "activityrelay setup".')
|
if not s.get_config('privkey') or app.config.host.endswith('example.com'):
|
||||||
|
return click.echo('Relay is not set up. Please run "activityrelay setup".')
|
||||||
|
|
||||||
vers_split = platform.python_version().split('.')
|
vers_split = platform.python_version().split('.')
|
||||||
pip_command = 'pip3 uninstall pycrypto && pip3 install pycryptodome'
|
pip_command = 'pip3 uninstall pycrypto && pip3 install pycryptodome'
|
||||||
|
@ -76,7 +222,7 @@ def cli_run():
|
||||||
click.echo('Warning: PyCrypto is old and should be replaced with pycryptodome')
|
click.echo('Warning: PyCrypto is old and should be replaced with pycryptodome')
|
||||||
return click.echo(pip_command)
|
return click.echo(pip_command)
|
||||||
|
|
||||||
if not misc.check_open_port(app.config.listen, app.config.port):
|
if not check_open_port(app.config.listen, app.config.port):
|
||||||
return click.echo(f'Error: A server is already running on port {app.config.port}')
|
return click.echo(f'Error: A server is already running on port {app.config.port}')
|
||||||
|
|
||||||
app.run()
|
app.run()
|
||||||
|
@ -95,22 +241,28 @@ def cli_config_list():
|
||||||
|
|
||||||
click.echo('Relay Config:')
|
click.echo('Relay Config:')
|
||||||
|
|
||||||
for key, value in app.config.items():
|
with app.database.session as s:
|
||||||
if key not in CONFIG_IGNORE:
|
config = s.get_config_all()
|
||||||
key = f'{key}:'.ljust(20)
|
|
||||||
click.echo(f'- {key} {value}')
|
for key in DEFAULT_CONFIG.keys():
|
||||||
|
if key in CONFIG_IGNORE:
|
||||||
|
continue
|
||||||
|
|
||||||
|
keystr = f'{key}:'.ljust(20)
|
||||||
|
click.echo(f'- {keystr} {config[key]}')
|
||||||
|
|
||||||
|
|
||||||
@cli_config.command('set')
|
@cli_config.command('set')
|
||||||
@click.argument('key')
|
@click.argument('key')
|
||||||
@click.argument('value')
|
@click.argument('value', nargs=-1)
|
||||||
def cli_config_set(key, value):
|
def cli_config_set(key, value):
|
||||||
'Set a config value'
|
'Set a config value'
|
||||||
|
|
||||||
app.config[key] = value
|
with app.database.session as s:
|
||||||
app.config.save()
|
s.put_config(key, ' '.join(value))
|
||||||
|
value = s.get_config(key)
|
||||||
|
|
||||||
print(f'{key}: {app.config[key]}')
|
print(f'{key}: {value}')
|
||||||
|
|
||||||
|
|
||||||
@cli.group('inbox')
|
@cli.group('inbox')
|
||||||
|
@ -125,8 +277,9 @@ def cli_inbox_list():
|
||||||
|
|
||||||
click.echo('Connected to the following instances or relays:')
|
click.echo('Connected to the following instances or relays:')
|
||||||
|
|
||||||
for inbox in app.database.inboxes:
|
with app.database.session as s:
|
||||||
click.echo(f'- {inbox}')
|
for instance in s.get_instances():
|
||||||
|
click.echo(f'- {instance.inbox}')
|
||||||
|
|
||||||
|
|
||||||
@cli_inbox.command('follow')
|
@cli_inbox.command('follow')
|
||||||
|
@ -134,9 +287,6 @@ def cli_inbox_list():
|
||||||
def cli_inbox_follow(actor):
|
def cli_inbox_follow(actor):
|
||||||
'Follow an actor (Relay must be running)'
|
'Follow an actor (Relay must be running)'
|
||||||
|
|
||||||
if app.config.is_banned(actor):
|
|
||||||
return click.echo(f'Error: Refusing to follow banned actor: {actor}')
|
|
||||||
|
|
||||||
if not actor.startswith('http'):
|
if not actor.startswith('http'):
|
||||||
domain = actor
|
domain = actor
|
||||||
actor = f'https://{actor}/actor'
|
actor = f'https://{actor}/actor'
|
||||||
|
@ -144,24 +294,32 @@ def cli_inbox_follow(actor):
|
||||||
else:
|
else:
|
||||||
domain = urlparse(actor).hostname
|
domain = urlparse(actor).hostname
|
||||||
|
|
||||||
try:
|
with app.database.session as s:
|
||||||
inbox_data = app.database['relay-list'][domain]
|
if s.get_ban('domain', domain):
|
||||||
inbox = inbox_data['inbox']
|
return click.echo(f'Error: Refusing to follow banned actor: {actor}')
|
||||||
|
|
||||||
except KeyError:
|
instance = s.get_instance(domain)
|
||||||
actor_data = asyncio.run(http.get(app.database, actor, sign_headers=True))
|
|
||||||
|
|
||||||
if not actor_data:
|
if not instance:
|
||||||
return click.echo(f'Failed to fetch actor: {actor}')
|
actor_data = asyncio.run(get(actor, sign_headers=True))
|
||||||
|
|
||||||
inbox = actor_data.shared_inbox
|
if not actor_data:
|
||||||
|
return click.echo(f'Failed to fetch actor: {actor}')
|
||||||
|
|
||||||
message = misc.Message.new_follow(
|
inbox = actor_data.shared_inbox
|
||||||
|
|
||||||
|
else:
|
||||||
|
inbox = instance.inbox
|
||||||
|
|
||||||
|
if instance.actor:
|
||||||
|
actor = instance.actor
|
||||||
|
|
||||||
|
message = Message.new_follow(
|
||||||
host = app.config.host,
|
host = app.config.host,
|
||||||
actor = actor
|
actor = actor
|
||||||
)
|
)
|
||||||
|
|
||||||
asyncio.run(http.post(app.database, inbox, message))
|
asyncio.run(post(inbox, message))
|
||||||
click.echo(f'Sent follow message to actor: {actor}')
|
click.echo(f'Sent follow message to actor: {actor}')
|
||||||
|
|
||||||
|
|
||||||
|
@ -170,6 +328,8 @@ def cli_inbox_follow(actor):
|
||||||
def cli_inbox_unfollow(actor):
|
def cli_inbox_unfollow(actor):
|
||||||
'Unfollow an actor (Relay must be running)'
|
'Unfollow an actor (Relay must be running)'
|
||||||
|
|
||||||
|
followid = None
|
||||||
|
|
||||||
if not actor.startswith('http'):
|
if not actor.startswith('http'):
|
||||||
domain = actor
|
domain = actor
|
||||||
actor = f'https://{actor}/actor'
|
actor = f'https://{actor}/actor'
|
||||||
|
@ -177,68 +337,176 @@ def cli_inbox_unfollow(actor):
|
||||||
else:
|
else:
|
||||||
domain = urlparse(actor).hostname
|
domain = urlparse(actor).hostname
|
||||||
|
|
||||||
try:
|
with app.database.session as s:
|
||||||
inbox_data = app.database['relay-list'][domain]
|
instance = s.get_instance(domain)
|
||||||
inbox = inbox_data['inbox']
|
|
||||||
message = misc.Message.new_unfollow(
|
if not instance:
|
||||||
|
actor_data = asyncio.run(get(actor, sign_headers=True))
|
||||||
|
|
||||||
|
if not actor_data:
|
||||||
|
return click.echo(f'Failed to fetch actor: {actor}')
|
||||||
|
|
||||||
|
inbox = actor_data.shared_inbox
|
||||||
|
|
||||||
|
else:
|
||||||
|
inbox = instance.inbox
|
||||||
|
followid = instance.followid
|
||||||
|
|
||||||
|
if instance.actor:
|
||||||
|
actor = instance.actor
|
||||||
|
|
||||||
|
if followid:
|
||||||
|
message = Message.new_unfollow(
|
||||||
host = app.config.host,
|
host = app.config.host,
|
||||||
actor = actor,
|
actor = actor,
|
||||||
follow = inbox_data['followid']
|
follow = followid
|
||||||
)
|
)
|
||||||
|
|
||||||
except KeyError:
|
else:
|
||||||
actor_data = asyncio.run(http.get(app.database, actor, sign_headers=True))
|
|
||||||
inbox = actor_data.shared_inbox
|
|
||||||
message = misc.Message.new_unfollow(
|
message = misc.Message.new_unfollow(
|
||||||
host = app.config.host,
|
host = app.config.host,
|
||||||
actor = actor,
|
actor = actor,
|
||||||
follow = {
|
follow = {
|
||||||
'type': 'Follow',
|
'type': 'Follow',
|
||||||
'object': actor,
|
'object': actor,
|
||||||
'actor': f'https://{app.config.host}/actor'
|
'actor': app.config.actor
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
asyncio.run(http.post(app.database, inbox, message))
|
asyncio.run(post(inbox, message))
|
||||||
click.echo(f'Sent unfollow message to: {actor}')
|
click.echo(f'Sent unfollow message to: {actor}')
|
||||||
|
|
||||||
|
|
||||||
@cli_inbox.command('add')
|
@cli_inbox.command('add')
|
||||||
@click.argument('inbox')
|
@click.argument('actor')
|
||||||
def cli_inbox_add(inbox):
|
def cli_inbox_add(actor):
|
||||||
'Add an inbox to the database'
|
'Add an instance to the database'
|
||||||
|
|
||||||
if not inbox.startswith('http'):
|
if not actor.startswith('http'):
|
||||||
inbox = f'https://{inbox}/inbox'
|
domain = actor
|
||||||
|
actor = f'https://{actor}/inbox'
|
||||||
|
|
||||||
if app.config.is_banned(inbox):
|
else:
|
||||||
return click.echo(f'Error: Refusing to add banned inbox: {inbox}')
|
domain = urlparse(actor).hostname
|
||||||
|
|
||||||
if app.database.get_inbox(inbox):
|
with app.database.session as s:
|
||||||
return click.echo(f'Error: Inbox already in database: {inbox}')
|
data = {
|
||||||
|
'domain': domain,
|
||||||
|
'actor': actor,
|
||||||
|
'inbox': f'https://{domain}/inbox'
|
||||||
|
}
|
||||||
|
|
||||||
app.database.add_inbox(inbox)
|
if s.get_instance(domain):
|
||||||
app.database.save()
|
return click.echo(f'Error: Instance already in database: {domain}')
|
||||||
|
|
||||||
click.echo(f'Added inbox to the database: {inbox}')
|
if s.get_ban('domain', domain):
|
||||||
|
return click.echo(f'Error: Refusing to add banned domain: {domain}')
|
||||||
|
|
||||||
|
nodeinfo = asyncio.run(fetch_nodeinfo(domain))
|
||||||
|
|
||||||
|
if nodeinfo:
|
||||||
|
if s.get_ban('software', nodeinfo.sw_name):
|
||||||
|
return click.echo(f'Error: Refusing to add banned software: {nodeinfo.sw_name}')
|
||||||
|
|
||||||
|
data['software'] = nodeinfo.sw_name
|
||||||
|
|
||||||
|
actor_data = asyncio.run(get(actor, sign_headers=True))
|
||||||
|
|
||||||
|
if actor_data:
|
||||||
|
instance = s.put_instance_actor(actor, nodeinfo)
|
||||||
|
|
||||||
|
else:
|
||||||
|
instance = s.put_instance(**data)
|
||||||
|
|
||||||
|
click.echo(f'Added instance to the database: {instance.domain}')
|
||||||
|
|
||||||
|
|
||||||
@cli_inbox.command('remove')
|
@cli_inbox.command('remove')
|
||||||
@click.argument('inbox')
|
@click.argument('domain')
|
||||||
def cli_inbox_remove(inbox):
|
def cli_inbox_remove(domain):
|
||||||
'Remove an inbox from the database'
|
'Remove an inbox from the database'
|
||||||
|
|
||||||
try:
|
if domain.startswith('http'):
|
||||||
dbinbox = app.database.get_inbox(inbox, fail=True)
|
domain = urlparse(domain).hostname
|
||||||
|
|
||||||
except KeyError:
|
with app.database.session as s:
|
||||||
click.echo(f'Error: Inbox does not exist: {inbox}')
|
try:
|
||||||
return
|
s.delete_instance(domain)
|
||||||
|
click.echo(f'Removed inbox from the database: {domain}')
|
||||||
|
|
||||||
app.database.del_inbox(dbinbox['domain'])
|
except KeyError:
|
||||||
app.database.save()
|
return click.echo(f'Error: Inbox does not exist: {domain}')
|
||||||
|
|
||||||
click.echo(f'Removed inbox from the database: {inbox}')
|
|
||||||
|
@cli.group('request')
|
||||||
|
def cli_request():
|
||||||
|
'Manage follow requests'
|
||||||
|
|
||||||
|
|
||||||
|
@cli_request.command('list')
|
||||||
|
def cli_request_list():
|
||||||
|
'List all the current follow requests'
|
||||||
|
|
||||||
|
click.echo('Follow requests:')
|
||||||
|
|
||||||
|
with app.database.session as s:
|
||||||
|
for row in s.get_requests():
|
||||||
|
click.echo(f'- {row.domain}')
|
||||||
|
|
||||||
|
|
||||||
|
@cli_request.command('approve')
|
||||||
|
@click.argument('domain')
|
||||||
|
def cli_request_approve(domain):
|
||||||
|
'Approve a follow request'
|
||||||
|
|
||||||
|
with app.database.session as s:
|
||||||
|
try:
|
||||||
|
instance = s.get_request(domain)
|
||||||
|
|
||||||
|
except KeyError:
|
||||||
|
return click.echo(f'No request for domain exists: {domain}')
|
||||||
|
|
||||||
|
data = {'joined': datetime.now()}
|
||||||
|
s.update('instances', data, id=instance.id)
|
||||||
|
|
||||||
|
asyncio.run(post(
|
||||||
|
instance.inbox,
|
||||||
|
Message.new_response(
|
||||||
|
host = app.config.host,
|
||||||
|
actor = instance.actor,
|
||||||
|
followid = instance.followid,
|
||||||
|
accept = True
|
||||||
|
)
|
||||||
|
))
|
||||||
|
|
||||||
|
return click.echo(f'Accepted follow request for domain: {domain}')
|
||||||
|
|
||||||
|
|
||||||
|
@cli_request.command('deny')
|
||||||
|
@click.argument('domain')
|
||||||
|
def cli_request_deny(domain):
|
||||||
|
'Deny a follow request'
|
||||||
|
|
||||||
|
with app.database.session as s:
|
||||||
|
try:
|
||||||
|
instance = s.get_request(domain)
|
||||||
|
|
||||||
|
except KeyError:
|
||||||
|
return click.echo(f'No request for domain exists: {domain}')
|
||||||
|
|
||||||
|
s.delete_instance(domain)
|
||||||
|
|
||||||
|
asyncio.run(post(
|
||||||
|
instance.inbox,
|
||||||
|
Message.new_response(
|
||||||
|
host = app.config.host,
|
||||||
|
actor = instance.actor,
|
||||||
|
followid = instance.followid,
|
||||||
|
accept = False
|
||||||
|
)
|
||||||
|
))
|
||||||
|
|
||||||
|
return click.echo(f'Denied follow request for domain: {domain}')
|
||||||
|
|
||||||
|
|
||||||
@cli.group('instance')
|
@cli.group('instance')
|
||||||
|
@ -253,42 +521,50 @@ def cli_instance_list():
|
||||||
|
|
||||||
click.echo('Banned instances or relays:')
|
click.echo('Banned instances or relays:')
|
||||||
|
|
||||||
for domain in app.config.blocked_instances:
|
with app.database.session as s:
|
||||||
click.echo(f'- {domain}')
|
for row in s.get_bans('domain'):
|
||||||
|
click.echo(f'- {row.name}')
|
||||||
|
|
||||||
|
|
||||||
@cli_instance.command('ban')
|
@cli_instance.command('ban')
|
||||||
@click.argument('target')
|
@click.argument('domain')
|
||||||
def cli_instance_ban(target):
|
def cli_instance_ban(domain):
|
||||||
'Ban an instance and remove the associated inbox if it exists'
|
'Ban an instance and remove the associated inbox if it exists'
|
||||||
|
|
||||||
if target.startswith('http'):
|
if domain.startswith('http'):
|
||||||
target = urlparse(target).hostname
|
domain = urlparse(domain).hostname
|
||||||
|
|
||||||
if app.config.ban_instance(target):
|
with app.database.session as s:
|
||||||
app.config.save()
|
try:
|
||||||
|
s.put_ban('domain', domain)
|
||||||
|
|
||||||
if app.database.del_inbox(target):
|
except KeyError:
|
||||||
app.database.save()
|
return click.echo(f'Instance already banned: {domain}')
|
||||||
|
|
||||||
click.echo(f'Banned instance: {target}')
|
try:
|
||||||
return
|
s.delete_instance(domain)
|
||||||
|
|
||||||
click.echo(f'Instance already banned: {target}')
|
except KeyError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
click.echo(f'Banned instance: {domain}')
|
||||||
|
|
||||||
|
|
||||||
@cli_instance.command('unban')
|
@cli_instance.command('unban')
|
||||||
@click.argument('target')
|
@click.argument('domain')
|
||||||
def cli_instance_unban(target):
|
def cli_instance_unban(domain):
|
||||||
'Unban an instance'
|
'Unban an instance'
|
||||||
|
|
||||||
if app.config.unban_instance(target):
|
if domain.startswith('http'):
|
||||||
app.config.save()
|
domain = urlparse(domain).hostname
|
||||||
|
|
||||||
click.echo(f'Unbanned instance: {target}')
|
with app.database.session as s:
|
||||||
return
|
try:
|
||||||
|
s.delete_ban('domain', domain)
|
||||||
|
click.echo(f'Unbanned instance: {domain}')
|
||||||
|
|
||||||
click.echo(f'Instance wasn\'t banned: {target}')
|
except KeyError:
|
||||||
|
click.echo(f'Instance wasn\'t banned: {domain}')
|
||||||
|
|
||||||
|
|
||||||
@cli.group('software')
|
@cli.group('software')
|
||||||
|
@ -303,8 +579,9 @@ def cli_software_list():
|
||||||
|
|
||||||
click.echo('Banned software:')
|
click.echo('Banned software:')
|
||||||
|
|
||||||
for software in app.config.blocked_software:
|
with app.database.session as s:
|
||||||
click.echo(f'- {software}')
|
for row in s.get_bans('software'):
|
||||||
|
click.echo(f'- {row.name}')
|
||||||
|
|
||||||
|
|
||||||
@cli_software.command('ban')
|
@cli_software.command('ban')
|
||||||
|
@ -315,26 +592,27 @@ def cli_software_list():
|
||||||
def cli_software_ban(name, fetch_nodeinfo):
|
def cli_software_ban(name, fetch_nodeinfo):
|
||||||
'Ban software. Use RELAYS for NAME to ban relays'
|
'Ban software. Use RELAYS for NAME to ban relays'
|
||||||
|
|
||||||
if name == 'RELAYS':
|
with app.database.session as s:
|
||||||
for name in RELAY_SOFTWARE:
|
if name == 'RELAYS':
|
||||||
app.config.ban_software(name)
|
for name in RELAY_SOFTWARE:
|
||||||
|
s.put_ban('software', name)
|
||||||
|
|
||||||
app.config.save()
|
return click.echo('Banned all relay software')
|
||||||
return click.echo('Banned all relay software')
|
|
||||||
|
|
||||||
if fetch_nodeinfo:
|
if fetch_nodeinfo:
|
||||||
nodeinfo = asyncio.run(http.fetch_nodeinfo(app.database, name))
|
nodeinfo = asyncio.run(fetch_nodeinfo(name))
|
||||||
|
|
||||||
if not nodeinfo:
|
if not nodeinfo:
|
||||||
click.echo(f'Failed to fetch software name from domain: {name}')
|
return click.echo(f'Failed to fetch software name from domain: {name}')
|
||||||
|
|
||||||
name = nodeinfo.sw_name
|
name = nodeinfo.sw_name
|
||||||
|
|
||||||
if app.config.ban_software(name):
|
try:
|
||||||
app.config.save()
|
s.put_ban('software', name)
|
||||||
return click.echo(f'Banned software: {name}')
|
click.echo(f'Banned software: {name}')
|
||||||
|
|
||||||
click.echo(f'Software already banned: {name}')
|
except KeyError:
|
||||||
|
click.echo(f'Software already banned: {name}')
|
||||||
|
|
||||||
|
|
||||||
@cli_software.command('unban')
|
@cli_software.command('unban')
|
||||||
|
@ -345,26 +623,27 @@ def cli_software_ban(name, fetch_nodeinfo):
|
||||||
def cli_software_unban(name, fetch_nodeinfo):
|
def cli_software_unban(name, fetch_nodeinfo):
|
||||||
'Ban software. Use RELAYS for NAME to unban relays'
|
'Ban software. Use RELAYS for NAME to unban relays'
|
||||||
|
|
||||||
if name == 'RELAYS':
|
with app.database.session as s:
|
||||||
for name in RELAY_SOFTWARE:
|
if name == 'RELAYS':
|
||||||
app.config.unban_software(name)
|
for name in RELAY_SOFTWARE:
|
||||||
|
s.put_ban('software', name)
|
||||||
|
|
||||||
app.config.save()
|
return click.echo('Unbanned all relay software')
|
||||||
return click.echo('Unbanned all relay software')
|
|
||||||
|
|
||||||
if fetch_nodeinfo:
|
if fetch_nodeinfo:
|
||||||
nodeinfo = asyncio.run(http.fetch_nodeinfo(app.database, name))
|
nodeinfo = asyncio.run(fetch_nodeinfo(name))
|
||||||
|
|
||||||
if not nodeinfo:
|
if not nodeinfo:
|
||||||
click.echo(f'Failed to fetch software name from domain: {name}')
|
return click.echo(f'Failed to fetch software name from domain: {name}')
|
||||||
|
|
||||||
name = nodeinfo.sw_name
|
name = nodeinfo.sw_name
|
||||||
|
|
||||||
if app.config.unban_software(name):
|
try:
|
||||||
app.config.save()
|
s.put_ban('software', name)
|
||||||
return click.echo(f'Unbanned software: {name}')
|
click.echo(f'Unbanned software: {name}')
|
||||||
|
|
||||||
click.echo(f'Software wasn\'t banned: {name}')
|
except KeyError:
|
||||||
|
click.echo(f'Software wasn\'t banned: {name}')
|
||||||
|
|
||||||
|
|
||||||
@cli.group('whitelist')
|
@cli.group('whitelist')
|
||||||
|
@ -377,47 +656,61 @@ def cli_whitelist():
|
||||||
def cli_whitelist_list():
|
def cli_whitelist_list():
|
||||||
'List all the instances in the whitelist'
|
'List all the instances in the whitelist'
|
||||||
|
|
||||||
click.echo('Current whitelisted domains')
|
click.echo('Current whitelisted domains:')
|
||||||
|
|
||||||
for domain in app.config.whitelist:
|
with app.database.session as s:
|
||||||
click.echo(f'- {domain}')
|
for row in s.get_whitelist():
|
||||||
|
click.echo(f'- {row.domain}')
|
||||||
|
|
||||||
|
|
||||||
@cli_whitelist.command('add')
|
@cli_whitelist.command('add')
|
||||||
@click.argument('instance')
|
@click.argument('domain')
|
||||||
def cli_whitelist_add(instance):
|
def cli_whitelist_add(domain):
|
||||||
'Add an instance to the whitelist'
|
'Add a domain to the whitelist'
|
||||||
|
|
||||||
if not app.config.add_whitelist(instance):
|
with app.database.session as s:
|
||||||
return click.echo(f'Instance already in the whitelist: {instance}')
|
try:
|
||||||
|
s.put_whitelist(domain)
|
||||||
|
click.echo(f'Instance added to the whitelist: {domain}')
|
||||||
|
|
||||||
app.config.save()
|
except KeyError:
|
||||||
click.echo(f'Instance added to the whitelist: {instance}')
|
return click.echo(f'Instance already in the whitelist: {domain}')
|
||||||
|
|
||||||
|
|
||||||
@cli_whitelist.command('remove')
|
@cli_whitelist.command('remove')
|
||||||
@click.argument('instance')
|
@click.argument('domain')
|
||||||
def cli_whitelist_remove(instance):
|
def cli_whitelist_remove(domain):
|
||||||
'Remove an instance from the whitelist'
|
'Remove a domain from the whitelist'
|
||||||
|
|
||||||
if not app.config.del_whitelist(instance):
|
with app.database.session as s:
|
||||||
return click.echo(f'Instance not in the whitelist: {instance}')
|
try:
|
||||||
|
s.delete_whitelist(domain)
|
||||||
|
click.echo(f'Removed instance from the whitelist: {domain}')
|
||||||
|
|
||||||
app.config.save()
|
except KeyError:
|
||||||
|
click.echo(f'Instance not in the whitelist: {domain}')
|
||||||
if app.config.whitelist_enabled:
|
|
||||||
if app.database.del_inbox(instance):
|
|
||||||
app.database.save()
|
|
||||||
|
|
||||||
click.echo(f'Removed instance from the whitelist: {instance}')
|
|
||||||
|
|
||||||
|
|
||||||
@cli_whitelist.command('import')
|
@cli_whitelist.command('import')
|
||||||
def cli_whitelist_import():
|
def cli_whitelist_import():
|
||||||
'Add all current inboxes to the whitelist'
|
'Add all current inboxes to the whitelist'
|
||||||
|
|
||||||
for domain in app.database.hostnames:
|
with app.database.session as s:
|
||||||
cli_whitelist_add.callback(domain)
|
for row in s.get_instances():
|
||||||
|
try:
|
||||||
|
s.put_whitelist(row.domain)
|
||||||
|
click.echo(f'Instance added to the whitelist: {row.domain}')
|
||||||
|
|
||||||
|
except KeyError:
|
||||||
|
click.echo(f'Instance already in the whitelist: {row.domain}')
|
||||||
|
|
||||||
|
|
||||||
|
@cli_whitelist.command('clear')
|
||||||
|
def cli_whitelist_clear():
|
||||||
|
'Clear all items out of the whitelist'
|
||||||
|
|
||||||
|
with app.database.session as s:
|
||||||
|
s.delete('whitelist')
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
|
@ -36,6 +36,9 @@ def set_app(new_app):
|
||||||
|
|
||||||
|
|
||||||
def boolean(value):
|
def boolean(value):
|
||||||
|
if isinstance(value, bytes):
|
||||||
|
value = str(value, 'utf-8')
|
||||||
|
|
||||||
if isinstance(value, str):
|
if isinstance(value, str):
|
||||||
if value.lower() in ['on', 'y', 'yes', 'true', 'enable', 'enabled', '1']:
|
if value.lower() in ['on', 'y', 'yes', 'true', 'enable', 'enabled', '1']:
|
||||||
return True
|
return True
|
||||||
|
@ -63,7 +66,7 @@ def boolean(value):
|
||||||
return value.__bool__()
|
return value.__bool__()
|
||||||
|
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
raise TypeError(f'Cannot convert object of type "{clsname(value)}"')
|
raise TypeError(f'Cannot convert object of type "{type(value).__name__}"')
|
||||||
|
|
||||||
|
|
||||||
def check_open_port(host, port):
|
def check_open_port(host, port):
|
||||||
|
@ -78,6 +81,32 @@ def check_open_port(host, port):
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
class AppBase:
|
||||||
|
@property
|
||||||
|
def app(self):
|
||||||
|
return app
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def client(self):
|
||||||
|
return app.client
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def config(self):
|
||||||
|
return app.config
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def database(self):
|
||||||
|
return app.database
|
||||||
|
|
||||||
|
|
||||||
|
@property
|
||||||
|
def signer(self):
|
||||||
|
return app.signer
|
||||||
|
|
||||||
|
|
||||||
class DotDict(dict):
|
class DotDict(dict):
|
||||||
def __init__(self, _data, **kwargs):
|
def __init__(self, _data, **kwargs):
|
||||||
dict.__init__(self)
|
dict.__init__(self)
|
||||||
|
@ -163,14 +192,15 @@ class DotDict(dict):
|
||||||
|
|
||||||
class Message(DotDict):
|
class Message(DotDict):
|
||||||
@classmethod
|
@classmethod
|
||||||
def new_actor(cls, host, pubkey, description=None):
|
def new_actor(cls, host, pubkey, name=None, description=None, locked=False):
|
||||||
return cls({
|
return cls({
|
||||||
'@context': 'https://www.w3.org/ns/activitystreams',
|
'@context': 'https://www.w3.org/ns/activitystreams',
|
||||||
'id': f'https://{host}/actor',
|
'id': f'https://{host}/actor',
|
||||||
'type': 'Application',
|
'type': 'Application',
|
||||||
'preferredUsername': 'relay',
|
'preferredUsername': 'relay',
|
||||||
'name': 'ActivityRelay',
|
'name': name or 'ActivityRelay',
|
||||||
'summary': description or 'ActivityRelay bot',
|
'summary': description or 'ActivityRelay bot',
|
||||||
|
'manuallyApprovesFollowers': locked,
|
||||||
'followers': f'https://{host}/followers',
|
'followers': f'https://{host}/followers',
|
||||||
'following': f'https://{host}/following',
|
'following': f'https://{host}/following',
|
||||||
'inbox': f'https://{host}/inbox',
|
'inbox': f'https://{host}/inbox',
|
||||||
|
@ -310,31 +340,3 @@ class Response(AiohttpResponse):
|
||||||
@location.setter
|
@location.setter
|
||||||
def location(self, value):
|
def location(self, value):
|
||||||
self.headers['Location'] = value
|
self.headers['Location'] = value
|
||||||
|
|
||||||
|
|
||||||
class View(AiohttpView):
|
|
||||||
async def _iter(self):
|
|
||||||
if self.request.method not in METHODS:
|
|
||||||
self._raise_allowed_methods()
|
|
||||||
|
|
||||||
method = getattr(self, self.request.method.lower(), None)
|
|
||||||
|
|
||||||
if method is None:
|
|
||||||
self._raise_allowed_methods()
|
|
||||||
|
|
||||||
return await method(**self.request.match_info)
|
|
||||||
|
|
||||||
|
|
||||||
@property
|
|
||||||
def app(self):
|
|
||||||
return self._request.app
|
|
||||||
|
|
||||||
|
|
||||||
@property
|
|
||||||
def config(self):
|
|
||||||
return self.app.config
|
|
||||||
|
|
||||||
|
|
||||||
@property
|
|
||||||
def database(self):
|
|
||||||
return self.app.database
|
|
||||||
|
|
|
@ -4,6 +4,7 @@ import logging
|
||||||
from cachetools import LRUCache
|
from cachetools import LRUCache
|
||||||
from uuid import uuid4
|
from uuid import uuid4
|
||||||
|
|
||||||
|
from .database import RELAY_SOFTWARE
|
||||||
from .misc import Message
|
from .misc import Message
|
||||||
|
|
||||||
|
|
||||||
|
@ -20,11 +21,16 @@ def person_check(actor, software):
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
async def handle_relay(request):
|
async def handle_relay(request, s):
|
||||||
if request.message.objectid in cache:
|
if request.message.objectid in cache:
|
||||||
logging.verbose(f'already relayed {request.message.objectid}')
|
logging.verbose(f'already relayed {request.message.objectid}')
|
||||||
return
|
return
|
||||||
|
|
||||||
|
if request.message.get('to') != ['https://www.w3.org/ns/activitystreams#Public']:
|
||||||
|
logging.verbose('Message was not public')
|
||||||
|
logging.verbose(request.message.get('to'))
|
||||||
|
return
|
||||||
|
|
||||||
message = Message.new_announce(
|
message = Message.new_announce(
|
||||||
host = request.config.host,
|
host = request.config.host,
|
||||||
object = request.message.objectid
|
object = request.message.objectid
|
||||||
|
@ -33,13 +39,13 @@ async def handle_relay(request):
|
||||||
cache[request.message.objectid] = message.id
|
cache[request.message.objectid] = message.id
|
||||||
logging.debug(f'>> relay: {message}')
|
logging.debug(f'>> relay: {message}')
|
||||||
|
|
||||||
inboxes = request.database.distill_inboxes(request.message)
|
inboxes = s.distill_inboxes(request.message)
|
||||||
|
|
||||||
for inbox in inboxes:
|
for inbox in inboxes:
|
||||||
request.app.push_message(inbox, message)
|
request.app.push_message(inbox, message)
|
||||||
|
|
||||||
|
|
||||||
async def handle_forward(request):
|
async def handle_forward(request, s):
|
||||||
if request.message.id in cache:
|
if request.message.id in cache:
|
||||||
logging.verbose(f'already forwarded {request.message.id}')
|
logging.verbose(f'already forwarded {request.message.id}')
|
||||||
return
|
return
|
||||||
|
@ -52,57 +58,73 @@ async def handle_forward(request):
|
||||||
cache[request.message.id] = message.id
|
cache[request.message.id] = message.id
|
||||||
logging.debug(f'>> forward: {message}')
|
logging.debug(f'>> forward: {message}')
|
||||||
|
|
||||||
inboxes = request.database.distill_inboxes(request.message)
|
inboxes = s.distill_inboxes(request.message)
|
||||||
|
|
||||||
for inbox in inboxes:
|
for inbox in inboxes:
|
||||||
request.app.push_message(inbox, message)
|
request.app.push_message(inbox, message)
|
||||||
|
|
||||||
|
|
||||||
async def handle_follow(request):
|
async def handle_follow(request, s):
|
||||||
|
approve = True
|
||||||
|
|
||||||
nodeinfo = await request.app.client.fetch_nodeinfo(request.actor.domain)
|
nodeinfo = await request.app.client.fetch_nodeinfo(request.actor.domain)
|
||||||
software = nodeinfo.sw_name if nodeinfo else None
|
software = nodeinfo.sw_name if nodeinfo else None
|
||||||
|
|
||||||
## reject if software used by actor is banned
|
## reject if the actor isn't whitelisted while the whiltelist is enabled
|
||||||
if request.config.is_banned_software(software):
|
if s.get_config('whitelist') and not s.get_whitelist(request.actor.domain):
|
||||||
request.app.push_message(
|
logging.verbose(f'Rejected actor for not being in the whitelist: {request.actor.id}')
|
||||||
request.actor.shared_inbox,
|
approve = False
|
||||||
Message.new_response(
|
|
||||||
host = request.config.host,
|
|
||||||
actor = request.actor.id,
|
|
||||||
followid = request.message.id,
|
|
||||||
accept = False
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
return logging.verbose(f'Rejected follow from actor for using specific software: actor={request.actor.id}, software={software}')
|
## reject if software used by actor is banned
|
||||||
|
if s.get_ban('software', software):
|
||||||
|
logging.verbose(f'Rejected follow from actor for using specific software: actor={request.actor.id}, software={software}')
|
||||||
|
approve = False
|
||||||
|
|
||||||
## reject if the actor is not an instance actor
|
## reject if the actor is not an instance actor
|
||||||
if person_check(request.actor, software):
|
if person_check(request.actor, software):
|
||||||
request.app.push_message(
|
logging.verbose(f'Non-application actor tried to follow: {request.actor.id}')
|
||||||
request.actor.shared_inbox,
|
approve = False
|
||||||
Message.new_response(
|
|
||||||
host = request.config.host,
|
if approve:
|
||||||
|
if not request.instance:
|
||||||
|
s.put_instance(
|
||||||
|
domain = request.actor.domain,
|
||||||
actor = request.actor.id,
|
actor = request.actor.id,
|
||||||
|
inbox = request.actor.shared_inbox,
|
||||||
|
actor_data = request.actor,
|
||||||
|
software = software,
|
||||||
followid = request.message.id,
|
followid = request.message.id,
|
||||||
accept = False
|
accept = not s.get_config('require_approval')
|
||||||
)
|
)
|
||||||
)
|
|
||||||
|
|
||||||
return logging.verbose(f'Non-application actor tried to follow: {request.actor.id}')
|
if s.get_config('require_approval'):
|
||||||
|
return
|
||||||
|
|
||||||
request.database.add_inbox(request.actor.shared_inbox, request.message.id, software)
|
else:
|
||||||
request.database.save()
|
s.put_instance(
|
||||||
|
domain = request.actor.domain,
|
||||||
|
followid = request.message.id
|
||||||
|
)
|
||||||
|
|
||||||
|
# Rejects don't seem to work right with mastodon
|
||||||
request.app.push_message(
|
request.app.push_message(
|
||||||
request.actor.shared_inbox,
|
request.actor.inbox,
|
||||||
Message.new_response(
|
Message.new_response(
|
||||||
host = request.config.host,
|
host = request.config.host,
|
||||||
actor = request.actor.id,
|
actor = request.message.actorid,
|
||||||
followid = request.message.id,
|
followid = request.message.id,
|
||||||
accept = True
|
accept = approve
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
## Don't send a follow if the the follow has been rejected
|
||||||
|
if not approve:
|
||||||
|
return
|
||||||
|
|
||||||
|
## Make sure two relays aren't continuously following each other
|
||||||
|
if software in RELAY_SOFTWARE and not request.instance:
|
||||||
|
return
|
||||||
|
|
||||||
# Are Akkoma and Pleroma the only two that expect a follow back?
|
# Are Akkoma and Pleroma the only two that expect a follow back?
|
||||||
# Ignoring only Mastodon for now
|
# Ignoring only Mastodon for now
|
||||||
if software != 'mastodon':
|
if software != 'mastodon':
|
||||||
|
@ -115,24 +137,32 @@ async def handle_follow(request):
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
async def handle_undo(request):
|
async def handle_undo(request, s):
|
||||||
## If the object is not a Follow, forward it
|
## If the object is not a Follow, forward it
|
||||||
if request.message.object.type != 'Follow':
|
if request.message.object.type != 'Follow':
|
||||||
return await handle_forward(request)
|
return await handle_forward(request)
|
||||||
|
|
||||||
if not request.database.del_inbox(request.actor.domain, request.message.id):
|
instance_follow = request.instance.followid
|
||||||
return
|
message_follow = request.message.object.id
|
||||||
|
|
||||||
request.database.save()
|
if person_check(request.actor, request.instance.software):
|
||||||
|
return logging.verbose(f'Non-application actor tried to unfollow: {request.actor.id}')
|
||||||
|
|
||||||
request.app.push_message(
|
if instance_follow and instance_follow != message_follow:
|
||||||
request.actor.shared_inbox,
|
return logging.verbose(f'Followid does not match: {instance_follow}, {message_follow}')
|
||||||
Message.new_unfollow(
|
|
||||||
host = request.config.host,
|
s.delete('instances', id=request.instance.id)
|
||||||
actor = request.actor.id,
|
logging.verbose(f'Removed inbox: {request.instance.inbox}')
|
||||||
follow = request.message
|
|
||||||
|
if request.instance.software != 'mastodon':
|
||||||
|
request.app.push_message(
|
||||||
|
request.actor.shared_inbox,
|
||||||
|
Message.new_unfollow(
|
||||||
|
host = request.config.host,
|
||||||
|
actor = request.actor.id,
|
||||||
|
follow = request.message
|
||||||
|
)
|
||||||
)
|
)
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
processors = {
|
processors = {
|
||||||
|
@ -149,12 +179,26 @@ async def run_processor(request):
|
||||||
if request.message.type not in processors:
|
if request.message.type not in processors:
|
||||||
return
|
return
|
||||||
|
|
||||||
if request.instance and not request.instance.get('software'):
|
with request.database.session as s:
|
||||||
nodeinfo = await request.app.client.fetch_nodeinfo(request.instance['domain'])
|
if request.instance:
|
||||||
|
new_data = {}
|
||||||
|
|
||||||
if nodeinfo:
|
if not request.instance.software:
|
||||||
request.instance['software'] = nodeinfo.sw_name
|
logging.verbose(f'Fetching nodeinfo for instance: {request.instance.domain}')
|
||||||
request.database.save()
|
nodeinfo = await request.app.client.fetch_nodeinfo(request.instance.domain)
|
||||||
|
|
||||||
logging.verbose(f'New "{request.message.type}" from actor: {request.actor.id}')
|
if nodeinfo:
|
||||||
return await processors[request.message.type](request)
|
new_data['software'] = nodeinfo.sw_name
|
||||||
|
|
||||||
|
if not request.instance.actor:
|
||||||
|
logging.verbose(f'Fetching actor for instance: {request.instance.domain}')
|
||||||
|
new_data['actor'] = request.signature.keyid.split('#', 1)[0]
|
||||||
|
|
||||||
|
if not request.instance.actor_data:
|
||||||
|
new_data['actor_data'] = request.actor
|
||||||
|
|
||||||
|
if new_data:
|
||||||
|
s.put_instance(request.actor.domain, **new_data)
|
||||||
|
|
||||||
|
logging.verbose(f'New "{request.message.type}" from actor: {request.actor.id}')
|
||||||
|
return await processors[request.message.type](request, s)
|
||||||
|
|
|
@ -5,6 +5,7 @@ import subprocess
|
||||||
import traceback
|
import traceback
|
||||||
|
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
from . import __version__, misc
|
from . import __version__, misc
|
||||||
from .misc import DotDict, Message, Response
|
from .misc import DotDict, Message, Response
|
||||||
|
@ -24,19 +25,24 @@ if Path(__file__).parent.parent.joinpath('.git').exists():
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
def register_route(method, path):
|
def register_route(method, *paths):
|
||||||
def wrapper(func):
|
def wrapper(func):
|
||||||
routes.append([method, path, func])
|
for path in paths:
|
||||||
|
routes.append([method, path, func])
|
||||||
|
|
||||||
return func
|
return func
|
||||||
|
|
||||||
return wrapper
|
return wrapper
|
||||||
|
|
||||||
|
|
||||||
@register_route('GET', '/')
|
@register_route('GET', '/')
|
||||||
async def home(request):
|
async def home(request, s):
|
||||||
targets = '<br>'.join(request.database.hostnames)
|
hostnames = s.get_hostnames()
|
||||||
note = request.config.note
|
config = s.get_config_all()
|
||||||
count = len(request.database.hostnames)
|
|
||||||
|
targets = '<br>'.join(hostnames)
|
||||||
|
note = config.description
|
||||||
|
count = len(hostnames)
|
||||||
host = request.config.host
|
host = request.config.host
|
||||||
|
|
||||||
text = f"""
|
text = f"""
|
||||||
|
@ -61,28 +67,33 @@ a:hover {{ color: #8AF; }}
|
||||||
return Response.new(text, ctype='html')
|
return Response.new(text, ctype='html')
|
||||||
|
|
||||||
|
|
||||||
@register_route('GET', '/inbox')
|
@register_route('GET', '/actor', '/inbox')
|
||||||
@register_route('GET', '/actor')
|
async def actor(request, s):
|
||||||
async def actor(request):
|
|
||||||
data = Message.new_actor(
|
data = Message.new_actor(
|
||||||
host = request.config.host,
|
host = request.config.host,
|
||||||
pubkey = request.database.signer.pubkey
|
pubkey = request.app.signer.pubkey,
|
||||||
|
name = s.get_config('name'),
|
||||||
|
description = s.get_config('description'),
|
||||||
|
locked = s.get_config('require_approval')
|
||||||
)
|
)
|
||||||
|
|
||||||
return Response.new(data, ctype='activity')
|
return Response.new(data, ctype='activity')
|
||||||
|
|
||||||
|
|
||||||
@register_route('POST', '/inbox')
|
@register_route('POST', '/actor', '/inbox')
|
||||||
@register_route('POST', '/actor')
|
async def inbox(request, s):
|
||||||
async def inbox(request):
|
|
||||||
config = request.config
|
|
||||||
database = request.database
|
|
||||||
|
|
||||||
## reject if missing signature header
|
## reject if missing signature header
|
||||||
if not request.signature:
|
if not request.signature:
|
||||||
logging.verbose('Actor missing signature header')
|
logging.verbose('Actor missing signature header')
|
||||||
raise HTTPUnauthorized(body='missing signature')
|
raise HTTPUnauthorized(body='missing signature')
|
||||||
|
|
||||||
|
domain = urlparse(request.signature.keyid).hostname
|
||||||
|
|
||||||
|
## reject if actor is banned
|
||||||
|
if s.get_ban('domain', domain):
|
||||||
|
logging.verbose(f'Ignored request from banned actor: {domain}')
|
||||||
|
return Response.new_error(403, 'access denied', 'json')
|
||||||
|
|
||||||
try:
|
try:
|
||||||
request['message'] = await request.json(loads=Message.new_from_json)
|
request['message'] = await request.json(loads=Message.new_from_json)
|
||||||
|
|
||||||
|
@ -114,17 +125,8 @@ async def inbox(request):
|
||||||
logging.verbose(f'Failed to fetch actor: {request.signature.keyid}')
|
logging.verbose(f'Failed to fetch actor: {request.signature.keyid}')
|
||||||
return Response.new_error(400, 'failed to fetch actor', 'json')
|
return Response.new_error(400, 'failed to fetch actor', 'json')
|
||||||
|
|
||||||
request['instance'] = request.database.get_inbox(request['actor'].inbox)
|
request['instance'] = s.get_instance(request.actor.shared_inbox)
|
||||||
|
config = s.get_config_all()
|
||||||
## reject if the actor isn't whitelisted while the whiltelist is enabled
|
|
||||||
if config.whitelist_enabled and not config.is_whitelisted(request.actor.domain):
|
|
||||||
logging.verbose(f'Rejected actor for not being in the whitelist: {request.actor.id}')
|
|
||||||
return Response.new_error(403, 'access denied', 'json')
|
|
||||||
|
|
||||||
## reject if actor is banned
|
|
||||||
if request.config.is_banned(request.actor.domain):
|
|
||||||
logging.verbose(f'Ignored request from banned actor: {actor.id}')
|
|
||||||
return Response.new_error(403, 'access denied', 'json')
|
|
||||||
|
|
||||||
## reject if the signature is invalid
|
## reject if the signature is invalid
|
||||||
try:
|
try:
|
||||||
|
@ -136,7 +138,7 @@ async def inbox(request):
|
||||||
return Response.new_error(401, str(e), 'json')
|
return Response.new_error(401, str(e), 'json')
|
||||||
|
|
||||||
## reject if activity type isn't 'Follow' and the actor isn't following
|
## reject if activity type isn't 'Follow' and the actor isn't following
|
||||||
if request.message.type != 'Follow' and not database.get_inbox(request.actor.domain):
|
if request.message.type != 'Follow' and (not request.instance or not request.instance.joined):
|
||||||
logging.verbose(f'Rejected actor for trying to post while not following: {request.actor.id}')
|
logging.verbose(f'Rejected actor for trying to post while not following: {request.actor.id}')
|
||||||
return Response.new_error(401, 'access denied', 'json')
|
return Response.new_error(401, 'access denied', 'json')
|
||||||
|
|
||||||
|
@ -167,16 +169,16 @@ async def webfinger(request):
|
||||||
|
|
||||||
|
|
||||||
@register_route('GET', '/nodeinfo/{version:\d.\d\.json}')
|
@register_route('GET', '/nodeinfo/{version:\d.\d\.json}')
|
||||||
async def nodeinfo(request):
|
async def nodeinfo(request, s):
|
||||||
niversion = request.match_info['version'][:3]
|
niversion = request.match_info['version'][:3]
|
||||||
|
|
||||||
data = dict(
|
data = dict(
|
||||||
name = 'activityrelay',
|
name = 'activityrelay',
|
||||||
version = version,
|
version = version,
|
||||||
protocols = ['activitypub'],
|
protocols = ['activitypub'],
|
||||||
open_regs = not request.config.whitelist_enabled,
|
open_regs = not s.get_config('whitelist'),
|
||||||
users = 1,
|
users = 1,
|
||||||
metadata = {'peers': request.database.hostnames}
|
metadata = {'peers': s.get_hostnames()}
|
||||||
)
|
)
|
||||||
|
|
||||||
if niversion == '2.1':
|
if niversion == '2.1':
|
||||||
|
|
|
@ -3,3 +3,4 @@ aputils@https://git.barkshark.xyz/barkshark/aputils/archive/0.1.3.tar.gz
|
||||||
cachetools>=5.2.0
|
cachetools>=5.2.0
|
||||||
click>=8.1.2
|
click>=8.1.2
|
||||||
pyyaml>=6.0
|
pyyaml>=6.0
|
||||||
|
tinysql[all]@https://git.barkshark.xyz/barkshark/tinysql/archive/0.1.0.tar.gz
|
||||||
|
|
Loading…
Reference in a new issue