From bd9a2f41817f6c8aee9b7a31da9c0807ced2ea8d Mon Sep 17 00:00:00 2001
From: Georg Krause <mail@georg-krause.net>
Date: Sun, 27 Nov 2022 14:01:55 +0100
Subject: [PATCH] style: Format everything with black

---
 funkwhale_network/cli.py    |  8 ++++----
 funkwhale_network/routes.py |  2 +-
 funkwhale_network/worker.py | 10 +++++-----
 tests/test_routes.py        | 13 ++++++++-----
 4 files changed, 18 insertions(+), 15 deletions(-)

diff --git a/funkwhale_network/cli.py b/funkwhale_network/cli.py
index 68a631f..dbf0c57 100644
--- a/funkwhale_network/cli.py
+++ b/funkwhale_network/cli.py
@@ -305,15 +305,15 @@ def aggregate_crawl_results(domains_info):
 @click.option("-v", "--verbose", is_flag=True)
 @click.option("--check", is_flag=True)
 def start(*, check, verbose):
-    #worker = arq.worker.import_string("funkwhale_network.worker", "Worker")
-    #logging.config.dictConfig(worker.logging_config(verbose))
+    # worker = arq.worker.import_string("funkwhale_network.worker", "Worker")
+    # logging.config.dictConfig(worker.logging_config(verbose))
 
     if check:
         pass
-        #exit(worker.check_health())
+        # exit(worker.check_health())
     else:
         pass
-        #arq.RunWorkerProcess("funkwhale_network.worker", "Worker", burst=False)
+        # arq.RunWorkerProcess("funkwhale_network.worker", "Worker", burst=False)
 
 
 def main():
diff --git a/funkwhale_network/routes.py b/funkwhale_network/routes.py
index bd26f69..5226cc9 100644
--- a/funkwhale_network/routes.py
+++ b/funkwhale_network/routes.py
@@ -26,7 +26,7 @@ domain_filters = {
 
 
 def validate_domain(raw):
-    raw = raw.split('://')[-1]
+    raw = raw.split("://")[-1]
     if not raw:
         raise ValueError()
 
diff --git a/funkwhale_network/worker.py b/funkwhale_network/worker.py
index bc3a5e0..8a2cace 100644
--- a/funkwhale_network/worker.py
+++ b/funkwhale_network/worker.py
@@ -9,20 +9,20 @@ from arq import cron
 
 
 async def poll(ctx, domain):
-    session: ClientSession = ctx['session']
+    session: ClientSession = ctx["session"]
     pool = await db.get_pool(settings.DB_DSN)
     async with pool as conn:
         return await crawler.check(conn=conn, session=session, domain=domain)
 
+
 async def update_all(ctx):
     pool = await db.get_pool(settings.DB_DSN)
     for check in await db.get_latest_check_by_domain(pool):
         await poll(ctx, check["domain"])
-    
+
+
 class WorkerSettings:
-    cron_jobs = [
-        cron(update_all, minute=None)
-    ]
+    cron_jobs = [cron(update_all, minute=None)]
     max_concurrent_tasks = 20
     shutdown_delay = 5
     timeout_seconds = 15
diff --git a/tests/test_routes.py b/tests/test_routes.py
index 0d8103e..a1a9cb7 100644
--- a/tests/test_routes.py
+++ b/tests/test_routes.py
@@ -132,10 +132,13 @@ async def test_domains_rss(db_conn, factories, client, mocker):
     assert resp.headers["content-type"] == "application/rss+xml"
 
 
-@pytest.mark.parametrize('input,expected', [
-    ('example.com', 'example.com'),
-    ('http://example.com', 'example.com'),
-    ('https://example.com', 'example.com'),
-])
+@pytest.mark.parametrize(
+    "input,expected",
+    [
+        ("example.com", "example.com"),
+        ("http://example.com", "example.com"),
+        ("https://example.com", "example.com"),
+    ],
+)
 def test_validate_domain(input, expected):
     assert routes.validate_domain(input) == expected
-- 
GitLab