Skip to content
Snippets Groups Projects
Commit 244e753d authored by Erik Reid's avatar Erik Reid
Browse files

ruffed

parent 2443fa66
Branches
Tags
No related merge requests found
...@@ -12,7 +12,7 @@ from fastapi import FastAPI ...@@ -12,7 +12,7 @@ from fastapi import FastAPI
from mapping_provider import config, environment from mapping_provider import config, environment
from mapping_provider.api import common, map from mapping_provider.api import common, map
from mapping_provider.backends import correlator, inventory, cache from mapping_provider.backends import cache, correlator, inventory
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
......
...@@ -7,6 +7,7 @@ from pydantic import BaseModel ...@@ -7,6 +7,7 @@ from pydantic import BaseModel
from mapping_provider import config from mapping_provider import config
from mapping_provider.backends import services from mapping_provider.backends import services
router = APIRouter() router = APIRouter()
......
...@@ -6,10 +6,10 @@ The global variable `cache_dir` is set during FastAPI server startup. ...@@ -6,10 +6,10 @@ The global variable `cache_dir` is set during FastAPI server startup.
Could be replaced with a proper caching module in the future, or maybe Could be replaced with a proper caching module in the future, or maybe
just a simple dict. just a simple dict.
""" """
from typing import Any
import json import json
import os
import logging import logging
import os
from typing import Any
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
_cache_dir: str | None = None _cache_dir: str | None = None
...@@ -17,7 +17,7 @@ _cache_dir: str | None = None ...@@ -17,7 +17,7 @@ _cache_dir: str | None = None
def init(cache_dir: str) -> None: def init(cache_dir: str) -> None:
global _cache_dir global _cache_dir
assert _cache_dir is None, f"cache_dir has already been initialized" assert _cache_dir is None, "cache_dir has already been initialized"
_cache_dir = cache_dir _cache_dir = cache_dir
logger.debug(f"set cache directory: {_cache_dir}") logger.debug(f"set cache directory: {_cache_dir}")
...@@ -26,7 +26,7 @@ def set(filename: str, data: dict[str, Any]) -> None: ...@@ -26,7 +26,7 @@ def set(filename: str, data: dict[str, Any]) -> None:
""" """
data must be a JSON-serializable dict. data must be a JSON-serializable dict.
""" """
assert _cache_dir is not None, f"cache_dir hasn't been initialized" assert _cache_dir is not None, "cache_dir hasn't been initialized"
with open(os.path.join(_cache_dir, filename), 'w') as f: with open(os.path.join(_cache_dir, filename), 'w') as f:
f.write(json.dumps(data)) f.write(json.dumps(data))
logger.debug(f"wrote cached data: {filename}") logger.debug(f"wrote cached data: {filename}")
...@@ -36,7 +36,7 @@ def get(filename: str) -> Any: ...@@ -36,7 +36,7 @@ def get(filename: str) -> Any:
""" """
Loads the cached data, parses it as json & returns the object. Loads the cached data, parses it as json & returns the object.
""" """
assert _cache_dir is not None, f"cache_dir hasn't been initialized" assert _cache_dir is not None, "cache_dir hasn't been initialized"
with open(os.path.join(_cache_dir, filename), 'r') as f: with open(os.path.join(_cache_dir, filename)) as f:
logger.debug(f"reading cached data: {filename}") logger.debug(f"reading cached data: {filename}")
return json.load(f) return json.load(f)
import json
import logging import logging
import os import os
import socket import socket
import threading import threading
import time import time
from functools import partial
from typing import TYPE_CHECKING, Any from typing import TYPE_CHECKING, Any
if TYPE_CHECKING: if TYPE_CHECKING:
...@@ -12,8 +10,8 @@ if TYPE_CHECKING: ...@@ -12,8 +10,8 @@ if TYPE_CHECKING:
from pika.exchange_type import ExchangeType from pika.exchange_type import ExchangeType
from .rmq.exchange import RabbitMQClient, default_rmq_connection_params
from . import cache from . import cache
from .rmq.exchange import RabbitMQClient, default_rmq_connection_params
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
......
import json import concurrent.futures
import logging import logging
import os
import time import time
from threading import Event from threading import Event
import concurrent.futures
from typing import Any from typing import Any
import requests import requests
......
from pydantic import BaseModel
from typing import Any, Generator
import logging import logging
from collections.abc import Generator
from pydantic import BaseModel
from . import cache, correlator, inventory from . import cache, correlator, inventory
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
...@@ -29,7 +31,7 @@ class ServiceList(BaseModel): ...@@ -29,7 +31,7 @@ class ServiceList(BaseModel):
services: list[Service] services: list[Service]
def _services(service_type: str | None = None) -> Generator[Service, None, None]: def _services(service_type: str | None = None) -> Generator[Service]:
""" """
load the cached backend data and yield map service records load the cached backend data and yield map service records
...@@ -45,7 +47,7 @@ def _services(service_type: str | None = None) -> Generator[Service, None, None] ...@@ -45,7 +47,7 @@ def _services(service_type: str | None = None) -> Generator[Service, None, None]
logger.exception('not enough data available to build the service list') logger.exception('not enough data available to build the service list')
return return
def _get_down_correlator_services() -> Generator[str, None, None]: def _get_down_correlator_services() -> Generator[str]:
for _e in correlator_state['endpoints']: for _e in correlator_state['endpoints']:
if _e['up']: if _e['up']:
continue continue
......
...@@ -7,7 +7,8 @@ import pytest ...@@ -7,7 +7,8 @@ import pytest
from fastapi.testclient import TestClient from fastapi.testclient import TestClient
from mapping_provider import create_app from mapping_provider import create_app
from mapping_provider.backends import cache, inventory, correlator from mapping_provider.backends import cache, correlator, inventory
from .common import load_test_data from .common import load_test_data
......
import json
import os
import tempfile import tempfile
from mapping_provider.backends import cache, correlator from mapping_provider.backends import cache, correlator
......
...@@ -2,7 +2,8 @@ import tempfile ...@@ -2,7 +2,8 @@ import tempfile
import responses import responses
from mapping_provider.backends import inventory, cache from mapping_provider.backends import cache, inventory
from .common import load_test_data from .common import load_test_data
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment