feat: initial commit
This commit is contained in:
commit
075939142f
63 changed files with 9494 additions and 0 deletions
12
.env.example
Normal file
12
.env.example
Normal file
|
|
@ -0,0 +1,12 @@
|
|||
DATABASE_URL=postgresql://user@localhost:5432/lawa
|
||||
DATABASE_ASYNC_URL=postgresql+asyncpg://user@localhost:5432/lawa
|
||||
|
||||
ENVIRONMENT=LOCAL
|
||||
|
||||
CORS_HEADERS=["*"]
|
||||
CORS_ORIGINS=["http://localhost:3000"]
|
||||
|
||||
OIDC_ENDPOINT=https://www.example.com/.well-known/openid-configuration
|
||||
OIDC_CLIENT_ID=6c1243fd-00c3-4a3b-9081-5d8403352612
|
||||
|
||||
OPENTOFU_PATH=/opt/homebrew/bin/tofu
|
||||
143
.gitignore
vendored
Normal file
143
.gitignore
vendored
Normal file
|
|
@ -0,0 +1,143 @@
|
|||
# Byte-compiled / optimized / DLL files
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*$py.class
|
||||
|
||||
# C extensions
|
||||
*.so
|
||||
|
||||
# Distribution / packaging
|
||||
.Python
|
||||
build/
|
||||
develop-eggs/
|
||||
dist/
|
||||
downloads/
|
||||
eggs/
|
||||
.eggs/
|
||||
lib/
|
||||
lib64/
|
||||
parts/
|
||||
sdist/
|
||||
var/
|
||||
wheels/
|
||||
pip-wheel-metadata/
|
||||
share/python-wheels/
|
||||
*.egg-info/
|
||||
.installed.cfg
|
||||
*.egg
|
||||
MANIFEST
|
||||
|
||||
# PyInstaller
|
||||
# Usually these files are written by a python script from a template
|
||||
# before PyInstaller builds the exe, so as to inject date/other infos into it.
|
||||
*.manifest
|
||||
*.spec
|
||||
|
||||
# Installer logs
|
||||
pip-log.txt
|
||||
pip-delete-this-directory.txt
|
||||
|
||||
# Unit test / coverage reports
|
||||
htmlcov/
|
||||
.tox/
|
||||
.nox/
|
||||
.coverage
|
||||
.coverage.*
|
||||
.cache
|
||||
nosetests.xml
|
||||
coverage.xml
|
||||
*.cover
|
||||
*.py,cover
|
||||
.hypothesis/
|
||||
.pytest_cache/
|
||||
|
||||
# Translations
|
||||
*.mo
|
||||
*.pot
|
||||
|
||||
# Django stuff:
|
||||
*.log
|
||||
local_settings.py
|
||||
db.sqlite3
|
||||
db.sqlite3-journal
|
||||
|
||||
# Flask stuff:
|
||||
instance/
|
||||
.webassets-cache
|
||||
|
||||
# Scrapy stuff:
|
||||
.scrapy
|
||||
|
||||
# Sphinx documentation
|
||||
docs/_build/
|
||||
|
||||
# PyBuilder
|
||||
target/
|
||||
|
||||
# Jupyter Notebook
|
||||
.ipynb_checkpoints
|
||||
|
||||
# IPython
|
||||
profile_default/
|
||||
ipython_config.py
|
||||
|
||||
# pyenv
|
||||
.python-version
|
||||
|
||||
# pipenv
|
||||
# According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
|
||||
# However, in case of collaboration, if having platform-specific dependencies or dependencies
|
||||
# having no cross-platform support, pipenv may install dependencies that don't work, or not
|
||||
# install all needed dependencies.
|
||||
#Pipfile.lock
|
||||
|
||||
# PEP 582; used by e.g. github.com/David-OConnor/pyflow
|
||||
__pypackages__/
|
||||
|
||||
# Celery stuff
|
||||
celerybeat-schedule
|
||||
celerybeat.pid
|
||||
|
||||
# SageMath parsed files
|
||||
*.sage.py
|
||||
|
||||
# Environments
|
||||
.env
|
||||
.env.prefect
|
||||
.venv
|
||||
env/
|
||||
venv/
|
||||
ENV/
|
||||
env.bak/
|
||||
venv.bak/
|
||||
|
||||
# Spyder project settings
|
||||
.spyderproject
|
||||
.spyproject
|
||||
|
||||
# Rope project settings
|
||||
.ropeproject
|
||||
|
||||
# mkdocs documentation
|
||||
/site
|
||||
|
||||
# mypy
|
||||
.mypy_cache/
|
||||
.dmypy.json
|
||||
dmypy.json
|
||||
|
||||
# Pyre type checker
|
||||
.pyre/
|
||||
postgres-data*
|
||||
node_modules/
|
||||
.DS_Store
|
||||
|
||||
# PyCharm
|
||||
.idea/
|
||||
backup.dump
|
||||
backup.dump.gz
|
||||
.vscode
|
||||
|
||||
# Terraform
|
||||
.terraform
|
||||
.terraform.lock.hcl
|
||||
48
Dockerfile
Normal file
48
Dockerfile
Normal file
|
|
@ -0,0 +1,48 @@
|
|||
FROM python:3.12-bullseye
|
||||
|
||||
ENV PYTHONDONTWRITEBYTECODE=1 \
|
||||
PYTHONUNBUFFERED=1 \
|
||||
PYTHONIOENCODING=utf-8 \
|
||||
# pip
|
||||
PIP_NO_CACHE_DIR=1 \
|
||||
PIP_DISABLE_PIP_VERSION_CHECK=1 \
|
||||
PIP_DEFAULT_TIMEOUT=100 \
|
||||
PIP_ROOT_USER_ACTION=ignore \
|
||||
# poetry
|
||||
POETRY_VERSION=1.8.3 \
|
||||
POETRY_NO_INTERACTION=1 \
|
||||
POETRY_VIRTUALENVS_CREATE=false \
|
||||
POETRY_CACHE_DIR='/var/cache/pypoetry' \
|
||||
POETRY_HOME='/usr/local' \
|
||||
# app
|
||||
PROMETHEUS_MULTIPROC_DIR=/tmp/prometheus_multiproc_dir
|
||||
|
||||
SHELL ["/bin/bash", "-eo", "pipefail", "-c"]
|
||||
|
||||
RUN apt-get update && \
|
||||
apt-get install -y gcc libpq-dev curl && \
|
||||
curl -sSL 'https://install.python-poetry.org' | python - \
|
||||
&& poetry --version \
|
||||
apt clean && \
|
||||
rm -rf /var/cache/apt/*
|
||||
|
||||
COPY poetry.lock pyproject.toml /src/
|
||||
|
||||
WORKDIR /src
|
||||
|
||||
RUN --mount=type=cache,target="$POETRY_CACHE_DIR" \
|
||||
echo "$ENVIRONMENT" \
|
||||
# Install deps:
|
||||
&& poetry run pip install -U pip \
|
||||
&& poetry install \
|
||||
--no-interaction --no-ansi --sync --with prod
|
||||
|
||||
COPY . .
|
||||
|
||||
RUN useradd -m -d /src -s /bin/bash app \
|
||||
&& chown -R app:app /src/* && chown -R app:app /src \
|
||||
&& chmod +x entrypoints/* \
|
||||
&& rm -rf /tmp/prometheus_multiproc_dir && mkdir -p /tmp/prometheus_multiproc_dir \
|
||||
&& chown -R app:app /tmp/prometheus_multiproc_dir
|
||||
|
||||
USER app
|
||||
21
LICENSE
Normal file
21
LICENSE
Normal file
|
|
@ -0,0 +1,21 @@
|
|||
(C) SR2 Communications Limited 2021-2025.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without
|
||||
modification, are permitted provided that the following conditions are met:
|
||||
|
||||
1. Redistributions of source code must retain the above copyright notice,
|
||||
this list of conditions and the following disclaimer.
|
||||
2. Redistributions in binary form must reproduce the above copyright
|
||||
notice, this list of conditions and the following disclaimer in the
|
||||
documentation and/or other materials provided with the distribution.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
||||
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
99
README.md
Normal file
99
README.md
Normal file
|
|
@ -0,0 +1,99 @@
|
|||
# ilo pali lili
|
||||
|
||||
Backend for orchestrating deployment of censorship circumvention resources.
|
||||
|
||||
## 🪓 Local Development
|
||||
|
||||
### Setup just
|
||||
MacOS:
|
||||
```shell
|
||||
brew install just
|
||||
```
|
||||
|
||||
Debian/Ubuntu:
|
||||
```shell
|
||||
apt install just
|
||||
````
|
||||
|
||||
Others: [link](https://github.com/casey/just?tab=readme-ov-file#packages)
|
||||
|
||||
### Setup poetry
|
||||
```shell
|
||||
pip install poetry
|
||||
```
|
||||
|
||||
Other ways: [link](https://python-poetry.org/docs/#installation)
|
||||
|
||||
### Setup Postgres (16.3)
|
||||
```shell
|
||||
just up
|
||||
```
|
||||
### Copy the environment file and install dependencies
|
||||
1. `cp .env.example .env`
|
||||
2. `poetry install`
|
||||
|
||||
### Run the uvicorn server
|
||||
With default settings:
|
||||
```shell
|
||||
just run
|
||||
```
|
||||
With extra configs (e.g. logging file)
|
||||
```shell
|
||||
just run --log-config logging.ini
|
||||
```
|
||||
|
||||
### 🔍 Linters
|
||||
Format the code with `ruff --fix` and `ruff format`
|
||||
```shell
|
||||
just lint
|
||||
```
|
||||
|
||||
### Migrations
|
||||
- Create an automatic migration from changes in `src/database.py`
|
||||
```shell
|
||||
just mm *migration_name*
|
||||
```
|
||||
- Run migrations
|
||||
```shell
|
||||
just migrate
|
||||
```
|
||||
- Downgrade migrations
|
||||
```shell
|
||||
just downgrade downgrade -1 # or -2 or base or hash of the migration
|
||||
```
|
||||
|
||||
## 🏃🏻♀️➡️ Deployment
|
||||
Deployment is done with Docker and Gunicorn. The Dockerfile is optimized for small size and fast builds with a non-root user. The gunicorn configuration is set to use the number of workers based on the number of CPU cores.
|
||||
|
||||
Example of running the app with docker compose:
|
||||
```shell
|
||||
docker compose -f docker-compose.prod.yml up -d --build
|
||||
```
|
||||
|
||||
## 📜 Conventions
|
||||
|
||||
This project uses the
|
||||
[FastAPI best practices](https://github.com/zhanymkanov/fastapi-best-practices)
|
||||
documented by GitHub user zhanymkanov and was originally created from their
|
||||
[template](https://github.com/zhanymkanov/fastapi_production_template).
|
||||
|
||||
|
||||
## 🖋️ Copyright
|
||||
|
||||
Copyright © 2025 SR2 Communications Limited.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without modification, are permitted provided that the
|
||||
following conditions are met:
|
||||
|
||||
1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following
|
||||
disclaimer.
|
||||
2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following
|
||||
disclaimer in the documentation and/or other materials provided with the distribution.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS “AS IS” AND ANY EXPRESS OR IMPLIED WARRANTIES,
|
||||
INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
||||
WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
||||
THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
104
alembic.ini
Normal file
104
alembic.ini
Normal file
|
|
@ -0,0 +1,104 @@
|
|||
# A generic, single database configuration.
|
||||
|
||||
[alembic]
|
||||
# path to migration scripts
|
||||
script_location = alembic
|
||||
|
||||
# template used to generate migration file names; The default value is %%(rev)s_%%(slug)s
|
||||
# Uncomment the line below if you want the files to be prepended with date and time
|
||||
# see https://alembic.sqlalchemy.org/en/latest/tutorial.html#editing-the-ini-file
|
||||
# for all available tokens
|
||||
file_template = %%(year)d-%%(month).2d-%%(day).2d_%%(slug)s
|
||||
|
||||
# sys.path path, will be prepended to sys.path if present.
|
||||
# defaults to the current working directory.
|
||||
prepend_sys_path = .
|
||||
|
||||
# timezone to use when rendering the date within the migration file
|
||||
# as well as the filename.
|
||||
# If specified, requires the python-dateutil library that can be
|
||||
# installed by adding `alembic[tz]` to the pip requirements
|
||||
# string value is passed to dateutil.tz.gettz()
|
||||
# leave blank for localtime
|
||||
# timezone =
|
||||
|
||||
# max length of characters to apply to the
|
||||
# "slug" field
|
||||
# truncate_slug_length = 40
|
||||
|
||||
# set to 'true' to run the environment during
|
||||
# the 'revision' command, regardless of autogenerate
|
||||
# revision_environment = false
|
||||
|
||||
# set to 'true' to allow .pyc and .pyo files without
|
||||
# a source .py file to be detected as revisions in the
|
||||
# versions/ directory
|
||||
# sourceless = false
|
||||
|
||||
# version location specification; This defaults
|
||||
# to alembic/versions. When using multiple version
|
||||
# directories, initial revisions must be specified with --version-path.
|
||||
# The path separator used here should be the separator specified by "version_path_separator" below.
|
||||
# version_locations = %(here)s/bar:%(here)s/bat:alembic/versions
|
||||
|
||||
# version path separator; As mentioned above, this is the character used to split
|
||||
# version_locations. The default within new alembic.ini files is "os", which uses os.pathsep.
|
||||
# If this key is omitted entirely, it falls back to the legacy behavior of splitting on spaces and/or commas.
|
||||
# Valid values for version_path_separator are:
|
||||
#
|
||||
# version_path_separator = :
|
||||
# version_path_separator = ;
|
||||
# version_path_separator = space
|
||||
version_path_separator = os # Use os.pathsep. Default configuration used for new projects.
|
||||
|
||||
# the output encoding used when revision files
|
||||
# are written from script.py.mako
|
||||
# output_encoding = utf-8
|
||||
|
||||
sqlalchemy.url = driver://user:pass@localhost/dbname
|
||||
|
||||
|
||||
[post_write_hooks]
|
||||
hooks = ruff_format,ruff
|
||||
ruff_format.type = exec
|
||||
ruff_format.executable = ruff
|
||||
ruff_format.options = format REVISION_SCRIPT_FILENAME
|
||||
|
||||
ruff.type = exec
|
||||
ruff.executable = ruff
|
||||
ruff.options = check --fix REVISION_SCRIPT_FILENAME
|
||||
|
||||
# Logging configuration
|
||||
[loggers]
|
||||
keys = root,sqlalchemy,alembic
|
||||
|
||||
[handlers]
|
||||
keys = console
|
||||
|
||||
[formatters]
|
||||
keys = generic
|
||||
|
||||
[logger_root]
|
||||
level = WARN
|
||||
handlers = console
|
||||
qualname =
|
||||
|
||||
[logger_sqlalchemy]
|
||||
level = WARN
|
||||
handlers =
|
||||
qualname = sqlalchemy.engine
|
||||
|
||||
[logger_alembic]
|
||||
level = INFO
|
||||
handlers =
|
||||
qualname = alembic
|
||||
|
||||
[handler_console]
|
||||
class = StreamHandler
|
||||
args = (sys.stderr,)
|
||||
level = NOTSET
|
||||
formatter = generic
|
||||
|
||||
[formatter_generic]
|
||||
format = %(levelname)-5.5s [%(name)s] %(message)s
|
||||
datefmt = %H:%M:%S
|
||||
1
alembic/README
Normal file
1
alembic/README
Normal file
|
|
@ -0,0 +1 @@
|
|||
Generic single-database configuration.
|
||||
BIN
alembic/__pycache__/env.cpython-312.pyc
Normal file
BIN
alembic/__pycache__/env.cpython-312.pyc
Normal file
Binary file not shown.
92
alembic/env.py
Normal file
92
alembic/env.py
Normal file
|
|
@ -0,0 +1,92 @@
|
|||
from logging.config import fileConfig
|
||||
|
||||
from sqlalchemy import engine_from_config, pool
|
||||
|
||||
from alembic import context
|
||||
|
||||
import src.main as _
|
||||
from src.config import settings
|
||||
from src.database import metadata
|
||||
|
||||
# this is the Alembic Config object, which provides
|
||||
# access to the values within the .ini file in use.
|
||||
config = context.config
|
||||
|
||||
# Interpret the config file for Python logging.
|
||||
# This line sets up loggers basically.
|
||||
if config.config_file_name is not None:
|
||||
fileConfig(config.config_file_name)
|
||||
|
||||
# add your model's MetaData object here
|
||||
# for 'autogenerate' support
|
||||
# from myapp import mymodel
|
||||
# target_metadata = mymodel.Base.metadata
|
||||
target_metadata = metadata
|
||||
|
||||
# other values from the config, defined by the needs of env.py,
|
||||
# can be acquired:
|
||||
# my_important_option = config.get_main_option("my_important_option")
|
||||
# ... etc.
|
||||
DATABASE_URL = str(settings.DATABASE_URL)
|
||||
|
||||
db_driver = settings.DATABASE_URL.scheme
|
||||
db_driver_parts = db_driver.split("+")
|
||||
if len(db_driver_parts) > 1: # e.g. postgresql+asyncpg
|
||||
sync_scheme = db_driver_parts[0].strip()
|
||||
DATABASE_URL = DATABASE_URL.replace( # replace with sync driver
|
||||
db_driver, sync_scheme
|
||||
)
|
||||
|
||||
config.set_main_option("sqlalchemy.url", DATABASE_URL)
|
||||
config.compare_type = True
|
||||
config.compare_server_default = True
|
||||
|
||||
|
||||
def run_migrations_offline() -> None:
|
||||
"""Run migrations in 'offline' mode.
|
||||
|
||||
This configures the context with just a URL
|
||||
and not an Engine, though an Engine is acceptable
|
||||
here as well. By skipping the Engine creation
|
||||
we don't even need a DBAPI to be available.
|
||||
|
||||
Calls to context.execute() here emit the given string to the
|
||||
script output.
|
||||
|
||||
"""
|
||||
url = config.get_main_option("sqlalchemy.url")
|
||||
context.configure(
|
||||
url=url,
|
||||
target_metadata=target_metadata,
|
||||
literal_binds=True,
|
||||
dialect_opts={"paramstyle": "named"},
|
||||
)
|
||||
|
||||
with context.begin_transaction():
|
||||
context.run_migrations()
|
||||
|
||||
|
||||
def run_migrations_online() -> None:
|
||||
"""Run migrations in 'online' mode.
|
||||
|
||||
In this scenario we need to create an Engine
|
||||
and associate a connection with the context.
|
||||
|
||||
"""
|
||||
connectable = engine_from_config(
|
||||
config.get_section(config.config_ini_section),
|
||||
prefix="sqlalchemy.",
|
||||
poolclass=pool.NullPool,
|
||||
)
|
||||
|
||||
with connectable.connect() as connection:
|
||||
context.configure(connection=connection, target_metadata=target_metadata)
|
||||
|
||||
with context.begin_transaction():
|
||||
context.run_migrations()
|
||||
|
||||
|
||||
if context.is_offline_mode():
|
||||
run_migrations_offline()
|
||||
else:
|
||||
run_migrations_online()
|
||||
24
alembic/script.py.mako
Normal file
24
alembic/script.py.mako
Normal file
|
|
@ -0,0 +1,24 @@
|
|||
"""${message}
|
||||
|
||||
Revision ID: ${up_revision}
|
||||
Revises: ${down_revision | comma,n}
|
||||
Create Date: ${create_date}
|
||||
|
||||
"""
|
||||
from alembic import op
|
||||
import sqlalchemy as sa
|
||||
${imports if imports else ""}
|
||||
|
||||
# revision identifiers, used by Alembic.
|
||||
revision = ${repr(up_revision)}
|
||||
down_revision = ${repr(down_revision)}
|
||||
branch_labels = ${repr(branch_labels)}
|
||||
depends_on = ${repr(depends_on)}
|
||||
|
||||
|
||||
def upgrade() -> None:
|
||||
${upgrades if upgrades else "pass"}
|
||||
|
||||
|
||||
def downgrade() -> None:
|
||||
${downgrades if downgrades else "pass"}
|
||||
0
alembic/versions/.gitkeep
Normal file
0
alembic/versions/.gitkeep
Normal file
BIN
alembic/versions/__pycache__/2025-12-14_hi.cpython-312.pyc
Normal file
BIN
alembic/versions/__pycache__/2025-12-14_hi.cpython-312.pyc
Normal file
Binary file not shown.
16
docker-compose.prod.yml
Normal file
16
docker-compose.prod.yml
Normal file
|
|
@ -0,0 +1,16 @@
|
|||
version: '3.7'
|
||||
|
||||
services:
|
||||
app:
|
||||
container_name: app
|
||||
image: app
|
||||
env_file:
|
||||
- .env
|
||||
ports:
|
||||
- "8000:8000"
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile
|
||||
command:
|
||||
- entrypoints/gunicorn.sh
|
||||
|
||||
19
docker-compose.yml
Normal file
19
docker-compose.yml
Normal file
|
|
@ -0,0 +1,19 @@
|
|||
version: '3.7'
|
||||
|
||||
services:
|
||||
app_db:
|
||||
container_name: app_db
|
||||
hostname: app_db
|
||||
image: library/postgres:17.5
|
||||
environment:
|
||||
- POSTGRES_USER=app
|
||||
- POSTGRES_PASSWORD=app
|
||||
- POSTGRES_DB=app
|
||||
volumes:
|
||||
- app_db_data:/var/lib/postgresql/data
|
||||
ports:
|
||||
- "5432:5432"
|
||||
|
||||
volumes:
|
||||
app_db_data:
|
||||
driver: "local"
|
||||
16
entrypoints/gunicorn.sh
Executable file
16
entrypoints/gunicorn.sh
Executable file
|
|
@ -0,0 +1,16 @@
|
|||
#!/usr/bin/env bash
|
||||
|
||||
set -e
|
||||
|
||||
DEFAULT_MODULE_NAME=src.main
|
||||
|
||||
MODULE_NAME=${MODULE_NAME:-$DEFAULT_MODULE_NAME}
|
||||
VARIABLE_NAME=${VARIABLE_NAME:-app}
|
||||
export APP_MODULE=${APP_MODULE:-"$MODULE_NAME:$VARIABLE_NAME"}
|
||||
|
||||
DEFAULT_GUNICORN_CONF=/src/gunicorn/gunicorn_conf.py
|
||||
export GUNICORN_CONF=${GUNICORN_CONF:-$DEFAULT_GUNICORN_CONF}
|
||||
export WORKER_CLASS=${WORKER_CLASS:-"uvicorn.workers.UvicornWorker"}
|
||||
|
||||
# Start Gunicorn
|
||||
gunicorn --forwarded-allow-ips "*" -k "$WORKER_CLASS" -c "$GUNICORN_CONF" "$APP_MODULE"
|
||||
24
frontend/.gitignore
vendored
Normal file
24
frontend/.gitignore
vendored
Normal file
|
|
@ -0,0 +1,24 @@
|
|||
# Logs
|
||||
logs
|
||||
*.log
|
||||
npm-debug.log*
|
||||
yarn-debug.log*
|
||||
yarn-error.log*
|
||||
pnpm-debug.log*
|
||||
lerna-debug.log*
|
||||
|
||||
node_modules
|
||||
dist
|
||||
dist-ssr
|
||||
*.local
|
||||
|
||||
# Editor directories and files
|
||||
.vscode/*
|
||||
!.vscode/extensions.json
|
||||
.idea
|
||||
.DS_Store
|
||||
*.suo
|
||||
*.ntvs*
|
||||
*.njsproj
|
||||
*.sln
|
||||
*.sw?
|
||||
73
frontend/README.md
Normal file
73
frontend/README.md
Normal file
|
|
@ -0,0 +1,73 @@
|
|||
# React + TypeScript + Vite
|
||||
|
||||
This template provides a minimal setup to get React working in Vite with HMR and some ESLint rules.
|
||||
|
||||
Currently, two official plugins are available:
|
||||
|
||||
- [@vitejs/plugin-react](https://github.com/vitejs/vite-plugin-react/blob/main/packages/plugin-react) uses [Babel](https://babeljs.io/) (or [oxc](https://oxc.rs) when used in [rolldown-vite](https://vite.dev/guide/rolldown)) for Fast Refresh
|
||||
- [@vitejs/plugin-react-swc](https://github.com/vitejs/vite-plugin-react/blob/main/packages/plugin-react-swc) uses [SWC](https://swc.rs/) for Fast Refresh
|
||||
|
||||
## React Compiler
|
||||
|
||||
The React Compiler is not enabled on this template because of its impact on dev & build performances. To add it, see [this documentation](https://react.dev/learn/react-compiler/installation).
|
||||
|
||||
## Expanding the ESLint configuration
|
||||
|
||||
If you are developing a production application, we recommend updating the configuration to enable type-aware lint rules:
|
||||
|
||||
```js
|
||||
export default defineConfig([
|
||||
globalIgnores(['dist']),
|
||||
{
|
||||
files: ['**/*.{ts,tsx}'],
|
||||
extends: [
|
||||
// Other configs...
|
||||
|
||||
// Remove tseslint.configs.recommended and replace with this
|
||||
tseslint.configs.recommendedTypeChecked,
|
||||
// Alternatively, use this for stricter rules
|
||||
tseslint.configs.strictTypeChecked,
|
||||
// Optionally, add this for stylistic rules
|
||||
tseslint.configs.stylisticTypeChecked,
|
||||
|
||||
// Other configs...
|
||||
],
|
||||
languageOptions: {
|
||||
parserOptions: {
|
||||
project: ['./tsconfig.node.json', './tsconfig.app.json'],
|
||||
tsconfigRootDir: import.meta.dirname,
|
||||
},
|
||||
// other options...
|
||||
},
|
||||
},
|
||||
])
|
||||
```
|
||||
|
||||
You can also install [eslint-plugin-react-x](https://github.com/Rel1cx/eslint-react/tree/main/packages/plugins/eslint-plugin-react-x) and [eslint-plugin-react-dom](https://github.com/Rel1cx/eslint-react/tree/main/packages/plugins/eslint-plugin-react-dom) for React-specific lint rules:
|
||||
|
||||
```js
|
||||
// eslint.config.js
|
||||
import reactX from 'eslint-plugin-react-x'
|
||||
import reactDom from 'eslint-plugin-react-dom'
|
||||
|
||||
export default defineConfig([
|
||||
globalIgnores(['dist']),
|
||||
{
|
||||
files: ['**/*.{ts,tsx}'],
|
||||
extends: [
|
||||
// Other configs...
|
||||
// Enable lint rules for React
|
||||
reactX.configs['recommended-typescript'],
|
||||
// Enable lint rules for React DOM
|
||||
reactDom.configs.recommended,
|
||||
],
|
||||
languageOptions: {
|
||||
parserOptions: {
|
||||
project: ['./tsconfig.node.json', './tsconfig.app.json'],
|
||||
tsconfigRootDir: import.meta.dirname,
|
||||
},
|
||||
// other options...
|
||||
},
|
||||
},
|
||||
])
|
||||
```
|
||||
23
frontend/eslint.config.js
Normal file
23
frontend/eslint.config.js
Normal file
|
|
@ -0,0 +1,23 @@
|
|||
import js from '@eslint/js'
|
||||
import globals from 'globals'
|
||||
import reactHooks from 'eslint-plugin-react-hooks'
|
||||
import reactRefresh from 'eslint-plugin-react-refresh'
|
||||
import tseslint from 'typescript-eslint'
|
||||
import { defineConfig, globalIgnores } from 'eslint/config'
|
||||
|
||||
export default defineConfig([
|
||||
globalIgnores(['dist']),
|
||||
{
|
||||
files: ['**/*.{ts,tsx}'],
|
||||
extends: [
|
||||
js.configs.recommended,
|
||||
tseslint.configs.recommended,
|
||||
reactHooks.configs.flat.recommended,
|
||||
reactRefresh.configs.vite,
|
||||
],
|
||||
languageOptions: {
|
||||
ecmaVersion: 2020,
|
||||
globals: globals.browser,
|
||||
},
|
||||
},
|
||||
])
|
||||
13
frontend/index.html
Normal file
13
frontend/index.html
Normal file
|
|
@ -0,0 +1,13 @@
|
|||
<!doctype html>
|
||||
<html lang="en">
|
||||
<head>
|
||||
<meta charset="UTF-8" />
|
||||
<link rel="icon" type="image/svg+xml" href="/vite.svg" />
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||
<title>frontend</title>
|
||||
</head>
|
||||
<body>
|
||||
<div id="root"></div>
|
||||
<script type="module" src="/src/main.tsx"></script>
|
||||
</body>
|
||||
</html>
|
||||
4217
frontend/package-lock.json
generated
Normal file
4217
frontend/package-lock.json
generated
Normal file
File diff suppressed because it is too large
Load diff
32
frontend/package.json
Normal file
32
frontend/package.json
Normal file
|
|
@ -0,0 +1,32 @@
|
|||
{
|
||||
"name": "frontend",
|
||||
"private": true,
|
||||
"version": "0.0.0",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "vite",
|
||||
"build": "tsc -b && vite build",
|
||||
"lint": "eslint .",
|
||||
"preview": "vite preview"
|
||||
},
|
||||
"dependencies": {
|
||||
"antd": "^6.1.0",
|
||||
"react": "^19.2.0",
|
||||
"react-dom": "^19.2.0",
|
||||
"react-router": "^7.10.1"
|
||||
},
|
||||
"devDependencies": {
|
||||
"@eslint/js": "^9.39.1",
|
||||
"@types/node": "^24.10.1",
|
||||
"@types/react": "^19.2.5",
|
||||
"@types/react-dom": "^19.2.3",
|
||||
"@vitejs/plugin-react": "^5.1.1",
|
||||
"eslint": "^9.39.1",
|
||||
"eslint-plugin-react-hooks": "^7.0.1",
|
||||
"eslint-plugin-react-refresh": "^0.4.24",
|
||||
"globals": "^16.5.0",
|
||||
"typescript": "~5.9.3",
|
||||
"typescript-eslint": "^8.46.4",
|
||||
"vite": "^7.2.4"
|
||||
}
|
||||
}
|
||||
1
frontend/public/vite.svg
Normal file
1
frontend/public/vite.svg
Normal file
|
|
@ -0,0 +1 @@
|
|||
<svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" role="img" class="iconify iconify--logos" width="31.88" height="32" preserveAspectRatio="xMidYMid meet" viewBox="0 0 256 257"><defs><linearGradient id="IconifyId1813088fe1fbc01fb466" x1="-.828%" x2="57.636%" y1="7.652%" y2="78.411%"><stop offset="0%" stop-color="#41D1FF"></stop><stop offset="100%" stop-color="#BD34FE"></stop></linearGradient><linearGradient id="IconifyId1813088fe1fbc01fb467" x1="43.376%" x2="50.316%" y1="2.242%" y2="89.03%"><stop offset="0%" stop-color="#FFEA83"></stop><stop offset="8.333%" stop-color="#FFDD35"></stop><stop offset="100%" stop-color="#FFA800"></stop></linearGradient></defs><path fill="url(#IconifyId1813088fe1fbc01fb466)" d="M255.153 37.938L134.897 252.976c-2.483 4.44-8.862 4.466-11.382.048L.875 37.958c-2.746-4.814 1.371-10.646 6.827-9.67l120.385 21.517a6.537 6.537 0 0 0 2.322-.004l117.867-21.483c5.438-.991 9.574 4.796 6.877 9.62Z"></path><path fill="url(#IconifyId1813088fe1fbc01fb467)" d="M185.432.063L96.44 17.501a3.268 3.268 0 0 0-2.634 3.014l-5.474 92.456a3.268 3.268 0 0 0 3.997 3.378l24.777-5.718c2.318-.535 4.413 1.507 3.936 3.838l-7.361 36.047c-.495 2.426 1.782 4.5 4.151 3.78l15.304-4.649c2.372-.72 4.652 1.36 4.15 3.788l-11.698 56.621c-.732 3.542 3.979 5.473 5.943 2.437l1.313-2.028l72.516-144.72c1.215-2.423-.88-5.186-3.54-4.672l-25.505 4.922c-2.396.462-4.435-1.77-3.759-4.114l16.646-57.705c.677-2.35-1.37-4.583-3.769-4.113Z"></path></svg>
|
||||
|
After Width: | Height: | Size: 1.5 KiB |
67
frontend/src/ApiKeyButton.tsx
Normal file
67
frontend/src/ApiKeyButton.tsx
Normal file
|
|
@ -0,0 +1,67 @@
|
|||
import {type Dispatch, type SetStateAction, useEffect} from 'react';
|
||||
import {Button, message} from "antd";
|
||||
|
||||
interface ApiKeyProps {
|
||||
apiKey?: string;
|
||||
setApiKey: Dispatch<SetStateAction<string | undefined>>;
|
||||
}
|
||||
|
||||
const ApiKeyButton: React.FC<ApiKeyProps> = ({apiKey, setApiKey}) => {
|
||||
// Check if API key already exists in localStorage on component mount
|
||||
useEffect(() => {
|
||||
const savedApiKey = localStorage.getItem('apiKey');
|
||||
if (savedApiKey) {
|
||||
setApiKey(savedApiKey);
|
||||
}
|
||||
}, [setApiKey]);
|
||||
|
||||
const handleSetApiKey = async () => {
|
||||
const key = prompt('Please enter your API key:');
|
||||
|
||||
if (key) {
|
||||
const response = await fetch(
|
||||
"/api/v1/tofu/instances",
|
||||
{
|
||||
headers: {
|
||||
"Authorization": `Bearer ${key}`,
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
}
|
||||
);
|
||||
if (response.ok) {
|
||||
// Store the API key in localStorage
|
||||
localStorage.setItem('apiKey', key);
|
||||
setApiKey(key);
|
||||
message.info("API key added successfully");
|
||||
} else {
|
||||
message.error("Invalid API key");
|
||||
}
|
||||
} else if (key !== undefined) { // User clicked OK but left empty
|
||||
message.error('API key cannot be empty.');
|
||||
}
|
||||
// If user clicked Cancel, do nothing
|
||||
};
|
||||
|
||||
const handleRemoveApiKey = () => {
|
||||
localStorage.removeItem('apiKey');
|
||||
setApiKey(undefined);
|
||||
};
|
||||
|
||||
return (
|
||||
<>
|
||||
{apiKey !== undefined ? (
|
||||
<div>
|
||||
<Button type="default" danger onClick={handleRemoveApiKey}>
|
||||
Remove API Key
|
||||
</Button>
|
||||
</div>
|
||||
) : (
|
||||
<Button type='primary' onClick={handleSetApiKey}>
|
||||
Set API Key
|
||||
</Button>
|
||||
)}
|
||||
</>
|
||||
);
|
||||
};
|
||||
|
||||
export default ApiKeyButton;
|
||||
12
frontend/src/App.css
Normal file
12
frontend/src/App.css
Normal file
|
|
@ -0,0 +1,12 @@
|
|||
body {
|
||||
margin: 0;
|
||||
}
|
||||
|
||||
.demo-logo {
|
||||
width: 120px;
|
||||
min-width: 120px;
|
||||
height: 32px;
|
||||
background: rgba(255, 255, 255, .2);
|
||||
border-radius: 6px;
|
||||
margin-inline-end: 24px;
|
||||
}
|
||||
84
frontend/src/App.tsx
Normal file
84
frontend/src/App.tsx
Normal file
|
|
@ -0,0 +1,84 @@
|
|||
import React, {useState} from 'react';
|
||||
import {CloudServerOutlined, UnorderedListOutlined} from '@ant-design/icons';
|
||||
import {Breadcrumb, Layout, Menu, type MenuProps, theme, Typography} from 'antd';
|
||||
|
||||
import "./App.css";
|
||||
import ApiKeyButton from "./ApiKeyButton.tsx";
|
||||
import TofuInstanceList from "./TofuInstanceList.tsx";
|
||||
import {Navigate, Route, Routes, useNavigate} from "react-router";
|
||||
import TofuInstanceDetail from "./TofuInstanceDetail.tsx";
|
||||
|
||||
const {Header, Content, Footer, Sider} = Layout;
|
||||
const {Title, Paragraph} = Typography;
|
||||
|
||||
const App: React.FC = () => {
|
||||
const {
|
||||
token: {colorBgContainer, borderRadiusLG},
|
||||
} = theme.useToken();
|
||||
|
||||
const [apiKey, setApiKey] = useState<string | undefined>(undefined);
|
||||
const [breadcrumb, setBreadcrumb] = useState();
|
||||
const navigate = useNavigate();
|
||||
|
||||
const sideMenu: MenuProps['items'] = [{
|
||||
key: 'tofu',
|
||||
label: 'OpenTofu',
|
||||
icon: React.createElement(CloudServerOutlined),
|
||||
children: [
|
||||
{
|
||||
key: 'instances',
|
||||
label: 'Instances',
|
||||
icon: React.createElement(UnorderedListOutlined),
|
||||
onClick: () => {
|
||||
navigate("/tofu/instances/")
|
||||
}
|
||||
}
|
||||
]
|
||||
}]
|
||||
|
||||
return (
|
||||
<Layout>
|
||||
<Header style={{display: 'flex', alignItems: 'center'}}>
|
||||
<div className="demo-logo"/>
|
||||
<ApiKeyButton apiKey={apiKey} setApiKey={setApiKey}/>
|
||||
</Header>
|
||||
<div style={{padding: '0 48px'}}>
|
||||
{apiKey && <Breadcrumb
|
||||
style={{margin: '16px 0'}}
|
||||
items={breadcrumb}
|
||||
/>}
|
||||
<Layout
|
||||
style={{padding: '24px 0', background: colorBgContainer, borderRadius: borderRadiusLG}}
|
||||
>
|
||||
{apiKey && <Sider style={{background: colorBgContainer}} width={200}>
|
||||
<Menu
|
||||
mode="inline"
|
||||
defaultSelectedKeys={['instances']}
|
||||
defaultOpenKeys={['tofu']}
|
||||
style={{height: '100%'}}
|
||||
items={sideMenu}
|
||||
/>
|
||||
</Sider>}
|
||||
<Content style={{padding: '0 24px', minHeight: 280}}>
|
||||
{(apiKey && <Routes>
|
||||
<Route index element={<Navigate to="/tofu/instances"/>}/>
|
||||
<Route path="/tofu" element={<Navigate to="/tofu/instances"/>}/>
|
||||
<Route path="/tofu/instances"
|
||||
element={<TofuInstanceList apiKey={apiKey} setBreadcrumb={setBreadcrumb}/>}/>
|
||||
<Route path="/tofu/instances/:instanceId"
|
||||
element={<TofuInstanceDetail apiKey={apiKey} setBreadcrumb={setBreadcrumb}/>}/>
|
||||
</Routes>) || <>
|
||||
<Title>API Key Required</Title>
|
||||
<Paragraph>Use the button above to enter your API key.</Paragraph>
|
||||
</>}
|
||||
</Content>
|
||||
</Layout>
|
||||
</div>
|
||||
<Footer style={{textAlign: 'center'}}>
|
||||
Copyright © 2021-{new Date().getFullYear()} SR2 Communications Limited.
|
||||
</Footer>
|
||||
</Layout>
|
||||
);
|
||||
};
|
||||
|
||||
export default App;
|
||||
215
frontend/src/TofuInstanceDetail.tsx
Normal file
215
frontend/src/TofuInstanceDetail.tsx
Normal file
|
|
@ -0,0 +1,215 @@
|
|||
import {Alert, Button, Descriptions, Input, message, Modal, Space, Table, Tooltip, Typography} from "antd";
|
||||
import {type Dispatch, type SetStateAction, useEffect, useState} from "react";
|
||||
import {deleteInstance, statusTag, updateInstance} from "./client.tsx";
|
||||
import {useNavigate, useParams} from "react-router";
|
||||
import {DeleteOutlined, EyeOutlined, ReloadOutlined, ToolOutlined} from "@ant-design/icons";
|
||||
|
||||
const {Title, Paragraph, Text} = Typography;
|
||||
const {TextArea} = Input;
|
||||
|
||||
interface TofuInstanceDetailProps {
|
||||
apiKey: string
|
||||
setBreadcrumb: Dispatch<SetStateAction<object[]>>;
|
||||
}
|
||||
|
||||
const showLogs = async (apiKey: string, taskId: number, setLogs: Dispatch<SetStateAction<string>>) => {
|
||||
const response = await fetch("/api/v1/tofu/logs/" + taskId, {
|
||||
headers: {
|
||||
"Authorization": `Bearer ${apiKey}`, "Content-Type": "application/json"
|
||||
},
|
||||
});
|
||||
const reader = response.body?.getReader();
|
||||
if (!reader) {
|
||||
message.error("Could not open logs stream");
|
||||
return;
|
||||
}
|
||||
const decoder = new TextDecoder()
|
||||
let buffer = "";
|
||||
let messages = "";
|
||||
while (true) {
|
||||
const {value, done} = await reader.read();
|
||||
buffer += decoder.decode(value, {stream: true});
|
||||
let boundary;
|
||||
while ((boundary = buffer.indexOf('\n')) !== -1) {
|
||||
const line = buffer.slice(0, boundary).trim();
|
||||
buffer = buffer.slice(boundary + 1);
|
||||
|
||||
if (line) {
|
||||
try {
|
||||
const json = JSON.parse(line);
|
||||
if (json["@message"] === undefined) {
|
||||
messages += JSON.stringify(json) + "\n";
|
||||
} else {
|
||||
messages += json["@message"] + "\n";
|
||||
}
|
||||
} catch (e) {
|
||||
console.error('Failed to parse JSON line:', e, 'Line:', line);
|
||||
}
|
||||
}
|
||||
}
|
||||
setLogs(messages);
|
||||
if (done) break;
|
||||
}
|
||||
}
|
||||
|
||||
const TofuInstanceDetail: React.FC<TofuInstanceDetailProps> = ({apiKey, setBreadcrumb}) => {
|
||||
const [instance, setInstance] = useState(undefined);
|
||||
const [isLoading, setIsLoading] = useState(true);
|
||||
const [isConfigurationVisible, setIsConfigurationVisible] = useState(false);
|
||||
const [configuration, setConfiguration] = useState("");
|
||||
const [reloadCount, setReloadCount] = useState(0);
|
||||
const [errorMessage, setErrorMessage] = useState<string | undefined>(undefined);
|
||||
const [logs, setLogs] = useState("");
|
||||
const navigate = useNavigate();
|
||||
const {instanceId} = useParams();
|
||||
|
||||
useEffect(() => {
|
||||
const fetchInstance = async () => {
|
||||
try {
|
||||
setIsLoading(true);
|
||||
const response = await fetch("/api/v1/tofu/instances/" + instanceId, {
|
||||
headers: {
|
||||
"Authorization": `Bearer ${apiKey}`, "Content-Type": "application/json"
|
||||
}
|
||||
});
|
||||
|
||||
if (response.ok) {
|
||||
const data = await response.json();
|
||||
data.tasks = data.tasks.map((task, index: number) => ({
|
||||
...task, key: index + 1
|
||||
}));
|
||||
setInstance(data);
|
||||
if (data.outputs) {
|
||||
data.outputs = Object.entries<object>(data.outputs).map(([key, value]) => ({
|
||||
...value,
|
||||
output: key,
|
||||
key: key,
|
||||
}));
|
||||
}
|
||||
setConfiguration(JSON.stringify(data.configuration));
|
||||
setErrorMessage(undefined);
|
||||
} else {
|
||||
setErrorMessage("Failed to fetch instances:" + response.status);
|
||||
setInstance(undefined);
|
||||
}
|
||||
} catch (error) {
|
||||
setErrorMessage("Error fetching instances:" + error);
|
||||
setInstance(undefined);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
}
|
||||
};
|
||||
|
||||
setBreadcrumb([
|
||||
{"title": "OpenTofu", href: "/tofu"},
|
||||
{"title": "Instances", href: "/tofu/instances"},
|
||||
{"title": "Instance " + instanceId},
|
||||
])
|
||||
|
||||
fetchInstance();
|
||||
}, [apiKey, instanceId, reloadCount, setBreadcrumb, setConfiguration, setInstance, setIsLoading]);
|
||||
|
||||
const taskColumns = [{"title": "ID", "dataIndex": "id", "key": "id"}, {
|
||||
"title": "Status", "dataIndex": "status", "key": "state", render: statusTag
|
||||
}, {
|
||||
"title": "Created",
|
||||
dataIndex: "created_at",
|
||||
key: "created_at",
|
||||
sorter: (a, b) => a.created_at.localeCompare(b.created_at)
|
||||
}, {
|
||||
"title": "Started",
|
||||
dataIndex: "start_time",
|
||||
key: "start_time",
|
||||
render: (time) => (time || "Not yet"),
|
||||
sorter: (a, b) => a.start_time.localeCompare(b.start_time)
|
||||
}, {
|
||||
"title": "Ended",
|
||||
dataIndex: "end_time",
|
||||
key: "end_time",
|
||||
render: (time) => (time || "Not yet"),
|
||||
sorter: (a, b) => a.end_time.localeCompare(b.end_time)
|
||||
}, {
|
||||
title: 'Action', key: 'action', render: (_, record) => (<Space size="middle">
|
||||
<Button type="default" onClick={() => {
|
||||
showLogs(apiKey, record.id, setLogs)
|
||||
}}><EyeOutlined/> View Log</Button>
|
||||
</Space>),
|
||||
},]
|
||||
|
||||
const outputColumns = [
|
||||
{"title": "Output", "dataIndex": "output", "key": "output"},
|
||||
{"title": "Value", "dataIndex": "value", "key": "value", render: (_, record) => {
|
||||
if (record.sensitive) {
|
||||
return (<Tooltip placement="top" title={record.value}>
|
||||
<Text type="secondary">Sensitive Value</Text>
|
||||
</Tooltip>);
|
||||
} else {
|
||||
return record.value;
|
||||
}
|
||||
}},
|
||||
{"title": "Type", "dataIndex": "type", "key": "type"}
|
||||
]
|
||||
|
||||
return (<>
|
||||
<Title>Instance {instanceId}</Title>
|
||||
<Paragraph>
|
||||
{errorMessage && <Alert title={errorMessage} type="error"/>}
|
||||
</Paragraph>
|
||||
<Paragraph>
|
||||
<Space size="middle">
|
||||
<Button type="default" onClick={() => {
|
||||
setIsConfigurationVisible(true)
|
||||
}}><ToolOutlined/> Configuration</Button>
|
||||
<Button onClick={() => setReloadCount(reloadCount + 1)}
|
||||
loading={isLoading}><ReloadOutlined/> Reload</Button>
|
||||
<Button type="default" danger={true} onClick={() => {
|
||||
deleteInstance(apiKey, instanceId).then(() => {
|
||||
navigate("/tofu/instances")
|
||||
});
|
||||
}}><DeleteOutlined/> Delete</Button>
|
||||
</Space>
|
||||
</Paragraph>
|
||||
{instance && <>
|
||||
<Descriptions title="Properties">
|
||||
<Descriptions.Item label="ID">{instance.id}</Descriptions.Item>
|
||||
<Descriptions.Item label="Status">{statusTag(instance.status)}</Descriptions.Item>
|
||||
<Descriptions.Item label="Created">{instance.created_at}</Descriptions.Item>
|
||||
<Descriptions.Item label="Updated">{instance.updated_at}</Descriptions.Item>
|
||||
<Descriptions.Item label="Drift Checked">{instance.drift_checked_at || "Never"}</Descriptions.Item>
|
||||
<Descriptions.Item label="State Lock">
|
||||
{(instance.state_lock && "Since " + instance.state_lock.Created) || "Unlocked"}
|
||||
</Descriptions.Item>
|
||||
</Descriptions>
|
||||
<Title level={2}>Instance Tasks</Title>
|
||||
<Table columns={taskColumns} dataSource={instance.tasks} loading={isLoading}/>
|
||||
{instance.outputs && <>
|
||||
<Title level={2}>Instance Outputs</Title>
|
||||
<Table columns={outputColumns} dataSource={instance.outputs} loading={isLoading}/>
|
||||
</>}
|
||||
<Modal
|
||||
title="Configuration"
|
||||
open={isConfigurationVisible}
|
||||
onCancel={() => setIsConfigurationVisible(false)}
|
||||
footer={[<Button key="back" danger onClick={() => {
|
||||
updateInstance(apiKey, instanceId, configuration);
|
||||
setIsConfigurationVisible(false);
|
||||
}}>Update</Button>,
|
||||
<Button key="back" onClick={() => setIsConfigurationVisible(false)}>Cancel</Button>]}
|
||||
width="90%"
|
||||
>
|
||||
<TextArea rows={20} value={configuration} onChange={(e) => setConfiguration(e.target.value)}/>
|
||||
</Modal>
|
||||
<Modal
|
||||
title="Logs"
|
||||
footer={[<Button key="back" onClick={() => setLogs("")}>Close</Button>]}
|
||||
onCancel={() => setLogs("")}
|
||||
width="90%"
|
||||
open={logs !== ""}
|
||||
>
|
||||
<pre style={{"overflow": "auto"}}>{logs}</pre>
|
||||
</Modal>
|
||||
</>}
|
||||
</>)
|
||||
};
|
||||
|
||||
export default TofuInstanceDetail;
|
||||
119
frontend/src/TofuInstanceList.tsx
Normal file
119
frontend/src/TofuInstanceList.tsx
Normal file
|
|
@ -0,0 +1,119 @@
|
|||
import {Alert, Button, Input, Modal, Space, Table, Typography} from "antd";
|
||||
import {type Dispatch, type SetStateAction, useEffect, useState} from "react";
|
||||
import {createInstance, deleteInstance, statusTag} from "./client.tsx";
|
||||
import {useNavigate} from "react-router";
|
||||
import {PlusOutlined, ReloadOutlined} from "@ant-design/icons";
|
||||
|
||||
const {Title, Paragraph} = Typography;
|
||||
const {TextArea} = Input;
|
||||
|
||||
const defaultNewConfiguration = "{\"resource\":{\"random_password\":{\"example\":{\"length\":20}}},\"output\":{\"example\":{\"value\":\"${random_password.example.result}\",\"sensitive\":true}}}"
|
||||
|
||||
interface TofuInstanceListProps {
|
||||
apiKey?: string
|
||||
setBreadcrumb: Dispatch<SetStateAction<object[]>>
|
||||
}
|
||||
|
||||
const TofuInstanceList: React.FC<TofuInstanceListProps> = ({apiKey, setBreadcrumb}) => {
|
||||
const [instances, setInstances] = useState([]);
|
||||
const [isLoading, setIsLoading] = useState(true);
|
||||
const [isNewModalVisible, setIsNewModalVisible] = useState(false);
|
||||
const [configuration, setConfiguration] = useState<string>(defaultNewConfiguration);
|
||||
const [reloadCount, setReloadCount] = useState(0);
|
||||
const [errorMessage, setErrorMessage] = useState<string | undefined>(undefined);
|
||||
const navigate = useNavigate();
|
||||
|
||||
|
||||
useEffect(() => {
|
||||
const fetchInstances = async () => {
|
||||
try {
|
||||
setIsLoading(true);
|
||||
const response = await fetch("/api/v1/tofu/instances", {
|
||||
headers: {
|
||||
"Authorization": `Bearer ${apiKey}`, "Content-Type": "application/json"
|
||||
}
|
||||
});
|
||||
|
||||
if (response.ok) {
|
||||
const data = await response.json();
|
||||
setInstances((data.items || []).map((item: object, index: number) => ({
|
||||
...item, key: index + 1
|
||||
})));
|
||||
setErrorMessage(undefined);
|
||||
} else {
|
||||
setErrorMessage("Failed to fetch deployments:" + response.status);
|
||||
setInstances([]);
|
||||
}
|
||||
} catch (error) {
|
||||
setErrorMessage("Error fetching deployments:" + error);
|
||||
setInstances([]);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
}
|
||||
};
|
||||
|
||||
setBreadcrumb([{"title": "OpenTofu", href: "/tofu"}, {
|
||||
"title": "Instances",
|
||||
href: "/tofu/instances"
|
||||
}, {"title": "List"},])
|
||||
|
||||
if (apiKey) fetchInstances();
|
||||
}, [apiKey, reloadCount, setBreadcrumb, setInstances, setIsLoading]); // Reload when apiKey or reloadCount changes
|
||||
|
||||
if (!apiKey) return <></>;
|
||||
|
||||
const columns = [{
|
||||
"title": "ID",
|
||||
"dataIndex": "id",
|
||||
"key": "id",
|
||||
sorter: (a, b) => a.id < b.id,
|
||||
defaultSortOrder: 'descend'
|
||||
}, {
|
||||
"title": "Status", "dataIndex": "status", "key": "state", render: statusTag
|
||||
}, {
|
||||
"title": "Last Status Change",
|
||||
key: "status_changed_at",
|
||||
dataIndex: "status_changed_at"
|
||||
}, {"title": "Last Drift Check", key: "drift_checked_at", dataIndex: "drift_checked_at"}, {
|
||||
title: 'Action', key: 'action', render: (_, record: { id: string | number; }) => (<Space size="middle">
|
||||
<Button type="primary" onClick={() => {
|
||||
navigate("/tofu/instances/" + record.id)
|
||||
}}>Details</Button>
|
||||
<Button type="link" danger={true} onClick={() => {
|
||||
deleteInstance(apiKey, record.id).then(() => {
|
||||
setReloadCount(reloadCount + 1)
|
||||
});
|
||||
}}>Delete</Button>
|
||||
</Space>),
|
||||
},]
|
||||
|
||||
return (<>
|
||||
<Title>Instances</Title>
|
||||
<Paragraph>
|
||||
{errorMessage && <Alert title={errorMessage} type="error"/>}
|
||||
</Paragraph>
|
||||
<Paragraph>
|
||||
<Space size="middle">
|
||||
<Button type="dashed" onClick={() => setIsNewModalVisible(true)}><PlusOutlined /> New</Button>
|
||||
<Button onClick={() => setReloadCount(reloadCount + 1)} loading={isLoading}><ReloadOutlined /> Reload</Button>
|
||||
</Space>
|
||||
</Paragraph>
|
||||
<Table columns={columns} dataSource={instances} loading={isLoading}/>
|
||||
<Modal
|
||||
title="New Instance Configuration"
|
||||
open={isNewModalVisible}
|
||||
onCancel={() => setIsNewModalVisible(false)}
|
||||
footer={[<Button key="back" danger onClick={async () => {
|
||||
const instanceId = await createInstance(apiKey, configuration);
|
||||
setIsNewModalVisible(false);
|
||||
navigate("/tofu/instances/" + instanceId)
|
||||
}}>Update</Button>,
|
||||
<Button key="back" onClick={() => setIsNewModalVisible(false)}>Cancel</Button>]}
|
||||
width="90%"
|
||||
>
|
||||
<TextArea rows={20} value={configuration} onChange={(e) => setConfiguration(e.target.value)}/>
|
||||
</Modal>
|
||||
</>)
|
||||
};
|
||||
|
||||
export default TofuInstanceList;
|
||||
5
frontend/src/assets/react.svg
Normal file
5
frontend/src/assets/react.svg
Normal file
|
|
@ -0,0 +1,5 @@
|
|||
<svg xmlns="http://www.w3.org/2000/svg" aria-hidden="true" role="img"
|
||||
class="iconify iconify--logos" width="35.93" height="32" preserveAspectRatio="xMidYMid meet" viewBox="0 0 256 228">
|
||||
<path fill="#00D8FF"
|
||||
d="M210.483 73.824a171.49 171.49 0 0 0-8.24-2.597c.465-1.9.893-3.777 1.273-5.621c6.238-30.281 2.16-54.676-11.769-62.708c-13.355-7.7-35.196.329-57.254 19.526a171.23 171.23 0 0 0-6.375 5.848a155.866 155.866 0 0 0-4.241-3.917C100.759 3.829 77.587-4.822 63.673 3.233C50.33 10.957 46.379 33.89 51.995 62.588a170.974 170.974 0 0 0 1.892 8.48c-3.28.932-6.445 1.924-9.474 2.98C17.309 83.498 0 98.307 0 113.668c0 15.865 18.582 31.778 46.812 41.427a145.52 145.52 0 0 0 6.921 2.165a167.467 167.467 0 0 0-2.01 9.138c-5.354 28.2-1.173 50.591 12.134 58.266c13.744 7.926 36.812-.22 59.273-19.855a145.567 145.567 0 0 0 5.342-4.923a168.064 168.064 0 0 0 6.92 6.314c21.758 18.722 43.246 26.282 56.54 18.586c13.731-7.949 18.194-32.003 12.4-61.268a145.016 145.016 0 0 0-1.535-6.842c1.62-.48 3.21-.974 4.76-1.488c29.348-9.723 48.443-25.443 48.443-41.52c0-15.417-17.868-30.326-45.517-39.844Zm-6.365 70.984c-1.4.463-2.836.91-4.3 1.345c-3.24-10.257-7.612-21.163-12.963-32.432c5.106-11 9.31-21.767 12.459-31.957c2.619.758 5.16 1.557 7.61 2.4c23.69 8.156 38.14 20.213 38.14 29.504c0 9.896-15.606 22.743-40.946 31.14Zm-10.514 20.834c2.562 12.94 2.927 24.64 1.23 33.787c-1.524 8.219-4.59 13.698-8.382 15.893c-8.067 4.67-25.32-1.4-43.927-17.412a156.726 156.726 0 0 1-6.437-5.87c7.214-7.889 14.423-17.06 21.459-27.246c12.376-1.098 24.068-2.894 34.671-5.345a134.17 134.17 0 0 1 1.386 6.193ZM87.276 214.515c-7.882 2.783-14.16 2.863-17.955.675c-8.075-4.657-11.432-22.636-6.853-46.752a156.923 156.923 0 0 1 1.869-8.499c10.486 2.32 22.093 3.988 34.498 4.994c7.084 9.967 14.501 19.128 21.976 27.15a134.668 134.668 0 0 1-4.877 4.492c-9.933 8.682-19.886 14.842-28.658 17.94ZM50.35 144.747c-12.483-4.267-22.792-9.812-29.858-15.863c-6.35-5.437-9.555-10.836-9.555-15.216c0-9.322 13.897-21.212 37.076-29.293c2.813-.98 5.757-1.905 8.812-2.773c3.204 10.42 7.406 21.315 12.477 32.332c-5.137 11.18-9.399 22.249-12.634 32.792a134.718 134.718 0 0 1-6.318-1.979Zm12.378-84.26c-4.811-24.587-1.616-43.134 6.425-47.789c8.564-4.958 27.502 2.111 47.463 19.835a144.318 144.318 0 0 1 3.841 3.545c-7.438 7.987-14.787 17.08-21.808 26.988c-12.04 1.116-23.565 2.908-34.161 5.309a160.342 160.342 0 0 1-1.76-7.887Zm110.427 27.268a347.8 347.8 0 0 0-7.785-12.803c8.168 1.033 15.994 2.404 23.343 4.08c-2.206 7.072-4.956 14.465-8.193 22.045a381.151 381.151 0 0 0-7.365-13.322Zm-45.032-43.861c5.044 5.465 10.096 11.566 15.065 18.186a322.04 322.04 0 0 0-30.257-.006c4.974-6.559 10.069-12.652 15.192-18.18ZM82.802 87.83a323.167 323.167 0 0 0-7.227 13.238c-3.184-7.553-5.909-14.98-8.134-22.152c7.304-1.634 15.093-2.97 23.209-3.984a321.524 321.524 0 0 0-7.848 12.897Zm8.081 65.352c-8.385-.936-16.291-2.203-23.593-3.793c2.26-7.3 5.045-14.885 8.298-22.6a321.187 321.187 0 0 0 7.257 13.246c2.594 4.48 5.28 8.868 8.038 13.147Zm37.542 31.03c-5.184-5.592-10.354-11.779-15.403-18.433c4.902.192 9.899.29 14.978.29c5.218 0 10.376-.117 15.453-.343c-4.985 6.774-10.018 12.97-15.028 18.486Zm52.198-57.817c3.422 7.8 6.306 15.345 8.596 22.52c-7.422 1.694-15.436 3.058-23.88 4.071a382.417 382.417 0 0 0 7.859-13.026a347.403 347.403 0 0 0 7.425-13.565Zm-16.898 8.101a358.557 358.557 0 0 1-12.281 19.815a329.4 329.4 0 0 1-23.444.823c-7.967 0-15.716-.248-23.178-.732a310.202 310.202 0 0 1-12.513-19.846h.001a307.41 307.41 0 0 1-10.923-20.627a310.278 310.278 0 0 1 10.89-20.637l-.001.001a307.318 307.318 0 0 1 12.413-19.761c7.613-.576 15.42-.876 23.31-.876H128c7.926 0 15.743.303 23.354.883a329.357 329.357 0 0 1 12.335 19.695a358.489 358.489 0 0 1 11.036 20.54a329.472 329.472 0 0 1-11 20.722Zm22.56-122.124c8.572 4.944 11.906 24.881 6.52 51.026c-.344 1.668-.73 3.367-1.15 5.09c-10.622-2.452-22.155-4.275-34.23-5.408c-7.034-10.017-14.323-19.124-21.64-27.008a160.789 160.789 0 0 1 5.888-5.4c18.9-16.447 36.564-22.941 44.612-18.3ZM128 90.808c12.625 0 22.86 10.235 22.86 22.86s-10.235 22.86-22.86 22.86s-22.86-10.235-22.86-22.86s10.235-22.86 22.86-22.86Z"></path>
|
||||
</svg>
|
||||
|
After Width: | Height: | Size: 4 KiB |
81
frontend/src/client.tsx
Normal file
81
frontend/src/client.tsx
Normal file
|
|
@ -0,0 +1,81 @@
|
|||
import {message, Tag} from "antd";
|
||||
|
||||
const statusTag = (status: string) => {
|
||||
switch (status) {
|
||||
case "ACTIVE":
|
||||
case "COMPLETED":
|
||||
return (<Tag color="green">{status}</Tag>)
|
||||
case "PENDING":
|
||||
case "RUNNING":
|
||||
return (<Tag color="blue">{status}</Tag>);
|
||||
case "PENDING_DESTROY":
|
||||
return (<Tag color="yellow">{status.replace("_", " ")}</Tag>);
|
||||
case "FAILED":
|
||||
case "FAILED_DESTROY":
|
||||
return (<Tag color="red">{status.replace("_", " ")}</Tag>);
|
||||
default:
|
||||
return status;
|
||||
}
|
||||
}
|
||||
|
||||
const createInstance = async (apiKey: string, configuration: string) => {
|
||||
const response = await fetch(
|
||||
"/api/v1/tofu/instances",
|
||||
{
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Authorization": `Bearer ${apiKey}`,
|
||||
"Content-Type": "application/json"
|
||||
},
|
||||
body: JSON.stringify({"configuration": JSON.parse(configuration)})
|
||||
}
|
||||
);
|
||||
const status = response.status;
|
||||
const json = await response.json();
|
||||
if (status === 202) {
|
||||
message.info("OpenTofu instance " + json.id + " creation in progress!");
|
||||
return json.id;
|
||||
} else {
|
||||
message.error("Could not update OpenTofu instance " + json.id + ": " + status);
|
||||
}
|
||||
}
|
||||
|
||||
const updateInstance = async (apiKey: string, instanceId: number, configuration: string) => {
|
||||
const response = await fetch(
|
||||
"/api/v1/tofu/instances/" + instanceId,
|
||||
{
|
||||
method: "PUT",
|
||||
headers: {
|
||||
"Authorization": `Bearer ${apiKey}`,
|
||||
"Content-Type": "application/json"
|
||||
},
|
||||
body: JSON.stringify({"configuration": JSON.parse(configuration)})
|
||||
}
|
||||
);
|
||||
const status = response.status;
|
||||
if (status === 202) {
|
||||
message.info("OpenTofu instance " + instanceId + " update in progress!");
|
||||
} else {
|
||||
message.error("Could not update OpenTofu instance " + instanceId + ": " + status);
|
||||
}
|
||||
}
|
||||
|
||||
const deleteInstance = async (apiKey: string, instanceId: number) => {
|
||||
const response = await fetch(
|
||||
"/api/v1/tofu/instances/" + instanceId,
|
||||
{
|
||||
method: "DELETE",
|
||||
headers: {
|
||||
"Authorization": `Bearer ${apiKey}`,
|
||||
}
|
||||
}
|
||||
);
|
||||
const status = response.status;
|
||||
if (status === 202) {
|
||||
message.info("OpenTofu instance " + instanceId + " deletion in progress!");
|
||||
} else {
|
||||
message.error("Could not delete OpenTofu instance " + instanceId + ": " + status);
|
||||
}
|
||||
}
|
||||
|
||||
export {createInstance, deleteInstance, statusTag, updateInstance};
|
||||
12
frontend/src/main.tsx
Normal file
12
frontend/src/main.tsx
Normal file
|
|
@ -0,0 +1,12 @@
|
|||
import {StrictMode} from 'react'
|
||||
import {createRoot} from 'react-dom/client'
|
||||
import App from './App.tsx'
|
||||
import {BrowserRouter} from "react-router";
|
||||
|
||||
createRoot(document.getElementById('root')!).render(
|
||||
<StrictMode>
|
||||
<BrowserRouter>
|
||||
<App/>
|
||||
</BrowserRouter>
|
||||
</StrictMode>,
|
||||
)
|
||||
28
frontend/tsconfig.app.json
Normal file
28
frontend/tsconfig.app.json
Normal file
|
|
@ -0,0 +1,28 @@
|
|||
{
|
||||
"compilerOptions": {
|
||||
"tsBuildInfoFile": "./node_modules/.tmp/tsconfig.app.tsbuildinfo",
|
||||
"target": "ES2022",
|
||||
"useDefineForClassFields": true,
|
||||
"lib": ["ES2022", "DOM", "DOM.Iterable"],
|
||||
"module": "ESNext",
|
||||
"types": ["vite/client"],
|
||||
"skipLibCheck": true,
|
||||
|
||||
/* Bundler mode */
|
||||
"moduleResolution": "bundler",
|
||||
"allowImportingTsExtensions": true,
|
||||
"verbatimModuleSyntax": true,
|
||||
"moduleDetection": "force",
|
||||
"noEmit": true,
|
||||
"jsx": "react-jsx",
|
||||
|
||||
/* Linting */
|
||||
"strict": true,
|
||||
"noUnusedLocals": true,
|
||||
"noUnusedParameters": true,
|
||||
"erasableSyntaxOnly": true,
|
||||
"noFallthroughCasesInSwitch": true,
|
||||
"noUncheckedSideEffectImports": true
|
||||
},
|
||||
"include": ["src"]
|
||||
}
|
||||
7
frontend/tsconfig.json
Normal file
7
frontend/tsconfig.json
Normal file
|
|
@ -0,0 +1,7 @@
|
|||
{
|
||||
"files": [],
|
||||
"references": [
|
||||
{ "path": "./tsconfig.app.json" },
|
||||
{ "path": "./tsconfig.node.json" }
|
||||
]
|
||||
}
|
||||
26
frontend/tsconfig.node.json
Normal file
26
frontend/tsconfig.node.json
Normal file
|
|
@ -0,0 +1,26 @@
|
|||
{
|
||||
"compilerOptions": {
|
||||
"tsBuildInfoFile": "./node_modules/.tmp/tsconfig.node.tsbuildinfo",
|
||||
"target": "ES2023",
|
||||
"lib": ["ES2023"],
|
||||
"module": "ESNext",
|
||||
"types": ["node"],
|
||||
"skipLibCheck": true,
|
||||
|
||||
/* Bundler mode */
|
||||
"moduleResolution": "bundler",
|
||||
"allowImportingTsExtensions": true,
|
||||
"verbatimModuleSyntax": true,
|
||||
"moduleDetection": "force",
|
||||
"noEmit": true,
|
||||
|
||||
/* Linting */
|
||||
"strict": true,
|
||||
"noUnusedLocals": true,
|
||||
"noUnusedParameters": true,
|
||||
"erasableSyntaxOnly": true,
|
||||
"noFallthroughCasesInSwitch": true,
|
||||
"noUncheckedSideEffectImports": true
|
||||
},
|
||||
"include": ["vite.config.ts"]
|
||||
}
|
||||
15
frontend/vite.config.ts
Normal file
15
frontend/vite.config.ts
Normal file
|
|
@ -0,0 +1,15 @@
|
|||
import { defineConfig } from 'vite'
|
||||
import react from '@vitejs/plugin-react'
|
||||
|
||||
// https://vite.dev/config/
|
||||
export default defineConfig({
|
||||
plugins: [react()],
|
||||
server: {
|
||||
proxy: {
|
||||
'/api/': {
|
||||
target: 'http://localhost:8000',
|
||||
changeOrigin: true,
|
||||
}
|
||||
}
|
||||
}
|
||||
})
|
||||
65
gunicorn/gunicorn_conf.py
Normal file
65
gunicorn/gunicorn_conf.py
Normal file
|
|
@ -0,0 +1,65 @@
|
|||
import multiprocessing
|
||||
|
||||
from pydantic import Field
|
||||
from pydantic_settings import BaseSettings, SettingsConfigDict
|
||||
|
||||
try:
|
||||
from prometheus_client import multiprocess
|
||||
|
||||
def child_exit(_, worker):
|
||||
multiprocess.mark_process_dead(worker.pid)
|
||||
|
||||
except ImportError:
|
||||
pass
|
||||
|
||||
|
||||
class Settings(BaseSettings):
|
||||
model_config = SettingsConfigDict(
|
||||
env_file=".env", env_file_encoding="utf-8", extra="ignore"
|
||||
)
|
||||
|
||||
host: str = "0.0.0.0"
|
||||
port: int = 8000
|
||||
bind: str | None = None
|
||||
|
||||
workers_per_core: int = Field(1)
|
||||
max_workers: int | None = None
|
||||
web_concurrency: int | None = None
|
||||
|
||||
graceful_timeout: int = 120
|
||||
timeout: int = 120
|
||||
keepalive: int = 5
|
||||
log_level: str = "INFO"
|
||||
log_config: str = "/src/logging_production.ini"
|
||||
|
||||
@property
|
||||
def computed_bind(self) -> str:
|
||||
return self.bind if self.bind else f"{self.host}:{self.port}"
|
||||
|
||||
@property
|
||||
def computed_web_concurrency(self) -> int:
|
||||
cores = multiprocessing.cpu_count()
|
||||
default_web_concurrency = self.workers_per_core * cores + 1
|
||||
|
||||
if self.web_concurrency:
|
||||
assert self.web_concurrency > 0
|
||||
return self.web_concurrency
|
||||
else:
|
||||
web_concurrency = max(default_web_concurrency, 2)
|
||||
if self.max_workers:
|
||||
return min(web_concurrency, self.max_workers)
|
||||
|
||||
return web_concurrency
|
||||
|
||||
|
||||
settings = Settings()
|
||||
|
||||
# Gunicorn config variables
|
||||
loglevel = settings.log_level
|
||||
workers = settings.computed_web_concurrency
|
||||
bind = settings.computed_bind
|
||||
worker_tmp_dir = "/dev/shm"
|
||||
graceful_timeout = settings.graceful_timeout
|
||||
timeout = settings.timeout
|
||||
keepalive = settings.keepalive
|
||||
logconfig = settings.log_config
|
||||
40
justfile
Normal file
40
justfile
Normal file
|
|
@ -0,0 +1,40 @@
|
|||
default:
|
||||
just --list
|
||||
|
||||
run *args:
|
||||
poetry run uvicorn src.main:app --reload {{args}}
|
||||
|
||||
mm *args:
|
||||
poetry run alembic revision --autogenerate -m "{{args}}"
|
||||
|
||||
migrate:
|
||||
poetry run alembic upgrade head
|
||||
|
||||
downgrade *args:
|
||||
poetry run alembic downgrade {{args}}
|
||||
|
||||
black *args:
|
||||
poetry run black {{args}} src
|
||||
|
||||
ruff *args:
|
||||
poetry run ruff check {{args}} src
|
||||
|
||||
lint:
|
||||
poetry run ruff format src
|
||||
just ruff --fix
|
||||
|
||||
test:
|
||||
PYTHONPATH=. pytest tests
|
||||
|
||||
# docker
|
||||
up:
|
||||
docker-compose up -d
|
||||
|
||||
kill *args:
|
||||
docker-compose kill {{args}}
|
||||
|
||||
build:
|
||||
docker-compose build
|
||||
|
||||
ps:
|
||||
docker-compose ps
|
||||
28
logging.ini
Normal file
28
logging.ini
Normal file
|
|
@ -0,0 +1,28 @@
|
|||
[loggers]
|
||||
keys=root,uvicorn
|
||||
|
||||
[handlers]
|
||||
keys=console
|
||||
|
||||
[formatters]
|
||||
keys=standard
|
||||
|
||||
[logger_root]
|
||||
level=INFO
|
||||
handlers=console
|
||||
formatter=standard
|
||||
|
||||
[logger_uvicorn]
|
||||
propagate=0
|
||||
handlers=console
|
||||
formatter=standard
|
||||
qualname=uvicorn
|
||||
|
||||
[handler_console]
|
||||
class=logging.StreamHandler
|
||||
level=DEBUG
|
||||
formatter=standard
|
||||
stream=ext://sys.stderr
|
||||
|
||||
[formatter_standard]
|
||||
format=[%(asctime)s] [%(levelname)s] [%(name)s] %(message)s
|
||||
44
logging_production.ini
Normal file
44
logging_production.ini
Normal file
|
|
@ -0,0 +1,44 @@
|
|||
[loggers]
|
||||
keys=root,gunicorn.access,gunicorn.error,uvicorn
|
||||
|
||||
[handlers]
|
||||
keys=console
|
||||
|
||||
[formatters]
|
||||
keys=json
|
||||
|
||||
[logger_root]
|
||||
level=INFO
|
||||
handlers=console
|
||||
formatter=json
|
||||
|
||||
[logger_gunicorn.access]
|
||||
level=INFO
|
||||
handlers=console
|
||||
formatter=json
|
||||
qualname=gunicorn.access
|
||||
propagate=1
|
||||
|
||||
[logger_gunicorn.error]
|
||||
level=INFO
|
||||
handlers=console
|
||||
formatter=json
|
||||
qualname=gunicorn.error
|
||||
propagate=0
|
||||
|
||||
[logger_uvicorn]
|
||||
propagate=0
|
||||
handlers=console
|
||||
formatter=json
|
||||
qualname=uvicorn
|
||||
|
||||
[handler_console]
|
||||
class=logging.StreamHandler
|
||||
level=INFO
|
||||
formatter=json
|
||||
stream=ext://sys.stderr
|
||||
|
||||
[formatter_json]
|
||||
class=pythonjsonlogger.jsonlogger.JsonFormatter
|
||||
format=[%(asctime)s] [%(levelname)s] [%(name)s] %(message)s
|
||||
datefmt=%Y-%m-%dT%H:%M:%S
|
||||
1782
poetry.lock
generated
Normal file
1782
poetry.lock
generated
Normal file
File diff suppressed because it is too large
Load diff
48
pyproject.toml
Normal file
48
pyproject.toml
Normal file
|
|
@ -0,0 +1,48 @@
|
|||
[tool.poetry]
|
||||
name = "pali-lili"
|
||||
version = "0.0.0"
|
||||
description = "Cloud orchestration microservice based on FastAPI."
|
||||
authors = ["irl"]
|
||||
readme = "README.md"
|
||||
license = "BSD-2"
|
||||
package-mode = false
|
||||
|
||||
[tool.poetry.dependencies]
|
||||
python = "^3.12"
|
||||
aiofiles = "^24.1.0"
|
||||
bcrypt = "^5.0.0"
|
||||
alembic = "^1.13.1"
|
||||
asyncpg = "^0.29.0"
|
||||
fastapi = "^0.115.12"
|
||||
fastapi-pagination = "^0.13.1"
|
||||
httpx = "^0.27.0"
|
||||
psycopg2-binary = "^2.9.9"
|
||||
pydantic = {extras = ["email"], version = "^2.11.4"}
|
||||
pydantic-settings = "^2.9.1"
|
||||
python-jose = {extras=["cryptogrpahy"], version = "^3.4.0"}
|
||||
sqlalchemy = {extras = ["asyncio"], version = "^2.0.30"}
|
||||
uvicorn = {extras = ["standard"], version = "^0.30.6"}
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
black = "^25.1.0"
|
||||
ruff = "^0.4.10"
|
||||
pytest = "^8.3.5"
|
||||
|
||||
[tool.poetry.group.prod.dependencies]
|
||||
gunicorn = "^22.0.0"
|
||||
python-json-logger = "^2.0.7"
|
||||
prometheus-client = "^0.20.0"
|
||||
|
||||
[build-system]
|
||||
requires = ["poetry-core"]
|
||||
build-backend = "poetry.core.masonry.api"
|
||||
|
||||
[tool.black]
|
||||
line-length = 92
|
||||
|
||||
[tool.pytest.ini_options]
|
||||
asyncio_default_fixture_loop_scope = "module"
|
||||
|
||||
[tool.ruff]
|
||||
target-version = "py312"
|
||||
line-length = 92
|
||||
17
scripts/postgres/backup
Executable file
17
scripts/postgres/backup
Executable file
|
|
@ -0,0 +1,17 @@
|
|||
#!/bin/sh -e
|
||||
|
||||
echo "Backup process started."
|
||||
|
||||
export POSTGRES_USER="${POSTGRES_USER}"
|
||||
|
||||
# Save the current date in YYYY-MM-DD format to a variable
|
||||
current_datetime=$(date +%Y-%m-%d-%H%M%S)
|
||||
|
||||
backup_directory="/backups"
|
||||
backup_filename="${backup_directory}/backup-${current_datetime}.dump.gz"
|
||||
|
||||
# Run pg_dump and compress its output, then save to /backups with the current date in the filename
|
||||
pg_dump -Fc app -U "$POSTGRES_USER" | gzip > "$backup_filename"
|
||||
|
||||
|
||||
echo "Backup has been created and saved to ${backup_filename}"
|
||||
38
scripts/postgres/restore
Executable file
38
scripts/postgres/restore
Executable file
|
|
@ -0,0 +1,38 @@
|
|||
#!/bin/sh -e
|
||||
|
||||
# The directory where backups are stored
|
||||
BACKUP_DIRECTORY="/backups"
|
||||
|
||||
# Check if a file name was provided as a parameter
|
||||
if [ $# -eq 0 ]; then
|
||||
echo "No file name provided. Please provide a file name to check."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# The file name is taken from the first argument provided to the script
|
||||
file_name="$1"
|
||||
|
||||
# Full path to the file
|
||||
full_file_path="${BACKUP_DIRECTORY}/${file_name}"
|
||||
|
||||
# Check if the file exists
|
||||
if [ -f "$full_file_path" ]; then
|
||||
echo "File ${file_name} exists."
|
||||
else
|
||||
echo "File ${file_name} does not exist."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
export POSTGRES_USER="${POSTGRES_USER}"
|
||||
export POSTGRES_DB="${POSTGRES_DB}"
|
||||
|
||||
echo "Dropping the database..."
|
||||
dropdb "$POSTGRES_DB" -U "$POSTGRES_USER"
|
||||
|
||||
echo "Creating a new database..."
|
||||
createdb "$POSTGRES_DB" --owner="$POSTGRES_USER" -U "$POSTGRES_USER"
|
||||
|
||||
echo "Applying the backup to the new database..."
|
||||
gunzip -c "${full_file_path}" | pg_restore -U "$POSTGRES_USER" -d "$POSTGRES_DB"
|
||||
|
||||
echo "Backup applied successfully."
|
||||
70
src/API.md
Normal file
70
src/API.md
Normal file
|
|
@ -0,0 +1,70 @@
|
|||
:::warning[Under construction]
|
||||
This documentation is a work in progress. Please [get in touch](/contact) with us if you have any questions.
|
||||
:::
|
||||
|
||||
Learn how to use the jasima.app API.
|
||||
Everything that can be done within the portal can also be achieved using our API, documented in this section.
|
||||
|
||||
## Authentication
|
||||
|
||||
We use an API key for authentication.
|
||||
This should be passed to the API in an `Authorization` header:
|
||||
|
||||
```http request
|
||||
Authorization: Bearer <token>
|
||||
```
|
||||
|
||||
## Data Format
|
||||
|
||||
The API uses JSON as the default format for both requests and responses, ensuring consistency and simplicity in data
|
||||
handling:
|
||||
|
||||
* For GET requests, the API expects parameters to be passed via query strings.
|
||||
* For other HTTP methods (e.g., POST, PUT), the request body (if required) should be formatted as JSON.
|
||||
|
||||
There is no need to explicitly specify content encoding in the headers, as the API assumes JSON by default for all
|
||||
applicable requests and responses.
|
||||
|
||||
## Pagination
|
||||
|
||||
The API supports pagination to manage large datasets efficiently. Use the page and perPage query parameters to control the data returned in your requests.
|
||||
Parameters:
|
||||
|
||||
* `page`: The page number to retrieve.
|
||||
* `size`: The number of items to include per page.
|
||||
|
||||
Example request:
|
||||
|
||||
```http request
|
||||
GET https://api.jasima.app/api/v1/widgets?page=1&size=10
|
||||
```
|
||||
|
||||
Example response:
|
||||
|
||||
```http request
|
||||
{
|
||||
"items": [
|
||||
],
|
||||
"total": 0,
|
||||
"page": 1,
|
||||
"size": 10,
|
||||
"pages": 0
|
||||
}
|
||||
```
|
||||
|
||||
Bear in mind the following:
|
||||
|
||||
* The `page` field in the response indicates the current page being viewed.
|
||||
* The `total` field shows the total number of items available.
|
||||
* The `pages` field indicates the total number of pages, and can be used to determine whether there are additional pages
|
||||
to fetch.
|
||||
|
||||
Use this structure to efficiently navigate through large datasets while ensuring optimal performance.
|
||||
|
||||
## Errors
|
||||
|
||||
:::info[TODO]
|
||||
Document the common status codes, and the validation error response format.
|
||||
:::
|
||||
|
||||
## Schemas
|
||||
0
src/__init__.py
Normal file
0
src/__init__.py
Normal file
56
src/config.py
Normal file
56
src/config.py
Normal file
|
|
@ -0,0 +1,56 @@
|
|||
from os.path import abspath, dirname, join
|
||||
from typing import Any
|
||||
|
||||
from pydantic import PostgresDsn
|
||||
from pydantic_settings import BaseSettings, SettingsConfigDict
|
||||
|
||||
from src.constants import Environment
|
||||
|
||||
API_README_PATH = abspath(join(dirname(__file__), "API.md"))
|
||||
|
||||
with open(API_README_PATH, "r", encoding="utf-8") as f:
|
||||
API_README_MD = f.read()
|
||||
|
||||
|
||||
class CustomBaseSettings(BaseSettings):
|
||||
model_config = SettingsConfigDict(
|
||||
env_file=".env", env_file_encoding="utf-8", extra="ignore"
|
||||
)
|
||||
|
||||
|
||||
class Config(CustomBaseSettings):
|
||||
DATABASE_URL: PostgresDsn # Used for alembic migrations
|
||||
DATABASE_ASYNC_URL: PostgresDsn
|
||||
DATABASE_POOL_SIZE: int = 16
|
||||
DATABASE_POOL_TTL: int = 60 * 20 # 20 minutes
|
||||
DATABASE_POOL_PRE_PING: bool = True
|
||||
|
||||
ENVIRONMENT: Environment = Environment.PRODUCTION
|
||||
|
||||
CORS_ORIGINS: list[str] = ["*"]
|
||||
CORS_ORIGINS_REGEX: str | None = None
|
||||
CORS_HEADERS: list[str] = ["*"]
|
||||
|
||||
API_KEY: str
|
||||
|
||||
APP_VERSION: str = "0.0.0"
|
||||
|
||||
|
||||
settings = Config()
|
||||
|
||||
tags_metadata = [
|
||||
{
|
||||
"name": "OpenTofu",
|
||||
"description": "Operations for managing OpenTofu deployments.",
|
||||
},
|
||||
]
|
||||
|
||||
app_configs: dict[str, Any] = {
|
||||
"title": "pali lili API",
|
||||
"version": settings.APP_VERSION,
|
||||
"description": API_README_MD,
|
||||
"openapi_tags": tags_metadata,
|
||||
}
|
||||
|
||||
if not settings.ENVIRONMENT.is_debug:
|
||||
app_configs["openapi_url"] = None # hide docs
|
||||
32
src/constants.py
Normal file
32
src/constants.py
Normal file
|
|
@ -0,0 +1,32 @@
|
|||
from enum import Enum
|
||||
|
||||
DB_NAMING_CONVENTION = {
|
||||
"ix": "%(column_0_label)s_idx",
|
||||
"uq": "%(table_name)s_%(column_0_name)s_key",
|
||||
"ck": "%(table_name)s_%(constraint_name)s_check",
|
||||
"fk": "%(table_name)s_%(column_0_name)s_fkey",
|
||||
"pk": "%(table_name)s_pkey",
|
||||
}
|
||||
|
||||
|
||||
class Environment(str, Enum):
|
||||
LOCAL = "LOCAL"
|
||||
TESTING = "TESTING"
|
||||
STAGING = "STAGING"
|
||||
PRODUCTION = "PRODUCTION"
|
||||
|
||||
@property
|
||||
def is_debug(self):
|
||||
return self in (self.LOCAL, self.STAGING, self.TESTING)
|
||||
|
||||
@property
|
||||
def is_local(self):
|
||||
return self is Environment.LOCAL
|
||||
|
||||
@property
|
||||
def is_testing(self):
|
||||
return self == self.TESTING
|
||||
|
||||
@property
|
||||
def is_deployed(self) -> bool:
|
||||
return self in (self.STAGING, self.PRODUCTION)
|
||||
57
src/database.py
Normal file
57
src/database.py
Normal file
|
|
@ -0,0 +1,57 @@
|
|||
import contextlib
|
||||
from typing import AsyncIterator, AsyncGenerator, Annotated
|
||||
|
||||
from fastapi import Depends
|
||||
from sqlalchemy import (
|
||||
MetaData,
|
||||
)
|
||||
from sqlalchemy.ext.asyncio import (
|
||||
AsyncConnection,
|
||||
create_async_engine,
|
||||
AsyncSession,
|
||||
async_sessionmaker,
|
||||
)
|
||||
|
||||
from src.config import settings
|
||||
from src.constants import DB_NAMING_CONVENTION
|
||||
|
||||
DATABASE_URL = str(settings.DATABASE_ASYNC_URL)
|
||||
|
||||
engine = create_async_engine(
|
||||
DATABASE_URL,
|
||||
pool_size=settings.DATABASE_POOL_SIZE,
|
||||
pool_recycle=settings.DATABASE_POOL_TTL,
|
||||
pool_pre_ping=settings.DATABASE_POOL_PRE_PING,
|
||||
)
|
||||
metadata = MetaData(naming_convention=DB_NAMING_CONVENTION)
|
||||
sessionmaker = async_sessionmaker(autocommit=False, expire_on_commit=False, bind=engine)
|
||||
|
||||
|
||||
@contextlib.asynccontextmanager
|
||||
async def get_db_connection() -> AsyncIterator[AsyncConnection]:
|
||||
async with engine.begin() as connection:
|
||||
try:
|
||||
yield connection
|
||||
except Exception:
|
||||
await connection.rollback()
|
||||
raise
|
||||
|
||||
|
||||
@contextlib.asynccontextmanager
|
||||
async def get_db_session() -> AsyncIterator[AsyncSession]:
|
||||
session = sessionmaker()
|
||||
try:
|
||||
yield session
|
||||
except Exception:
|
||||
await session.rollback()
|
||||
raise
|
||||
finally:
|
||||
await session.close()
|
||||
|
||||
|
||||
async def get_db() -> AsyncGenerator[AsyncSession, None]:
|
||||
async with get_db_session() as session:
|
||||
yield session
|
||||
|
||||
|
||||
DbSession = Annotated[AsyncSession, Depends(get_db)]
|
||||
33
src/exceptions.py
Normal file
33
src/exceptions.py
Normal file
|
|
@ -0,0 +1,33 @@
|
|||
from typing import Any
|
||||
|
||||
from fastapi import HTTPException, status
|
||||
|
||||
|
||||
class DetailedHTTPException(HTTPException):
|
||||
STATUS_CODE = status.HTTP_500_INTERNAL_SERVER_ERROR
|
||||
DETAIL = "Server error"
|
||||
|
||||
def __init__(self, **kwargs: dict[str, Any]) -> None:
|
||||
super().__init__(status_code=self.STATUS_CODE, detail=self.DETAIL, **kwargs)
|
||||
|
||||
|
||||
class PermissionDenied(DetailedHTTPException):
|
||||
STATUS_CODE = status.HTTP_403_FORBIDDEN
|
||||
DETAIL = "Permission denied"
|
||||
|
||||
|
||||
class NotFound(DetailedHTTPException):
|
||||
STATUS_CODE = status.HTTP_404_NOT_FOUND
|
||||
|
||||
|
||||
class BadRequest(DetailedHTTPException):
|
||||
STATUS_CODE = status.HTTP_400_BAD_REQUEST
|
||||
DETAIL = "Bad Request"
|
||||
|
||||
|
||||
class NotAuthenticated(DetailedHTTPException):
|
||||
STATUS_CODE = status.HTTP_401_UNAUTHORIZED
|
||||
DETAIL = "User not authenticated"
|
||||
|
||||
def __init__(self) -> None:
|
||||
super().__init__(headers={"WWW-Authenticate": "Bearer"})
|
||||
44
src/main.py
Normal file
44
src/main.py
Normal file
|
|
@ -0,0 +1,44 @@
|
|||
from contextlib import asynccontextmanager
|
||||
from typing import AsyncGenerator
|
||||
|
||||
from fastapi import FastAPI
|
||||
from fastapi_pagination import add_pagination
|
||||
from starlette.middleware.cors import CORSMiddleware
|
||||
|
||||
from src.config import app_configs, settings
|
||||
from src.tofu.exceptions import TofuStateLocked
|
||||
from src.tofu.router import router as tofu_router, tofu_state_locked_handler
|
||||
from src.tofu.tasks import periodic_tofu_process_tasks
|
||||
|
||||
|
||||
@asynccontextmanager
|
||||
async def lifespan(_application: FastAPI) -> AsyncGenerator:
|
||||
await periodic_tofu_process_tasks()
|
||||
# Startup
|
||||
yield
|
||||
# Shutdown
|
||||
|
||||
|
||||
app = FastAPI(**app_configs, lifespan=lifespan)
|
||||
add_pagination(app)
|
||||
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=settings.CORS_ORIGINS,
|
||||
allow_origin_regex=settings.CORS_ORIGINS_REGEX,
|
||||
allow_credentials=True,
|
||||
allow_methods=("GET", "POST", "PUT", "PATCH", "DELETE", "OPTIONS"),
|
||||
allow_headers=settings.CORS_HEADERS,
|
||||
)
|
||||
|
||||
|
||||
app.add_exception_handler(TofuStateLocked, tofu_state_locked_handler)
|
||||
app.include_router(
|
||||
tofu_router,
|
||||
prefix="/api/v1/tofu",
|
||||
)
|
||||
|
||||
|
||||
@app.get("/healthcheck", include_in_schema=False)
|
||||
async def healthcheck() -> dict[str, str]:
|
||||
return {"status": "ok"}
|
||||
50
src/models.py
Normal file
50
src/models.py
Normal file
|
|
@ -0,0 +1,50 @@
|
|||
from datetime import datetime
|
||||
from typing import Any
|
||||
|
||||
from sqlalchemy import JSON, DateTime, func, ForeignKey
|
||||
from sqlalchemy.orm import (
|
||||
DeclarativeBase,
|
||||
Mapped,
|
||||
mapped_column,
|
||||
relationship,
|
||||
declared_attr,
|
||||
)
|
||||
|
||||
from src.database import metadata
|
||||
|
||||
|
||||
class CustomBase(DeclarativeBase):
|
||||
type_annotation_map = {
|
||||
datetime: DateTime(timezone=True),
|
||||
dict[str, Any]: JSON,
|
||||
}
|
||||
metadata = metadata
|
||||
|
||||
|
||||
class ActivatedMixin:
|
||||
active: Mapped[bool] = mapped_column(default=True)
|
||||
|
||||
|
||||
class DeletedTimestampMixin:
|
||||
deleted_at: Mapped[datetime | None] = mapped_column(nullable=True)
|
||||
|
||||
|
||||
class DescriptionMixin:
|
||||
description: Mapped[str]
|
||||
|
||||
|
||||
class IdMixin:
|
||||
id: Mapped[int] = mapped_column(primary_key=True, autoincrement=True)
|
||||
|
||||
|
||||
class TimestampMixin:
|
||||
created_at: Mapped[datetime] = mapped_column(default=func.now())
|
||||
updated_at: Mapped[datetime] = mapped_column(default=func.now(), onupdate=func.now())
|
||||
|
||||
|
||||
class TofuInstanceMixin:
|
||||
tofu_instance_id: Mapped[int] = mapped_column(ForeignKey("tofu_instance.id"))
|
||||
|
||||
@declared_attr
|
||||
def tofu_instance(cls):
|
||||
return relationship("TofuInstance")
|
||||
91
src/schemas.py
Normal file
91
src/schemas.py
Normal file
|
|
@ -0,0 +1,91 @@
|
|||
from datetime import datetime, timezone
|
||||
from typing import Annotated, Literal
|
||||
from uuid import UUID
|
||||
|
||||
from email_validator.rfc_constants import DOMAIN_NAME_REGEX
|
||||
from fastapi.encoders import jsonable_encoder
|
||||
from pydantic import BaseModel, ConfigDict, PlainSerializer, RootModel, StringConstraints
|
||||
|
||||
|
||||
def datetime_to_utc_str(dt: datetime) -> str:
|
||||
if not dt.tzinfo:
|
||||
dt = dt.replace(tzinfo=timezone.utc)
|
||||
else:
|
||||
dt = dt.astimezone(timezone.utc)
|
||||
return dt.strftime("%Y-%m-%dT%H:%M:%SZ")
|
||||
|
||||
|
||||
CustomDatetime = Annotated[datetime, PlainSerializer(datetime_to_utc_str, return_type=str)]
|
||||
|
||||
|
||||
class CustomModelMixin:
|
||||
model_config = ConfigDict(
|
||||
populate_by_name=True,
|
||||
)
|
||||
|
||||
def serializable_dict(self, **kwargs):
|
||||
"""Return a dict which contains only serializable fields."""
|
||||
default_dict = self.model_dump()
|
||||
|
||||
return jsonable_encoder(default_dict, custom_encoder={UUID: str})
|
||||
|
||||
|
||||
class CustomModel(BaseModel, CustomModelMixin):
|
||||
pass
|
||||
|
||||
|
||||
class CustomRootModel(RootModel, CustomModelMixin):
|
||||
pass
|
||||
|
||||
|
||||
class OkResponse(CustomModel):
|
||||
"""
|
||||
A generic payload to respond with to indicate success where no other information is required or available.
|
||||
Pairs well with HTTP status code 200.
|
||||
"""
|
||||
|
||||
ok: Literal[True]
|
||||
|
||||
|
||||
class AcceptedResponse(CustomModel):
|
||||
"""
|
||||
A generic payload to respond with where a request has been accepted but perhaps has not yet been processed.
|
||||
Pairs well with HTTP status code 201.
|
||||
"""
|
||||
|
||||
id: int
|
||||
|
||||
|
||||
class CreatedResponse(CustomModel):
|
||||
"""
|
||||
A generic payload to respond with where a request has resulted in a thing being created.
|
||||
Pairs well with HTTP status code 202.
|
||||
"""
|
||||
|
||||
id: int
|
||||
|
||||
|
||||
class ActivatedMixin:
|
||||
active: bool
|
||||
|
||||
|
||||
class DescriptionMixin:
|
||||
description: str
|
||||
|
||||
|
||||
class DeletedTimestampMixin:
|
||||
deleted_at: CustomDatetime
|
||||
|
||||
|
||||
class IdMixin:
|
||||
id: int
|
||||
|
||||
|
||||
class TimestampMixin:
|
||||
created_at: CustomDatetime
|
||||
updated_at: CustomDatetime
|
||||
|
||||
|
||||
DomainName = Annotated[
|
||||
str, StringConstraints(strip_whitespace=True, to_lower=True, pattern=DOMAIN_NAME_REGEX)
|
||||
]
|
||||
0
src/tofu/__init__.py
Normal file
0
src/tofu/__init__.py
Normal file
165
src/tofu/client.py
Normal file
165
src/tofu/client.py
Normal file
|
|
@ -0,0 +1,165 @@
|
|||
import asyncio
|
||||
import json
|
||||
import os
|
||||
import shutil
|
||||
import tempfile
|
||||
from datetime import datetime, timezone
|
||||
from json import JSONDecodeError
|
||||
from typing import Any, TYPE_CHECKING
|
||||
|
||||
import aiofiles
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
|
||||
from src.tofu.config import settings
|
||||
from src.tofu.exceptions import TofuExecutionError, TofuTypeError
|
||||
from src.tofu.models import TofuInstanceTask, TofuInstanceTaskLog
|
||||
from src.tofu.security import generate_password, generate_password_hash
|
||||
|
||||
|
||||
def _convert_python_to_tf(value: Any) -> Any:
|
||||
"""Convert Python types to Terraform-compatible types."""
|
||||
if isinstance(value, bool):
|
||||
return value
|
||||
elif isinstance(value, (int, float)):
|
||||
return value
|
||||
elif isinstance(value, str):
|
||||
return value
|
||||
elif isinstance(value, list):
|
||||
return [_convert_python_to_tf(item) for item in value]
|
||||
elif isinstance(value, dict):
|
||||
return {str(k): _convert_python_to_tf(v) for k, v in value.items()}
|
||||
elif value is None:
|
||||
return None
|
||||
else:
|
||||
raise TofuTypeError(f"Unsupported type for Terraform conversion: {type(value)}")
|
||||
|
||||
|
||||
class TofuManager:
|
||||
def __init__(
|
||||
self,
|
||||
db: AsyncSession,
|
||||
instance_task: TofuInstanceTask,
|
||||
working_dir: str | None = None,
|
||||
):
|
||||
self.db = db
|
||||
self.instance_task = instance_task
|
||||
self._is_temp_dir = working_dir is None
|
||||
self._working_dir = working_dir
|
||||
self.tofu_path = settings.OPENTOFU_PATH
|
||||
|
||||
async def __aenter__(self) -> "TofuManager":
|
||||
return self
|
||||
|
||||
async def __aexit__(self, exc_type: Any, exc_value: Any, traceback: Any) -> None:
|
||||
self.instance_task.instance.state_password = None
|
||||
await self.db.commit()
|
||||
# Only clean up temporary directories if there was no exception
|
||||
if self._is_temp_dir and self._working_dir and exc_type is None:
|
||||
shutil.rmtree(self._working_dir)
|
||||
|
||||
async def _create_config(self) -> None:
|
||||
config = self.instance_task.instance.configuration
|
||||
self.password = generate_password()
|
||||
self.instance_task.instance.state_password = generate_password_hash(self.password)
|
||||
await self.db.commit()
|
||||
if "terraform" not in config:
|
||||
config["terraform"] = {}
|
||||
config["terraform"]["backend"] = {
|
||||
"http": {
|
||||
"address": f"http://localhost:8000/api/v1/tofu/instances/{self.instance_task.instance_id}/state",
|
||||
"lock_address": f"http://localhost:8000/api/v1/tofu/instances/{self.instance_task.instance_id}/state",
|
||||
"unlock_address": f"http://localhost:8000/api/v1/tofu/instances/{self.instance_task.instance_id}/state",
|
||||
}
|
||||
}
|
||||
async with aiofiles.open(await self.config_file(), "w") as f:
|
||||
await f.write(json.dumps(config, indent=2))
|
||||
|
||||
async def _process_output_line(self, line: str) -> None:
|
||||
try:
|
||||
data = json.loads(line)
|
||||
except json.decoder.JSONDecodeError:
|
||||
data = {
|
||||
"@level": "info",
|
||||
"@timestamp": datetime.now(tz=timezone.utc).strftime(
|
||||
"%Y-%m-%dT%H:%M:%S.%f%z"
|
||||
),
|
||||
"@module": "manager",
|
||||
"@message": line,
|
||||
"type": "decode-error",
|
||||
}
|
||||
log = TofuInstanceTaskLog(
|
||||
instance_task_id=self.instance_task.id,
|
||||
timestamp=datetime.strptime(data["@timestamp"], "%Y-%m-%dT%H:%M:%S.%f%z"),
|
||||
log=data,
|
||||
)
|
||||
self.db.add(log)
|
||||
await self.db.commit()
|
||||
|
||||
async def _run_command(
|
||||
self,
|
||||
command: str,
|
||||
*args: str,
|
||||
json_output: bool = False,
|
||||
log_output: bool = True,
|
||||
return_output: bool = False,
|
||||
) -> str | None:
|
||||
cmd = [self.tofu_path, command] + list(args)
|
||||
if json_output:
|
||||
cmd.append("-json")
|
||||
process = await asyncio.create_subprocess_exec(
|
||||
*cmd,
|
||||
cwd=await self.working_dir(),
|
||||
stdout=asyncio.subprocess.PIPE,
|
||||
stderr=asyncio.subprocess.DEVNULL,
|
||||
env={
|
||||
"TF_HTTP_USERNAME": "tofu",
|
||||
"TF_HTTP_PASSWORD": self.password,
|
||||
},
|
||||
)
|
||||
if TYPE_CHECKING:
|
||||
assert process.stdout is not None
|
||||
result = []
|
||||
async for line in process.stdout:
|
||||
if return_output:
|
||||
result.append(line.decode("utf-8"))
|
||||
if log_output:
|
||||
await self._process_output_line(line.decode("utf-8"))
|
||||
await process.wait()
|
||||
if process.returncode != 0:
|
||||
raise TofuExecutionError(f"Tofu command failed: {process.returncode}")
|
||||
return "\n".join(result) if return_output else None
|
||||
|
||||
async def config_file(self) -> str:
|
||||
return os.path.join(await self.working_dir(), "main.tf.json")
|
||||
|
||||
async def working_dir(self) -> str:
|
||||
if self._working_dir:
|
||||
return self._working_dir
|
||||
self._working_dir = await asyncio.to_thread(tempfile.mkdtemp, prefix="tofu-")
|
||||
return self._working_dir
|
||||
|
||||
async def init(self, upgrade: bool = True) -> None:
|
||||
await self._create_config()
|
||||
args = []
|
||||
if upgrade:
|
||||
args.append("-upgrade")
|
||||
await self._run_command("init", *args, json_output=True)
|
||||
|
||||
async def apply(self) -> None:
|
||||
await self._create_config()
|
||||
await self._run_command("apply", "-auto-approve", json_output=True)
|
||||
|
||||
async def destroy(self) -> None:
|
||||
await self._create_config()
|
||||
await self._run_command("destroy", "-auto-approve", json_output=True)
|
||||
|
||||
async def output(self) -> None:
|
||||
await self._create_config()
|
||||
try:
|
||||
outputs = await self._run_command(
|
||||
"output", json_output=True, log_output=False, return_output=True
|
||||
)
|
||||
self.instance_task.instance.outputs = json.loads(outputs)
|
||||
await self.db.commit()
|
||||
except JSONDecodeError:
|
||||
raise TofuExecutionError("Could not parse JSON output")
|
||||
13
src/tofu/config.py
Normal file
13
src/tofu/config.py
Normal file
|
|
@ -0,0 +1,13 @@
|
|||
from os.path import abspath, dirname, join
|
||||
|
||||
from src.config import CustomBaseSettings
|
||||
|
||||
|
||||
class Config(CustomBaseSettings):
|
||||
OPENTOFU_PATH: str = "/usr/bin/tofu"
|
||||
OPENTOFU_MODULES_PATH: str = abspath(
|
||||
join(dirname(__file__), "..", "..", "tofu")
|
||||
) # no trailing slash from abspath
|
||||
|
||||
|
||||
settings = Config()
|
||||
39
src/tofu/exceptions.py
Normal file
39
src/tofu/exceptions.py
Normal file
|
|
@ -0,0 +1,39 @@
|
|||
from typing import Any
|
||||
|
||||
from starlette import status
|
||||
from starlette.exceptions import HTTPException
|
||||
|
||||
|
||||
class TofuExecutionError(RuntimeError):
|
||||
"""An error occurred when executing an OpenTofu command."""
|
||||
|
||||
|
||||
class TofuTypeError(TypeError):
|
||||
"""A value provided used a type that cannot be converted to a supported OpenTofu type."""
|
||||
|
||||
|
||||
class TofuStateLocked(RuntimeError):
|
||||
"""A request could not be completed as the state is locked, and the requestor does not hold the lock."""
|
||||
|
||||
def __init__(self, lock: dict[str, Any]) -> None:
|
||||
self.lock = lock
|
||||
|
||||
|
||||
class TofuForbiddenException(HTTPException):
|
||||
def __init__(
|
||||
self,
|
||||
**kwargs: Any,
|
||||
) -> None:
|
||||
super().__init__(status_code=status.HTTP_403_FORBIDDEN, detail="Forbidden")
|
||||
|
||||
|
||||
class TofuUnauthorizedException(HTTPException):
|
||||
def __init__(
|
||||
self,
|
||||
**kwargs: Any,
|
||||
) -> None:
|
||||
super().__init__(
|
||||
status_code=status.HTTP_401_UNAUTHORIZED,
|
||||
detail="Unauthorized",
|
||||
headers={"WWW-Authenticate": 'Basic realm="jasima"'},
|
||||
)
|
||||
107
src/tofu/models.py
Normal file
107
src/tofu/models.py
Normal file
|
|
@ -0,0 +1,107 @@
|
|||
from datetime import datetime
|
||||
from enum import Enum
|
||||
from typing import Any
|
||||
|
||||
from sqlalchemy import ForeignKey, func, text
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||
|
||||
from src.models import CustomBase, TimestampMixin, IdMixin, DeletedTimestampMixin
|
||||
|
||||
|
||||
class TofuInstanceStatus(Enum):
|
||||
ACTIVE = "ACTIVE"
|
||||
DEPLOYING = "DEPLOYING"
|
||||
DESTROYED = "DESTROYED"
|
||||
DESTROYING = "DESTROYING"
|
||||
DRIFTED = "DRIFTED"
|
||||
FAILED = "FAILED"
|
||||
FAILED_DESTROY = "FAILED_DESTROY"
|
||||
PENDING = "PENDING"
|
||||
PENDING_DESTROY = "PENDING_DESTROY"
|
||||
PENDING_DRIFT_CHECK = "PENDING_DRIFT_CHECK"
|
||||
|
||||
|
||||
class TofuInstance(CustomBase, IdMixin, TimestampMixin, DeletedTimestampMixin):
|
||||
__tablename__ = "tofu_instance"
|
||||
|
||||
status: Mapped[TofuInstanceStatus] = mapped_column(default=TofuInstanceStatus.PENDING)
|
||||
configuration: Mapped[dict[str, Any]]
|
||||
outputs: Mapped[dict[str, Any] | None]
|
||||
plan: Mapped[dict[str, Any] | None]
|
||||
state: Mapped[dict[str, Any] | None]
|
||||
state_password: Mapped[bytes | None]
|
||||
state_lock: Mapped[dict[str, Any] | None]
|
||||
status_changed_at: Mapped[datetime] = mapped_column(default=func.now())
|
||||
drift_checked_at: Mapped[datetime | None]
|
||||
tasks = relationship("TofuInstanceTask", back_populates="instance")
|
||||
status_changes = relationship("TofuInstanceStatusChange", back_populates="instance")
|
||||
|
||||
|
||||
class TofuInstanceStatusChange(CustomBase, IdMixin):
|
||||
__tablename__ = "tofu_instance_status_change"
|
||||
|
||||
instance_id: Mapped[int] = mapped_column(ForeignKey("tofu_instance.id"))
|
||||
instance_task_id: Mapped[int] = mapped_column(ForeignKey("tofu_instance_task.id"))
|
||||
timestamp: Mapped[datetime] = mapped_column(default=func.now())
|
||||
old_status: Mapped[TofuInstanceStatus]
|
||||
new_status: Mapped[TofuInstanceStatus]
|
||||
instance = relationship("TofuInstance", back_populates="status_changes")
|
||||
|
||||
|
||||
class TofuInstanceTaskType(Enum):
|
||||
CHECK_DRIFT = "CHECK_DRIFT"
|
||||
DEPLOY = "DEPLOY"
|
||||
DESTROY = "DESTROY"
|
||||
|
||||
|
||||
class TofuInstanceTaskStatus(Enum):
|
||||
CANCELED = "CANCELED"
|
||||
COMPLETED = "COMPLETED"
|
||||
FAILED = "FAILED"
|
||||
PENDING = "PENDING"
|
||||
RUNNING = "RUNNING"
|
||||
|
||||
|
||||
class TofuInstanceTask(CustomBase, IdMixin, TimestampMixin):
|
||||
__tablename__ = "tofu_instance_task"
|
||||
|
||||
instance_id: Mapped[int] = mapped_column(ForeignKey("tofu_instance.id"))
|
||||
task: Mapped[TofuInstanceTaskType]
|
||||
status: Mapped[TofuInstanceTaskStatus] = mapped_column(
|
||||
default=TofuInstanceTaskStatus.PENDING
|
||||
)
|
||||
start_time: Mapped[datetime | None]
|
||||
end_time: Mapped[datetime | None]
|
||||
instance = relationship("TofuInstance", back_populates="tasks")
|
||||
|
||||
|
||||
class TofuInstanceTaskLog(CustomBase, IdMixin):
|
||||
__tablename__ = "tofu_instance_task_log"
|
||||
|
||||
instance_task_id: Mapped[int] = mapped_column(ForeignKey("tofu_instance_task.id"))
|
||||
timestamp: Mapped[datetime] = mapped_column(default=func.now())
|
||||
log: Mapped[dict[str, Any]]
|
||||
|
||||
|
||||
class TofuBruteForce(CustomBase, IdMixin, TimestampMixin):
|
||||
__tablename__ = "tofu_brute_force"
|
||||
|
||||
host: Mapped[str]
|
||||
expiry: Mapped[datetime] = mapped_column(default=func.now() + text("INTERVAL '1 hour'"))
|
||||
|
||||
|
||||
def update_tofu_instance_status(
|
||||
db: AsyncSession, instance: TofuInstance, task_id: int, new_status: TofuInstanceStatus
|
||||
) -> None:
|
||||
status_change = TofuInstanceStatusChange(
|
||||
instance_id=instance.id,
|
||||
instance_task_id=task_id,
|
||||
old_status=instance.status,
|
||||
new_status=new_status,
|
||||
)
|
||||
db.add(status_change)
|
||||
instance.status = new_status
|
||||
instance.status_changed_at = func.now()
|
||||
if new_status == TofuInstanceStatus.DESTROYED:
|
||||
instance.deleted_at = func.now()
|
||||
356
src/tofu/router.py
Normal file
356
src/tofu/router.py
Normal file
|
|
@ -0,0 +1,356 @@
|
|||
import asyncio
|
||||
import json
|
||||
from datetime import datetime, timezone, timedelta
|
||||
from typing import Any, Annotated, AsyncIterator
|
||||
|
||||
from fastapi import APIRouter, HTTPException, Request, Query
|
||||
from fastapi_pagination import Page
|
||||
from fastapi_pagination.ext.sqlalchemy import apaginate
|
||||
from sqlalchemy import select
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from sqlalchemy.orm import selectinload
|
||||
from starlette import status
|
||||
from starlette.responses import JSONResponse, Response, StreamingResponse
|
||||
|
||||
from src.database import DbSession
|
||||
from src.schemas import AcceptedResponse, OkResponse
|
||||
from src.tofu.exceptions import TofuStateLocked, TofuUnauthorizedException
|
||||
from src.tofu.models import (
|
||||
TofuInstance,
|
||||
TofuInstanceTask,
|
||||
TofuInstanceTaskType,
|
||||
TofuInstanceStatus,
|
||||
TofuInstanceTaskLog,
|
||||
TofuInstanceTaskStatus,
|
||||
update_tofu_instance_status,
|
||||
)
|
||||
from src.tofu.schemas import (
|
||||
TofuInstanceDetail,
|
||||
TofuInstanceCreate,
|
||||
TofuInstanceSummary,
|
||||
TofuInstanceUpdate,
|
||||
TofuInstanceState,
|
||||
TofuInstanceStateLock,
|
||||
)
|
||||
from src.tofu.security import (
|
||||
Password,
|
||||
validate_password,
|
||||
generate_password_hash,
|
||||
block_host_brute_force,
|
||||
APIKey,
|
||||
)
|
||||
|
||||
router = APIRouter()
|
||||
|
||||
TofuLockID = Annotated[str | None, Query(alias="ID", description="State lock ID")]
|
||||
|
||||
|
||||
async def tofu_state_locked_handler(request: Request, exc: Exception) -> Response:
|
||||
if isinstance(exc, TofuStateLocked):
|
||||
return JSONResponse(status_code=status.HTTP_423_LOCKED, content=exc.lock)
|
||||
raise RuntimeError(exc)
|
||||
|
||||
|
||||
@router.get(
|
||||
"/instances",
|
||||
response_model=Page[TofuInstanceSummary],
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def list_tofu_instances(db: DbSession, _: APIKey) -> Page[TofuInstance]:
|
||||
return await apaginate(
|
||||
db, select(TofuInstance).filter(TofuInstance.deleted_at == None) # noqa: E711
|
||||
)
|
||||
|
||||
|
||||
@router.post(
|
||||
"/instances",
|
||||
response_model=AcceptedResponse,
|
||||
status_code=status.HTTP_202_ACCEPTED,
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def create_tofu_instance(
|
||||
instance_data: TofuInstanceCreate, db: DbSession, _: APIKey
|
||||
) -> JSONResponse:
|
||||
hashed_password: str | None = (
|
||||
generate_password_hash(instance_data.password) if instance_data.password else None
|
||||
)
|
||||
instance = TofuInstance(
|
||||
configuration=instance_data.configuration, state_password=hashed_password
|
||||
)
|
||||
db.add(instance)
|
||||
await db.flush()
|
||||
task = TofuInstanceTask(instance_id=instance.id, task=TofuInstanceTaskType.DEPLOY)
|
||||
db.add(task)
|
||||
await db.commit()
|
||||
await db.refresh(instance)
|
||||
return JSONResponse(status_code=status.HTTP_202_ACCEPTED, content={"id": instance.id})
|
||||
|
||||
|
||||
async def _get_tofu_instance(
|
||||
db: AsyncSession,
|
||||
instance_id: int,
|
||||
password: str | None = None,
|
||||
host: str | None = None,
|
||||
*,
|
||||
detail: bool = False,
|
||||
password_required: bool = True,
|
||||
) -> TofuInstance:
|
||||
if detail:
|
||||
query = (
|
||||
select(TofuInstance)
|
||||
.filter(TofuInstance.id == instance_id)
|
||||
.options(
|
||||
selectinload(TofuInstance.status_changes), selectinload(TofuInstance.tasks)
|
||||
)
|
||||
)
|
||||
else:
|
||||
query = select(TofuInstance).filter(TofuInstance.id == instance_id)
|
||||
row = (await db.execute(query)).first()
|
||||
if not row:
|
||||
raise HTTPException(status_code=404, detail="No instance found")
|
||||
if password_required:
|
||||
if not host:
|
||||
raise TofuUnauthorizedException
|
||||
if not validate_password(password, row[0].state_password):
|
||||
await block_host_brute_force(db, host)
|
||||
raise TofuUnauthorizedException
|
||||
return row[0]
|
||||
|
||||
|
||||
@router.get(
|
||||
"/instances/{instance_id}",
|
||||
response_model=TofuInstanceDetail,
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def get_tofu_instance(instance_id: int, db: DbSession, _: APIKey) -> TofuInstance:
|
||||
return await _get_tofu_instance(db, instance_id, detail=True, password_required=False)
|
||||
|
||||
|
||||
@router.put(
|
||||
"/instances/{instance_id}",
|
||||
response_model=AcceptedResponse,
|
||||
status_code=status.HTTP_202_ACCEPTED,
|
||||
response_description="Update accepted (deployment task pending)",
|
||||
responses={
|
||||
status.HTTP_200_OK: {"description": "Update Successful", "model": AcceptedResponse},
|
||||
status.HTTP_404_NOT_FOUND: {"description": "Instance not found"},
|
||||
status.HTTP_412_PRECONDITION_FAILED: {
|
||||
"description": "Update failed due to current status"
|
||||
},
|
||||
},
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def update_tofu_instance(
|
||||
instance_id: int, instance_data: TofuInstanceUpdate, db: DbSession, _: APIKey
|
||||
) -> JSONResponse:
|
||||
query = select(TofuInstance).with_for_update().filter(TofuInstance.id == instance_id)
|
||||
row = (await db.execute(query)).first()
|
||||
if not row:
|
||||
raise HTTPException(status_code=404, detail="Instance not found")
|
||||
instance = row[0]
|
||||
if instance.status not in [TofuInstanceStatus.ACTIVE, TofuInstanceStatus.DRIFTED]:
|
||||
raise HTTPException(
|
||||
status_code=412, detail="Updates only allowed for active instances"
|
||||
)
|
||||
if instance_data.configuration:
|
||||
instance.configuration = instance_data.configuration
|
||||
task = TofuInstanceTask(instance_id=instance.id, task=TofuInstanceTaskType.DEPLOY)
|
||||
db.add(task)
|
||||
await db.flush()
|
||||
update_tofu_instance_status(db, instance, task.id, TofuInstanceStatus.PENDING)
|
||||
if instance_data.password:
|
||||
instance.password = generate_password_hash(instance_data.password)
|
||||
await db.commit() # release FOR UPDATE lock
|
||||
return JSONResponse(status_code=status.HTTP_202_ACCEPTED, content={"id": instance.id})
|
||||
|
||||
|
||||
@router.delete(
|
||||
"/instances/{instance_id}",
|
||||
status_code=202,
|
||||
response_model=AcceptedResponse,
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def destroy_tofu_instance(instance_id: int, db: DbSession, _: APIKey) -> Response:
|
||||
query = select(TofuInstance).with_for_update().filter(TofuInstance.id == instance_id)
|
||||
row = (await db.execute(query)).first()
|
||||
if not row:
|
||||
raise HTTPException(status_code=404, detail="Resource not found")
|
||||
instance = row[0]
|
||||
if instance.status not in [
|
||||
TofuInstanceStatus.ACTIVE,
|
||||
TofuInstanceStatus.DRIFTED,
|
||||
TofuInstanceStatus.FAILED,
|
||||
]:
|
||||
raise HTTPException(
|
||||
status_code=412,
|
||||
detail="Instance cannot be destroyed currently as it is pending update",
|
||||
)
|
||||
task = TofuInstanceTask(instance_id=instance.id, task=TofuInstanceTaskType.DESTROY)
|
||||
db.add(task)
|
||||
await db.flush()
|
||||
update_tofu_instance_status(db, instance, task.id, TofuInstanceStatus.PENDING_DESTROY)
|
||||
await db.commit() # release FOR UPDATE lock
|
||||
return JSONResponse(status_code=status.HTTP_202_ACCEPTED, content={"id": instance_id})
|
||||
|
||||
|
||||
async def _get_tofu_instance_state(
|
||||
db: AsyncSession,
|
||||
instance_id: int,
|
||||
password: str | None = None,
|
||||
host: str | None = None,
|
||||
*,
|
||||
password_required: bool = True,
|
||||
) -> dict[str, Any]:
|
||||
instance = await _get_tofu_instance(
|
||||
db, instance_id, password, host, password_required=password_required
|
||||
)
|
||||
if not instance.state:
|
||||
raise HTTPException(status_code=404, detail="No state found")
|
||||
return instance.state
|
||||
|
||||
|
||||
@router.get(
|
||||
"/instances/{instance_id}/state",
|
||||
response_model=TofuInstanceState,
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def get_tofu_instance_state_route(
|
||||
instance_id: int, db: DbSession, password: Password, request: Request
|
||||
) -> dict[str, Any]:
|
||||
return await _get_tofu_instance_state(db, instance_id, password, request.client.host)
|
||||
|
||||
|
||||
@router.post(
|
||||
"/instances/{instance_id}/state",
|
||||
response_model=None,
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def update_tofu_instance_state(
|
||||
instance_id: int,
|
||||
new_state: TofuInstanceState,
|
||||
db: DbSession,
|
||||
password: Password,
|
||||
request: Request,
|
||||
lock_id: TofuLockID = None,
|
||||
) -> Response:
|
||||
instance = await _get_tofu_instance(db, instance_id, password, request.client.host)
|
||||
if instance.state_lock and instance.state_lock["ID"] != lock_id:
|
||||
raise TofuStateLocked(instance.state_lock)
|
||||
instance.state = new_state.model_dump()
|
||||
await db.commit()
|
||||
return Response(status_code=200)
|
||||
|
||||
|
||||
@router.api_route(
|
||||
"/instances/{instance_id}/state",
|
||||
methods=["LOCK"],
|
||||
response_model=OkResponse,
|
||||
tags=["OpenTofu"],
|
||||
include_in_schema=False,
|
||||
)
|
||||
async def lock_tofu_instance_state(
|
||||
instance_id: int,
|
||||
lock: TofuInstanceStateLock,
|
||||
db: DbSession,
|
||||
password: Password,
|
||||
request: Request,
|
||||
) -> JSONResponse:
|
||||
instance = await _get_tofu_instance(db, instance_id, password, request.client.host)
|
||||
if instance.state_lock:
|
||||
return JSONResponse(status_code=423, content=instance.state_lock)
|
||||
instance.state_lock = lock.serializable_dict()
|
||||
await db.commit()
|
||||
return JSONResponse(
|
||||
status_code=200, content={"result": "ok"}
|
||||
) # A 204 would make sense here but the spec needs 200
|
||||
|
||||
|
||||
@router.api_route(
|
||||
"/instances/{instance_id}/state",
|
||||
response_model=OkResponse,
|
||||
methods=["UNLOCK"],
|
||||
tags=["OpenTofu"],
|
||||
include_in_schema=False,
|
||||
)
|
||||
async def unlock_tofu_instance_state(
|
||||
instance_id: int,
|
||||
db: DbSession,
|
||||
password: Password,
|
||||
request: Request,
|
||||
lock_id: TofuLockID = None,
|
||||
) -> JSONResponse:
|
||||
instance = await _get_tofu_instance(db, instance_id, password, request.client.host)
|
||||
if not instance.state_lock:
|
||||
JSONResponse(
|
||||
status_code=200, content={"result": "ok"}
|
||||
) # A 204 would make sense here but the spec needs 200
|
||||
if lock_id is None or instance.state_lock["ID"] == lock_id:
|
||||
# force-unlock seems to not give an ID to verify so accept no ID being present.
|
||||
# This may not be great for robustness but it's not a security issue as the protocol
|
||||
# requires returning the lock ID to clients anyway when the lock ID is not correct.
|
||||
instance.state_lock = None
|
||||
await db.commit()
|
||||
return JSONResponse(
|
||||
status_code=200, content={"result": "ok"}
|
||||
) # A 204 would make sense here but the spec needs 200
|
||||
raise TofuStateLocked(instance.state_lock)
|
||||
|
||||
|
||||
@router.delete(
|
||||
"/instances/{instance_id}/state",
|
||||
response_model=OkResponse,
|
||||
tags=["OpenTofu"],
|
||||
)
|
||||
async def purge_tofu_instance_state(
|
||||
instance_id: int,
|
||||
db: DbSession,
|
||||
password: Password,
|
||||
request: Request,
|
||||
lock_id: TofuLockID = None,
|
||||
) -> JSONResponse:
|
||||
instance = await _get_tofu_instance(db, instance_id, password, request.client.host)
|
||||
if instance.state_lock:
|
||||
if instance.state_lock["ID"] != lock_id:
|
||||
raise TofuStateLocked(instance.state_lock)
|
||||
instance.state = None
|
||||
instance.state_lock = None
|
||||
await db.commit()
|
||||
return JSONResponse(status_code=200, content={"result": "ok"})
|
||||
|
||||
|
||||
@router.get("/logs/{task_id}", tags=["OpenTofu"])
|
||||
async def get_logs(task_id: int, db: DbSession, _: APIKey) -> StreamingResponse:
|
||||
# TODO: This could definitely be optimised but I don't think it's going to see heavy usage, just for debugging
|
||||
async def stream_log() -> AsyncIterator[str]:
|
||||
start = datetime.now(tz=timezone.utc)
|
||||
seen_log_id = 0
|
||||
task_query = select(TofuInstanceTask).where(TofuInstanceTask.id == task_id)
|
||||
row = (await db.execute(task_query)).first()
|
||||
if not row:
|
||||
raise HTTPException(status_code=status.HTTP_404_NOT_FOUND, detail="Not found")
|
||||
task = row[0]
|
||||
while True:
|
||||
if datetime.now(tz=timezone.utc) - start > timedelta(minutes=5):
|
||||
return
|
||||
result = await db.execute(
|
||||
select(TofuInstanceTaskLog)
|
||||
.where(
|
||||
TofuInstanceTaskLog.instance_task_id == task_id,
|
||||
TofuInstanceTaskLog.id > seen_log_id,
|
||||
)
|
||||
.order_by(TofuInstanceTaskLog.timestamp)
|
||||
)
|
||||
for log in result.all():
|
||||
seen_log_id = log[0].id
|
||||
yield json.dumps(log[0].log) + "\n"
|
||||
await db.refresh(task)
|
||||
if row[0].status in [
|
||||
TofuInstanceTaskStatus.PENDING,
|
||||
TofuInstanceTaskStatus.FAILED,
|
||||
TofuInstanceTaskStatus.COMPLETED,
|
||||
]:
|
||||
print("done")
|
||||
return
|
||||
await asyncio.sleep(1)
|
||||
|
||||
return StreamingResponse(stream_log(), media_type="application/x-ndjson")
|
||||
76
src/tofu/schemas.py
Normal file
76
src/tofu/schemas.py
Normal file
|
|
@ -0,0 +1,76 @@
|
|||
from datetime import datetime
|
||||
from enum import Enum
|
||||
from typing import Any
|
||||
from uuid import UUID
|
||||
|
||||
from pydantic import ConfigDict
|
||||
|
||||
from src.schemas import CustomModel, TimestampMixin, CustomDatetime, IdMixin
|
||||
from src.tofu.models import TofuInstanceStatus, TofuInstanceTaskType, TofuInstanceTaskStatus
|
||||
|
||||
|
||||
class TofuOperationType(Enum):
|
||||
# https://github.com/opentofu/opentofu/blob/main/internal/backend/operation_type.go
|
||||
INVALID = "OperationTypeInvalid"
|
||||
REFRESH = "OperationTypeRefresh"
|
||||
PLAN = "OperationTypePlan"
|
||||
APPLY = "OperationTypeApply"
|
||||
|
||||
|
||||
class TofuInstanceState(CustomModel):
|
||||
# TODO: Do better
|
||||
model_config = ConfigDict(extra="allow")
|
||||
|
||||
|
||||
class TofuInstanceStateLock(CustomModel):
|
||||
model_config = ConfigDict(extra="allow")
|
||||
ID: UUID
|
||||
Operation: TofuOperationType
|
||||
Info: str
|
||||
Who: str
|
||||
Version: str
|
||||
Created: str
|
||||
Path: str
|
||||
|
||||
|
||||
class TofuInstanceSummary(CustomModel, IdMixin, TimestampMixin):
|
||||
status: TofuInstanceStatus
|
||||
status_changed_at: CustomDatetime
|
||||
drift_checked_at: CustomDatetime | None
|
||||
|
||||
|
||||
class TofuInstanceStatusChange(CustomModel):
|
||||
instance_task_id: int
|
||||
timestamp: datetime
|
||||
old_status: TofuInstanceStatus
|
||||
new_status: TofuInstanceStatus
|
||||
|
||||
|
||||
class TofuInstanceTask(CustomModel, TimestampMixin):
|
||||
id: int
|
||||
task: TofuInstanceTaskType
|
||||
status: TofuInstanceTaskStatus
|
||||
start_time: datetime | None
|
||||
end_time: datetime | None
|
||||
|
||||
|
||||
class TofuInstanceDetail(CustomModel, IdMixin, TimestampMixin):
|
||||
status: TofuInstanceStatus
|
||||
configuration: dict[str, Any]
|
||||
outputs: dict[str, Any] | None
|
||||
plan: dict[str, Any] | None
|
||||
status_changed_at: CustomDatetime
|
||||
drift_checked_at: CustomDatetime | None
|
||||
state_lock: TofuInstanceStateLock | None
|
||||
tasks: list[TofuInstanceTask]
|
||||
status_changes: list[TofuInstanceStatusChange]
|
||||
|
||||
|
||||
class TofuInstanceCreate(CustomModel):
|
||||
configuration: dict[str, Any]
|
||||
password: str | None = None
|
||||
|
||||
|
||||
class TofuInstanceUpdate(CustomModel):
|
||||
configuration: dict[str, Any] | None = None
|
||||
password: str | None = None
|
||||
93
src/tofu/security.py
Normal file
93
src/tofu/security.py
Normal file
|
|
@ -0,0 +1,93 @@
|
|||
from datetime import datetime, timezone
|
||||
from typing import Annotated
|
||||
|
||||
import bcrypt
|
||||
from fastapi import Depends
|
||||
from fastapi.security import (
|
||||
HTTPBasic,
|
||||
HTTPBasicCredentials,
|
||||
HTTPBearer,
|
||||
HTTPAuthorizationCredentials,
|
||||
)
|
||||
from sqlalchemy import select
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from starlette.requests import Request
|
||||
|
||||
from src.config import settings
|
||||
from src.database import DbSession
|
||||
from src.tofu.exceptions import TofuForbiddenException, TofuUnauthorizedException
|
||||
from src.tofu.models import TofuBruteForce
|
||||
from src.utils import generate_random_alphanum
|
||||
|
||||
basic = HTTPBasic()
|
||||
Credentials = Annotated[HTTPBasicCredentials, Depends(basic)]
|
||||
|
||||
|
||||
async def block_host_brute_force(db: AsyncSession, host: str) -> None:
|
||||
db.add(TofuBruteForce(host=host))
|
||||
await db.commit()
|
||||
|
||||
|
||||
async def is_blocked_brute_force(request: Request, db: DbSession) -> None:
|
||||
query = select(TofuBruteForce).where(TofuBruteForce.host == request.client.host)
|
||||
row = (await db.execute(query)).first()
|
||||
block: TofuBruteForce | None = row[0] if row else None
|
||||
if block:
|
||||
if block.expiry < datetime.now(tz=timezone.utc):
|
||||
db.delete(block)
|
||||
await db.commit()
|
||||
return await is_blocked_brute_force(
|
||||
request, db
|
||||
) # Just in case there's another block in the table
|
||||
raise TofuForbiddenException
|
||||
return None
|
||||
|
||||
|
||||
BruteForceProtection = Annotated[None, Depends(is_blocked_brute_force)]
|
||||
|
||||
|
||||
def generate_password() -> str:
|
||||
return generate_random_alphanum(40)
|
||||
|
||||
|
||||
def generate_password_hash(password_str: str) -> bytes:
|
||||
password = password_str.encode("utf-8")
|
||||
salt = bcrypt.gensalt()
|
||||
return bcrypt.hashpw(password, salt)
|
||||
|
||||
|
||||
def validate_password(password_str: str, password_hash: bytes) -> bool:
|
||||
if password_hash is None:
|
||||
return False
|
||||
password = password_str.encode("utf-8")
|
||||
return bcrypt.checkpw(password, password_hash)
|
||||
|
||||
|
||||
def get_password(
|
||||
db: DbSession, credentials: Credentials, request: Request, _: BruteForceProtection
|
||||
) -> str:
|
||||
if not credentials or not credentials.password:
|
||||
block_host_brute_force(db, request.client.host)
|
||||
raise TofuUnauthorizedException
|
||||
return credentials.password
|
||||
|
||||
|
||||
Password = Annotated[str, Depends(get_password)]
|
||||
|
||||
bearer = HTTPBearer()
|
||||
|
||||
|
||||
async def api_key(
|
||||
db: DbSession,
|
||||
request: Request,
|
||||
_: BruteForceProtection,
|
||||
credentials: HTTPAuthorizationCredentials = Depends(bearer),
|
||||
) -> None:
|
||||
api_key = credentials.credentials
|
||||
print(api_key)
|
||||
if api_key != settings.API_KEY:
|
||||
await block_host_brute_force(db, request.client.host)
|
||||
raise TofuUnauthorizedException
|
||||
|
||||
|
||||
APIKey = Annotated[None, Depends(api_key)]
|
||||
80
src/tofu/service.py
Normal file
80
src/tofu/service.py
Normal file
|
|
@ -0,0 +1,80 @@
|
|||
from typing import Any
|
||||
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from starlette.background import BackgroundTask
|
||||
from starlette.exceptions import HTTPException
|
||||
|
||||
from src.tofu.models import (
|
||||
TofuInstance,
|
||||
TofuInstanceTask,
|
||||
TofuInstanceTaskType,
|
||||
TofuInstanceStatus,
|
||||
update_tofu_instance_status,
|
||||
)
|
||||
from src.tofu.tasks import process_tasks
|
||||
|
||||
|
||||
async def create_tofu_instance(
|
||||
configuration: dict[str, Any],
|
||||
db: AsyncSession,
|
||||
*,
|
||||
commit: bool = False,
|
||||
create_task: bool = True
|
||||
) -> tuple[int, BackgroundTask | None]:
|
||||
instance = TofuInstance(configuration=configuration)
|
||||
db.add(instance)
|
||||
await db.flush()
|
||||
if create_task:
|
||||
task = TofuInstanceTask(instance_id=instance.id, task=TofuInstanceTaskType.DEPLOY)
|
||||
db.add(task)
|
||||
if commit:
|
||||
await db.commit()
|
||||
return instance.id, BackgroundTask(process_tasks) if create_task else None
|
||||
|
||||
|
||||
async def create_empty_tofu_instance(db: AsyncSession) -> int:
|
||||
config = {}
|
||||
tofu_instance_id, _ = await create_tofu_instance(config, db, create_task=False)
|
||||
return tofu_instance_id
|
||||
|
||||
|
||||
async def update_tofu_instance(
|
||||
instance: TofuInstance,
|
||||
configuration: dict[str, Any],
|
||||
db: AsyncSession,
|
||||
*,
|
||||
commit: bool = False,
|
||||
allow_pending: bool = False
|
||||
) -> BackgroundTask:
|
||||
allowed_status = [TofuInstanceStatus.ACTIVE, TofuInstanceStatus.DRIFTED]
|
||||
if allow_pending:
|
||||
allowed_status.append(TofuInstanceStatus.PENDING)
|
||||
if instance.status not in allowed_status:
|
||||
raise HTTPException(
|
||||
status_code=412, detail="Updates only allowed for active instances"
|
||||
)
|
||||
instance.configuration = configuration
|
||||
task = TofuInstanceTask(instance_id=instance.id, task=TofuInstanceTaskType.DEPLOY)
|
||||
db.add(task)
|
||||
await db.flush()
|
||||
update_tofu_instance_status(db, instance, task.id, TofuInstanceStatus.PENDING)
|
||||
if commit:
|
||||
await db.commit()
|
||||
return BackgroundTask(process_tasks)
|
||||
|
||||
|
||||
async def destroy_tofu_instance(
|
||||
instance: TofuInstance, db: AsyncSession, *, commit: bool = False
|
||||
) -> BackgroundTask:
|
||||
if instance.status not in [TofuInstanceStatus.ACTIVE, TofuInstanceStatus.DRIFTED]:
|
||||
raise HTTPException(
|
||||
status_code=412,
|
||||
detail="Instance cannot be destroyed currently as it is pending update",
|
||||
)
|
||||
task = TofuInstanceTask(instance_id=instance.id, task=TofuInstanceTaskType.DESTROY)
|
||||
db.add(task)
|
||||
await db.flush()
|
||||
update_tofu_instance_status(db, instance, task.id, TofuInstanceStatus.PENDING_DESTROY)
|
||||
if commit:
|
||||
await db.commit()
|
||||
return BackgroundTask(process_tasks)
|
||||
86
src/tofu/tasks.py
Normal file
86
src/tofu/tasks.py
Normal file
|
|
@ -0,0 +1,86 @@
|
|||
from datetime import datetime, timezone
|
||||
|
||||
from sqlalchemy import select
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
from sqlalchemy.orm import selectinload
|
||||
|
||||
from src.database import get_db_session
|
||||
from src.tofu.client import TofuManager
|
||||
from src.tofu.exceptions import TofuExecutionError
|
||||
from src.tofu.models import (
|
||||
TofuInstanceStatus,
|
||||
TofuInstanceTask,
|
||||
TofuInstanceTaskType,
|
||||
TofuInstanceTaskStatus,
|
||||
update_tofu_instance_status,
|
||||
)
|
||||
from src.utils import repeat_every
|
||||
|
||||
|
||||
async def deploy_tofu_instance(db: AsyncSession, task: TofuInstanceTask) -> None:
|
||||
try:
|
||||
async with TofuManager(db, task) as tofu:
|
||||
await tofu.init()
|
||||
await tofu.apply()
|
||||
await tofu.output()
|
||||
update_tofu_instance_status(db, task.instance, task.id, TofuInstanceStatus.ACTIVE)
|
||||
task.instance.drift_checked_at = task.end_time = datetime.now(tz=timezone.utc)
|
||||
except TofuExecutionError as exc:
|
||||
update_tofu_instance_status(db, task.instance, task.id, TofuInstanceStatus.FAILED)
|
||||
raise exc
|
||||
|
||||
|
||||
async def destroy_tofu_instance(db: AsyncSession, task: TofuInstanceTask) -> None:
|
||||
try:
|
||||
async with TofuManager(db, task) as tofu:
|
||||
await tofu.init()
|
||||
await tofu.destroy()
|
||||
update_tofu_instance_status(
|
||||
db, task.instance, task.id, TofuInstanceStatus.DESTROYED
|
||||
)
|
||||
task.instance.outputs = {}
|
||||
task.instance.drift_checked_at = task.end_time = datetime.now(tz=timezone.utc)
|
||||
except TofuExecutionError as exc:
|
||||
update_tofu_instance_status(
|
||||
db, task.instance, task.id, TofuInstanceStatus.FAILED_DESTROY
|
||||
)
|
||||
raise exc
|
||||
|
||||
|
||||
async def process_tasks() -> None:
|
||||
async with get_db_session() as db:
|
||||
while True:
|
||||
query = (
|
||||
select(TofuInstanceTask)
|
||||
.with_for_update(skip_locked=True)
|
||||
.options(selectinload(TofuInstanceTask.instance))
|
||||
.filter(TofuInstanceTask.status == TofuInstanceTaskStatus.PENDING)
|
||||
.limit(1)
|
||||
)
|
||||
row = (await db.execute(query)).first()
|
||||
if not row:
|
||||
break
|
||||
task = row[0]
|
||||
task.status = TofuInstanceTaskStatus.RUNNING
|
||||
task.start_time = datetime.now(tz=timezone.utc)
|
||||
await db.commit() # release FOR UPDATE lock
|
||||
try:
|
||||
match task.task:
|
||||
case TofuInstanceTaskType.DEPLOY:
|
||||
await deploy_tofu_instance(db, task)
|
||||
case TofuInstanceTaskType.DESTROY:
|
||||
await destroy_tofu_instance(db, task)
|
||||
case TofuInstanceTaskType.CHECK_DRIFT:
|
||||
print("NOT IMPLEMMENTED") # TODO: Implement
|
||||
task.status = TofuInstanceTaskStatus.COMPLETED
|
||||
task.end_time = datetime.now(tz=timezone.utc)
|
||||
except TofuExecutionError:
|
||||
task.status = TofuInstanceTaskStatus.FAILED
|
||||
task.end_time = datetime.now(tz=timezone.utc)
|
||||
finally:
|
||||
await db.commit()
|
||||
|
||||
|
||||
@repeat_every(seconds=5, wait_first=0)
|
||||
async def periodic_tofu_process_tasks():
|
||||
await process_tasks()
|
||||
199
src/utils.py
Normal file
199
src/utils.py
Normal file
|
|
@ -0,0 +1,199 @@
|
|||
import asyncio
|
||||
import logging
|
||||
import random
|
||||
import string
|
||||
from datetime import datetime, timedelta, timezone
|
||||
from functools import wraps
|
||||
from traceback import format_exception
|
||||
from typing import Callable, Coroutine, Any
|
||||
|
||||
from starlette.concurrency import run_in_threadpool
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
ALPHA_NUM = string.ascii_letters + string.digits
|
||||
LOWER_NUM = string.ascii_lowercase + string.digits
|
||||
|
||||
TOP_SUBDOMAINS = [
|
||||
"admin",
|
||||
"api",
|
||||
"app",
|
||||
"apps",
|
||||
"autodiscover",
|
||||
"baidu",
|
||||
"bbs",
|
||||
"beta",
|
||||
"blog",
|
||||
"cdn",
|
||||
"citrix",
|
||||
"cloud",
|
||||
"demo",
|
||||
"dev",
|
||||
"email",
|
||||
"en",
|
||||
"exchange",
|
||||
"forum",
|
||||
"ftp",
|
||||
"gateway",
|
||||
"gov",
|
||||
"gw",
|
||||
"home",
|
||||
"host",
|
||||
"images",
|
||||
"img",
|
||||
"info",
|
||||
"intranet",
|
||||
"login",
|
||||
"m",
|
||||
"mail",
|
||||
"mail1",
|
||||
"mail2",
|
||||
"mail3",
|
||||
"media",
|
||||
"mobile",
|
||||
"news",
|
||||
"office",
|
||||
"owa",
|
||||
"portal",
|
||||
"remote",
|
||||
"secure",
|
||||
"server",
|
||||
"server1",
|
||||
"shop",
|
||||
"ssl",
|
||||
"stage",
|
||||
"staging",
|
||||
"start",
|
||||
"static",
|
||||
"store",
|
||||
"support",
|
||||
"test",
|
||||
"web",
|
||||
"webmail",
|
||||
"wiki",
|
||||
"www1",
|
||||
"www2",
|
||||
]
|
||||
|
||||
NoArgsNoReturnFuncT = Callable[[], None]
|
||||
NoArgsNoReturnAsyncFuncT = Callable[[], Coroutine[Any, Any, None]]
|
||||
ExcArgNoReturnFuncT = Callable[[Exception], None]
|
||||
ExcArgNoReturnAsyncFuncT = Callable[[Exception], Coroutine[Any, Any, None]]
|
||||
NoArgsNoReturnAnyFuncT = NoArgsNoReturnFuncT | NoArgsNoReturnAsyncFuncT
|
||||
ExcArgNoReturnAnyFuncT = ExcArgNoReturnFuncT | ExcArgNoReturnAsyncFuncT
|
||||
NoArgsNoReturnDecorator = Callable[[NoArgsNoReturnAnyFuncT], NoArgsNoReturnAsyncFuncT]
|
||||
|
||||
|
||||
async def _handle_repeat_func(func: NoArgsNoReturnAnyFuncT) -> None:
|
||||
if asyncio.iscoroutinefunction(func):
|
||||
await func()
|
||||
else:
|
||||
await run_in_threadpool(func)
|
||||
|
||||
|
||||
async def _handle_repeat_exc(
|
||||
exc: Exception, on_exception: ExcArgNoReturnAnyFuncT | None
|
||||
) -> None:
|
||||
if on_exception:
|
||||
if asyncio.iscoroutinefunction(on_exception):
|
||||
await on_exception(exc)
|
||||
else:
|
||||
await run_in_threadpool(on_exception, exc)
|
||||
|
||||
|
||||
def repeat_every(
|
||||
*,
|
||||
seconds: float,
|
||||
wait_first: float | None = None,
|
||||
max_repetitions: int | None = None,
|
||||
on_complete: NoArgsNoReturnAnyFuncT | None = None,
|
||||
on_exception: ExcArgNoReturnAnyFuncT | None = None,
|
||||
) -> NoArgsNoReturnDecorator:
|
||||
"""
|
||||
This function returns a decorator that modifies a function so it is periodically re-executed after its first call.
|
||||
|
||||
The function it decorates should accept no arguments and return nothing. If necessary, this can be accomplished
|
||||
by using `functools.partial` or otherwise wrapping the target function prior to decoration.
|
||||
|
||||
Parameters
|
||||
----------
|
||||
seconds: float
|
||||
The number of seconds to wait between repeated calls
|
||||
wait_first: float (default None)
|
||||
If not None, the function will wait for the given duration before the first call
|
||||
max_repetitions: Optional[int] (default None)
|
||||
The maximum number of times to call the repeated function. If `None`, the function is repeated forever.
|
||||
on_complete: Optional[Callable[[], None]] (default None)
|
||||
A function to call after the final repetition of the decorated function.
|
||||
on_exception: Optional[Callable[[Exception], None]] (default None)
|
||||
A function to call when an exception is raised by the decorated function.
|
||||
"""
|
||||
|
||||
def decorator(func: NoArgsNoReturnAnyFuncT) -> NoArgsNoReturnAsyncFuncT:
|
||||
"""
|
||||
Converts the decorated function into a repeated, periodically-called version of itself.
|
||||
"""
|
||||
|
||||
@wraps(func)
|
||||
async def wrapped() -> None:
|
||||
async def loop() -> None:
|
||||
if wait_first is not None:
|
||||
await asyncio.sleep(wait_first)
|
||||
|
||||
repetitions = 0
|
||||
while max_repetitions is None or repetitions < max_repetitions:
|
||||
try:
|
||||
await _handle_repeat_func(func)
|
||||
|
||||
except Exception as exc:
|
||||
formatted_exception = "".join(
|
||||
format_exception(type(exc), exc, exc.__traceback__)
|
||||
)
|
||||
logger.error(formatted_exception)
|
||||
await _handle_repeat_exc(exc, on_exception)
|
||||
|
||||
repetitions += 1
|
||||
await asyncio.sleep(seconds)
|
||||
|
||||
if on_complete:
|
||||
await _handle_repeat_func(on_complete)
|
||||
|
||||
asyncio.ensure_future(loop())
|
||||
|
||||
return wrapped
|
||||
|
||||
return decorator
|
||||
|
||||
|
||||
def perishable_cache(expires: int, minimum: int):
|
||||
def decorator(func):
|
||||
cache_data = {}
|
||||
|
||||
@wraps(func)
|
||||
async def wrapper(*args, retry: bool = False, **kwargs):
|
||||
timeout = minimum if retry else expires
|
||||
if "cached_time" in cache_data:
|
||||
if datetime.now(tz=timezone.utc) - cache_data["cached_time"] < timedelta(
|
||||
seconds=timeout
|
||||
):
|
||||
return cache_data["cached_result"]
|
||||
result = await func(*args, retry=retry, **kwargs)
|
||||
cache_data["cached_result"] = result
|
||||
cache_data["cached_time"] = datetime.now(tz=timezone.utc)
|
||||
return result
|
||||
|
||||
return wrapper
|
||||
|
||||
return decorator
|
||||
|
||||
|
||||
def generate_random_alphanum(length: int = 20, prefix: str = "") -> str:
|
||||
return prefix + "".join(random.choices(ALPHA_NUM, k=length - len(prefix)))
|
||||
|
||||
|
||||
def generate_random_lowernum(length: int = 20, prefix: str = "") -> str:
|
||||
return prefix + "".join(random.choices(LOWER_NUM, k=length - len(prefix)))
|
||||
|
||||
|
||||
def generate_random_subdomain() -> str:
|
||||
return random.choice(TOP_SUBDOMAINS)
|
||||
136
tests/tofu/test_tofu_state.py
Normal file
136
tests/tofu/test_tofu_state.py
Normal file
|
|
@ -0,0 +1,136 @@
|
|||
from typing import AsyncGenerator
|
||||
|
||||
import pytest
|
||||
from httpx import AsyncClient, ASGITransport
|
||||
|
||||
from src.auth.service import get_admin
|
||||
from src.main import app
|
||||
|
||||
PASSWORD="password123"
|
||||
|
||||
AUTH=("tofu", PASSWORD)
|
||||
|
||||
CREATE_INSTANCE_PAYLOAD = {
|
||||
"suppress_deployment": True,
|
||||
"password": PASSWORD,
|
||||
"configuration": {
|
||||
"terraform": {"required_providers": {"random": {"source": "hashicorp/random", "version": ">= 3.0.0"}}},
|
||||
"provider": {"random": {}},
|
||||
"variable": {
|
||||
"password_length": {"description": "Length of the random password", "type": "number", "default": 16}
|
||||
},
|
||||
"resource": {"random_password": {"example": {"length": "${var.password_length}", "special": True}}},
|
||||
"output": {"generated_password": {"value": "${random_password.example.result}", "sensitive": True}},
|
||||
},
|
||||
}
|
||||
|
||||
INITIAL_STATE_PAYLOAD = {"key": "value"}
|
||||
|
||||
UPDATE_STATE_PAYLOAD = {"key": "value"}
|
||||
|
||||
STATE_LOCK_PAYLOAD_1 = {"ID": "bd812a7e-2297-4b70-acc9-d51015a9172c",
|
||||
"Operation": "OperationTypeInvalid",
|
||||
"Info": "",
|
||||
"Who": "",
|
||||
"Version": "",
|
||||
"Created": "1990-01-01T12:00:00Z",
|
||||
"Path": ""
|
||||
}
|
||||
|
||||
STATE_LOCK_PAYLOAD_2 = {"ID": "ab0eb55f-2f00-4e02-9bf5-e2c6658ab8af",
|
||||
"Operation": "OperationTypeInvalid",
|
||||
"Info": "",
|
||||
"Who": "",
|
||||
"Version": "",
|
||||
"Created": "1990-01-01T12:00:00Z",
|
||||
"Path": ""
|
||||
}
|
||||
|
||||
|
||||
@pytest.fixture(scope="module")
|
||||
async def client() -> AsyncGenerator[AsyncClient, None]:
|
||||
host, port = "127.0.0.1", "9000"
|
||||
|
||||
async def override_get_admin():
|
||||
return None
|
||||
|
||||
app.dependency_overrides[get_admin] = override_get_admin
|
||||
async with AsyncClient(transport=ASGITransport(app=app, client=(host, port)), base_url="http://test", auth=AUTH) as client:
|
||||
yield client
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
async def instance_id(client: AsyncClient):
|
||||
response = await client.post("/api/v1/tofu/instances", json=CREATE_INSTANCE_PAYLOAD)
|
||||
assert response.status_code == 202
|
||||
return response.json()["id"]
|
||||
|
||||
|
||||
@pytest.mark.anyio
|
||||
async def test_state_no_locking(client: AsyncClient, instance_id: int):
|
||||
# Initially there should be no state
|
||||
response = await client.get(f"/api/v1/tofu/instances/{instance_id}/state")
|
||||
assert response.status_code == 404
|
||||
# Let's create the state
|
||||
response = await client.post(f"/api/v1/tofu/instances/{instance_id}/state", json=INITIAL_STATE_PAYLOAD)
|
||||
assert response.status_code == 200
|
||||
# Now check the state is retrievable
|
||||
response = await client.get(f"/api/v1/tofu/instances/{instance_id}/state")
|
||||
assert response.status_code == 200
|
||||
assert response.json() == INITIAL_STATE_PAYLOAD
|
||||
# Now update the state
|
||||
response = await client.post(f"/api/v1/tofu/instances/{instance_id}/state", json=UPDATE_STATE_PAYLOAD)
|
||||
assert response.status_code == 200
|
||||
# Now check the state is retrievable
|
||||
response = await client.get(f"/api/v1/tofu/instances/{instance_id}/state")
|
||||
assert response.status_code == 200
|
||||
assert response.json() == UPDATE_STATE_PAYLOAD
|
||||
# Now purge the state
|
||||
response = await client.delete(f"/api/v1/tofu/instances/{instance_id}/state")
|
||||
assert response.status_code == 200
|
||||
# And check it is gone
|
||||
response = await client.get(f"/api/v1/tofu/instances/{instance_id}/state")
|
||||
assert response.status_code == 404
|
||||
|
||||
@pytest.mark.anyio
|
||||
async def test_state_double_locking(client: AsyncClient, instance_id: int):
|
||||
response = await client.request("LOCK", f"/api/v1/tofu/instances/{instance_id}/state", json=STATE_LOCK_PAYLOAD_1)
|
||||
assert response.status_code == 200
|
||||
response = await client.request("LOCK", f"/api/v1/tofu/instances/{instance_id}/state", json=STATE_LOCK_PAYLOAD_2)
|
||||
assert response.status_code == 423
|
||||
assert response.json() == STATE_LOCK_PAYLOAD_1
|
||||
response = await client.request("UNLOCK",
|
||||
f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_2['ID'])
|
||||
assert response.status_code == 423
|
||||
assert response.json() == STATE_LOCK_PAYLOAD_1
|
||||
response = await client.request("UNLOCK",
|
||||
f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_1['ID'])
|
||||
assert response.status_code == 200
|
||||
|
||||
@pytest.mark.anyio
|
||||
async def test_state_locked_update(client: AsyncClient, instance_id: int):
|
||||
response = await client.request("LOCK", f"/api/v1/tofu/instances/{instance_id}/state", json=STATE_LOCK_PAYLOAD_1)
|
||||
assert response.status_code == 200
|
||||
response = await client.post(f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_1['ID'],
|
||||
json=INITIAL_STATE_PAYLOAD)
|
||||
assert response.status_code == 200
|
||||
response = await client.post(f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_2['ID'],
|
||||
json=UPDATE_STATE_PAYLOAD)
|
||||
assert response.status_code == 423
|
||||
assert response.json() == STATE_LOCK_PAYLOAD_1
|
||||
response = await client.post(f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_1['ID'],
|
||||
json=UPDATE_STATE_PAYLOAD)
|
||||
assert response.status_code == 200
|
||||
response = await client.delete(f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_2['ID'])
|
||||
assert response.status_code == 423
|
||||
assert response.json() == STATE_LOCK_PAYLOAD_1
|
||||
response = await client.delete(f"/api/v1/tofu/instances/{instance_id}/state?ID=" + STATE_LOCK_PAYLOAD_1['ID'])
|
||||
assert response.status_code == 200
|
||||
|
||||
@pytest.mark.anyio
|
||||
async def test_state_allow_force_unlock(client: AsyncClient, instance_id: int):
|
||||
# force-unlock doesn't include the ID when calling the unlock endpoint
|
||||
response = await client.request("LOCK", f"/api/v1/tofu/instances/{instance_id}/state", json=STATE_LOCK_PAYLOAD_1)
|
||||
assert response.status_code == 200
|
||||
response = await client.request("UNLOCK", f"/api/v1/tofu/instances/{instance_id}/state", json=STATE_LOCK_PAYLOAD_1)
|
||||
assert response.status_code == 200
|
||||
Loading…
Add table
Add a link
Reference in a new issue