mirror of
https://github.com/GNS3/gns3-server
synced 2025-01-13 09:30:54 +00:00
4758431c76
* Checksum is updated in the database for updated disks. * It is not possible to prune them.
272 lines
9.9 KiB
Python
272 lines
9.9 KiB
Python
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2020 GNS3 Technologies Inc.
|
|
#
|
|
# This program is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
import asyncio
|
|
import time
|
|
import os
|
|
|
|
from fastapi import FastAPI
|
|
from pydantic import ValidationError
|
|
from typing import List
|
|
from sqlalchemy import event
|
|
from sqlalchemy.engine import Engine
|
|
from sqlalchemy.exc import SQLAlchemyError
|
|
from sqlalchemy.ext.asyncio import create_async_engine, AsyncSession
|
|
from alembic import command, config
|
|
from alembic.script import ScriptDirectory
|
|
from alembic.runtime.migration import MigrationContext
|
|
from alembic.util.exc import CommandError
|
|
from watchdog.observers import Observer
|
|
from watchdog.events import FileSystemEvent, PatternMatchingEventHandler
|
|
|
|
from gns3server.db.repositories.computes import ComputesRepository
|
|
from gns3server.db.repositories.images import ImagesRepository
|
|
from gns3server.utils.images import md5sum, discover_images, read_image_info, InvalidImageError
|
|
from gns3server.utils.asyncio import wait_run_in_executor
|
|
from gns3server import schemas
|
|
|
|
from .models import Base
|
|
from gns3server.config import Config
|
|
|
|
import logging
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
def run_upgrade(connection, cfg):
|
|
|
|
cfg.attributes["connection"] = connection
|
|
try:
|
|
command.upgrade(cfg, "head")
|
|
except CommandError as e:
|
|
log.error(f"Could not upgrade database: {e}")
|
|
|
|
|
|
def run_stamp(connection, cfg):
|
|
|
|
cfg.attributes["connection"] = connection
|
|
try:
|
|
command.stamp(cfg, "head")
|
|
except CommandError as e:
|
|
log.error(f"Could not stamp database: {e}")
|
|
|
|
|
|
def check_revision(connection, cfg):
|
|
|
|
script = ScriptDirectory.from_config(cfg)
|
|
head_rev = script.get_revision("head").revision
|
|
context = MigrationContext.configure(connection)
|
|
current_rev = context.get_current_revision()
|
|
return current_rev, head_rev
|
|
|
|
|
|
async def connect_to_db(app: FastAPI) -> None:
|
|
|
|
db_path = os.path.join(Config.instance().config_dir, "gns3_controller.db")
|
|
db_url = os.environ.get("GNS3_DATABASE_URI", f"sqlite+aiosqlite:///{db_path}")
|
|
engine = create_async_engine(db_url, connect_args={"check_same_thread": False, "timeout": 20}, future=True)
|
|
alembic_cfg = config.Config()
|
|
alembic_cfg.set_main_option("script_location", "gns3server:db_migrations")
|
|
#alembic_cfg.set_main_option('sqlalchemy.url', db_url)
|
|
try:
|
|
async with engine.connect() as conn:
|
|
current_rev, head_rev = await conn.run_sync(check_revision, alembic_cfg)
|
|
log.info(f"Current database revision is {current_rev}")
|
|
if current_rev is None:
|
|
await conn.run_sync(Base.metadata.create_all)
|
|
await conn.run_sync(run_stamp, alembic_cfg)
|
|
elif current_rev != head_rev:
|
|
# upgrade the database if needed
|
|
await conn.run_sync(run_upgrade, alembic_cfg)
|
|
await conn.commit()
|
|
app.state._db_engine = engine
|
|
except SQLAlchemyError as e:
|
|
log.fatal(f"Error while connecting to database '{db_url}: {e}")
|
|
|
|
|
|
async def disconnect_from_db(app: FastAPI) -> None:
|
|
|
|
# dispose of the connection pool used by the database engine
|
|
if getattr(app.state, "_db_engine"):
|
|
await app.state._db_engine.dispose()
|
|
log.info(f"Disconnected from database")
|
|
|
|
|
|
@event.listens_for(Engine, "connect")
|
|
def set_sqlite_pragma(dbapi_connection, connection_record):
|
|
|
|
# Enable SQL foreign key support for SQLite
|
|
# https://docs.sqlalchemy.org/en/14/dialects/sqlite.html#foreign-key-support
|
|
cursor = dbapi_connection.cursor()
|
|
cursor.execute("PRAGMA foreign_keys=ON")
|
|
cursor.close()
|
|
|
|
|
|
async def get_computes(app: FastAPI) -> List[dict]:
|
|
|
|
computes = []
|
|
async with AsyncSession(app.state._db_engine) as db_session:
|
|
db_computes = await ComputesRepository(db_session).get_computes()
|
|
for db_compute in db_computes:
|
|
try:
|
|
compute = schemas.Compute.model_validate(db_compute)
|
|
except ValidationError as e:
|
|
log.error(f"Could not load compute '{db_compute.compute_id}' from database: {e}")
|
|
continue
|
|
computes.append(compute)
|
|
return computes
|
|
|
|
|
|
async def discover_images_on_filesystem(app: FastAPI) -> None:
|
|
|
|
async with AsyncSession(app.state._db_engine) as db_session:
|
|
images_repository = ImagesRepository(db_session)
|
|
db_images = await images_repository.get_images()
|
|
existing_image_paths = []
|
|
for db_image in db_images:
|
|
try:
|
|
image = schemas.Image.model_validate(db_image)
|
|
existing_image_paths.append(image.path)
|
|
except ValidationError as e:
|
|
log.error(f"Could not load image '{db_image.filename}' from database: {e}")
|
|
continue
|
|
for image_type in ("qemu", "ios", "iou"):
|
|
discovered_images = await discover_images(image_type, existing_image_paths)
|
|
for image in discovered_images:
|
|
log.info(f"Adding discovered image '{image['path']}' to the database")
|
|
try:
|
|
await images_repository.add_image(**image)
|
|
except SQLAlchemyError as e:
|
|
log.warning(f"Error while adding image '{image['path']}' to the database: {e}")
|
|
|
|
# monitor if images have been manually added
|
|
asyncio.create_task(monitor_images_on_filesystem(app))
|
|
|
|
|
|
async def update_disk_checksums(updated_disks: List[str]) -> None:
|
|
"""
|
|
Update the checksum of a list of disks in the database.
|
|
|
|
:param updated_disks: list of updated disks
|
|
"""
|
|
|
|
from gns3server.api.server import app
|
|
async with AsyncSession(app.state._db_engine) as db_session:
|
|
images_repository = ImagesRepository(db_session)
|
|
for path in updated_disks:
|
|
image = await images_repository.get_image(path)
|
|
if image:
|
|
log.info(f"Updating image '{path}' in the database")
|
|
checksum = await wait_run_in_executor(md5sum, path, cache_to_md5file=False)
|
|
if image.checksum != checksum:
|
|
await images_repository.update_image(path, checksum, "md5")
|
|
|
|
class EventHandler(PatternMatchingEventHandler):
|
|
"""
|
|
Watchdog event handler.
|
|
"""
|
|
|
|
def __init__(self, queue: asyncio.Queue, loop: asyncio.BaseEventLoop, **kwargs):
|
|
|
|
self._loop = loop
|
|
self._queue = queue
|
|
|
|
# ignore temporary files, md5sum files, hidden files and directories
|
|
super().__init__(ignore_patterns=["*.tmp", "*.md5sum", ".*"], ignore_directories = True, **kwargs)
|
|
|
|
def on_closed(self, event: FileSystemEvent) -> None:
|
|
# monitor for closed files (e.g. when a file has finished to be copied)
|
|
if "/lib/" in event.src_path or "/lib64/" in event.src_path:
|
|
return # ignore custom IOU libraries
|
|
self._loop.call_soon_threadsafe(self._queue.put_nowait, event)
|
|
|
|
class EventIterator(object):
|
|
"""
|
|
Watchdog Event iterator.
|
|
"""
|
|
|
|
def __init__(self, queue: asyncio.Queue):
|
|
self.queue = queue
|
|
|
|
def __aiter__(self):
|
|
return self
|
|
|
|
async def __anext__(self):
|
|
|
|
item = await self.queue.get()
|
|
if item is None:
|
|
raise StopAsyncIteration
|
|
return item
|
|
|
|
async def monitor_images_on_filesystem(app: FastAPI):
|
|
|
|
def watchdog(
|
|
path: str,
|
|
queue: asyncio.Queue,
|
|
loop: asyncio.BaseEventLoop,
|
|
app: FastAPI, recursive: bool = False
|
|
) -> None:
|
|
"""
|
|
Thread to monitor a directory for new images.
|
|
"""
|
|
|
|
handler = EventHandler(queue, loop)
|
|
observer = Observer()
|
|
observer.schedule(handler, str(path), recursive=recursive)
|
|
observer.start()
|
|
log.info(f"Monitoring for new images in '{path}'")
|
|
while True:
|
|
time.sleep(1)
|
|
# stop when the app is exiting
|
|
if app.state.exiting:
|
|
observer.stop()
|
|
observer.join(10)
|
|
log.info(f"Stopping monitoring for new images in '{path}'")
|
|
loop.call_soon_threadsafe(queue.put_nowait, None)
|
|
break
|
|
|
|
queue = asyncio.Queue()
|
|
loop = asyncio.get_event_loop()
|
|
server_config = Config.instance().settings.Server
|
|
image_dir = os.path.expanduser(server_config.images_path)
|
|
asyncio.get_event_loop().run_in_executor(None, watchdog,image_dir, queue, loop, app, True)
|
|
|
|
async for filesystem_event in EventIterator(queue):
|
|
# read the file system event from the queue
|
|
image_path = filesystem_event.src_path
|
|
expected_image_type = None
|
|
if "IOU" in image_path:
|
|
expected_image_type = "iou"
|
|
elif "QEMU" in image_path:
|
|
expected_image_type = "qemu"
|
|
elif "IOS" in image_path:
|
|
expected_image_type = "ios"
|
|
async with AsyncSession(app.state._db_engine) as db_session:
|
|
images_repository = ImagesRepository(db_session)
|
|
try:
|
|
image = await read_image_info(image_path, expected_image_type)
|
|
except InvalidImageError as e:
|
|
log.warning(str(e))
|
|
continue
|
|
try:
|
|
if await images_repository.get_image(image_path):
|
|
continue
|
|
await images_repository.add_image(**image)
|
|
log.info(f"Discovered image '{image_path}' has been added to the database")
|
|
except SQLAlchemyError as e:
|
|
log.warning(f"Error while adding image '{image_path}' to the database: {e}")
|