mirror of
https://github.com/GNS3/gns3-server
synced 2024-11-12 03:18:55 +00:00
640 lines
21 KiB
Python
640 lines
21 KiB
Python
# -*- coding: utf-8 -*-
|
||
#
|
||
# Copyright (C) 2015 GNS3 Technologies Inc.
|
||
#
|
||
# This program is free software: you can redistribute it and/or modify
|
||
# it under the terms of the GNU General Public License as published by
|
||
# the Free Software Foundation, either version 3 of the License, or
|
||
# (at your option) any later version.
|
||
#
|
||
# This program is distributed in the hope that it will be useful,
|
||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
# GNU General Public License for more details.
|
||
#
|
||
# You should have received a copy of the GNU General Public License
|
||
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||
|
||
import os
|
||
import aiohttp
|
||
import shutil
|
||
import asyncio
|
||
import hashlib
|
||
import zipstream
|
||
import zipfile
|
||
import json
|
||
|
||
from uuid import UUID
|
||
from .port_manager import PortManager
|
||
from .notification_manager import NotificationManager
|
||
from ..config import Config
|
||
from ..utils.asyncio import wait_run_in_executor
|
||
from ..utils.path import check_path_allowed, get_default_project_directory
|
||
|
||
|
||
import logging
|
||
log = logging.getLogger(__name__)
|
||
|
||
|
||
class Project:
|
||
|
||
"""
|
||
A project contains a list of nodes.
|
||
In theory nodes are isolated project/project.
|
||
|
||
:param project_id: force project identifier (None by default auto generate an UUID)
|
||
:param path: path of the project. (None use the standard directory)
|
||
:param temporary: boolean to tell if the project is a temporary project (destroy when closed)
|
||
"""
|
||
|
||
def __init__(self, name=None, project_id=None, path=None, temporary=False):
|
||
|
||
self._name = name
|
||
try:
|
||
UUID(project_id, version=4)
|
||
except ValueError:
|
||
raise aiohttp.web.HTTPBadRequest(text="{} is not a valid UUID".format(project_id))
|
||
self._id = project_id
|
||
|
||
self._nodes = set()
|
||
self._nodes_to_destroy = set()
|
||
self.temporary = temporary
|
||
self._used_tcp_ports = set()
|
||
self._used_udp_ports = set()
|
||
|
||
if path is None:
|
||
location = get_default_project_directory()
|
||
path = os.path.join(location, self._id)
|
||
try:
|
||
os.makedirs(path, exist_ok=True)
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not create project directory: {}".format(e))
|
||
self.path = path
|
||
|
||
try:
|
||
if os.path.exists(self.tmp_working_directory()):
|
||
shutil.rmtree(self.tmp_working_directory())
|
||
except OSError:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not clean project directory: {}".format(e))
|
||
|
||
log.info("Project {id} with path '{path}' created".format(path=self._path, id=self._id))
|
||
|
||
def __json__(self):
|
||
|
||
return {
|
||
"name": self._name,
|
||
"project_id": self._id,
|
||
"temporary": self._temporary
|
||
}
|
||
|
||
def _config(self):
|
||
|
||
return Config.instance().get_section_config("Server")
|
||
|
||
def is_local(self):
|
||
|
||
return self._config().getboolean("local", False)
|
||
|
||
@property
|
||
def id(self):
|
||
|
||
return self._id
|
||
|
||
@property
|
||
def path(self):
|
||
|
||
return self._path
|
||
|
||
@path.setter
|
||
def path(self, path):
|
||
check_path_allowed(path)
|
||
|
||
if hasattr(self, "_path"):
|
||
if path != self._path and self.is_local() is False:
|
||
raise aiohttp.web.HTTPForbidden(text="You are not allowed to modify the project directory path")
|
||
|
||
self._path = path
|
||
self._update_temporary_file()
|
||
|
||
@asyncio.coroutine
|
||
def clean_old_path(self, old_path):
|
||
"""
|
||
Called after a project location change. All the compute should
|
||
have been notified before
|
||
"""
|
||
if self._temporary:
|
||
try:
|
||
yield from wait_run_in_executor(shutil.rmtree, old_path)
|
||
except OSError as e:
|
||
log.warn("Can't remove temporary directory {}: {}".format(old_path, e))
|
||
|
||
@property
|
||
def name(self):
|
||
|
||
return self._name
|
||
|
||
@name.setter
|
||
def name(self, name):
|
||
|
||
if "/" in name or "\\" in name:
|
||
raise aiohttp.web.HTTPForbidden(text="Name can not contain path separator")
|
||
self._name = name
|
||
|
||
@property
|
||
def nodes(self):
|
||
|
||
return self._nodes
|
||
|
||
@property
|
||
def temporary(self):
|
||
|
||
return self._temporary
|
||
|
||
@temporary.setter
|
||
def temporary(self, temporary):
|
||
|
||
if hasattr(self, 'temporary') and temporary == self._temporary:
|
||
return
|
||
|
||
self._temporary = temporary
|
||
self._update_temporary_file()
|
||
|
||
def record_tcp_port(self, port):
|
||
"""
|
||
Associate a reserved TCP port number with this project.
|
||
|
||
:param port: TCP port number
|
||
"""
|
||
|
||
if port not in self._used_tcp_ports:
|
||
self._used_tcp_ports.add(port)
|
||
|
||
def record_udp_port(self, port):
|
||
"""
|
||
Associate a reserved UDP port number with this project.
|
||
|
||
:param port: UDP port number
|
||
"""
|
||
|
||
if port not in self._used_udp_ports:
|
||
self._used_udp_ports.add(port)
|
||
|
||
def remove_tcp_port(self, port):
|
||
"""
|
||
Removes an associated TCP port number from this project.
|
||
|
||
:param port: TCP port number
|
||
"""
|
||
|
||
if port in self._used_tcp_ports:
|
||
self._used_tcp_ports.remove(port)
|
||
|
||
def remove_udp_port(self, port):
|
||
"""
|
||
Removes an associated UDP port number from this project.
|
||
|
||
:param port: UDP port number
|
||
"""
|
||
|
||
if port in self._used_udp_ports:
|
||
self._used_udp_ports.remove(port)
|
||
|
||
def _update_temporary_file(self):
|
||
"""
|
||
Update the .gns3_temporary file in order to reflect the current project status.
|
||
"""
|
||
|
||
if not hasattr(self, "_path"):
|
||
return
|
||
|
||
if self._temporary:
|
||
try:
|
||
with open(os.path.join(self._path, ".gns3_temporary"), 'w+') as f:
|
||
f.write("1")
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not create temporary project: {}".format(e))
|
||
else:
|
||
if os.path.exists(os.path.join(self._path, ".gns3_temporary")):
|
||
try:
|
||
os.remove(os.path.join(self._path, ".gns3_temporary"))
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not mark project as no longer temporary: {}".format(e))
|
||
|
||
def module_working_directory(self, module_name):
|
||
"""
|
||
Returns a working directory for the module
|
||
The directory is created if the directory doesn't exist.
|
||
|
||
:param module_name: name for the module
|
||
:returns: working directory
|
||
"""
|
||
|
||
workdir = self.module_working_path(module_name)
|
||
try:
|
||
os.makedirs(workdir, exist_ok=True)
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not create module working directory: {}".format(e))
|
||
return workdir
|
||
|
||
def module_working_path(self, module_name):
|
||
"""
|
||
Returns the working directory for the module. If you want
|
||
to be sure to have the directory on disk take a look on:
|
||
module_working_directory
|
||
"""
|
||
|
||
return os.path.join(self._path, "project-files", module_name)
|
||
|
||
def node_working_directory(self, node):
|
||
"""
|
||
Returns a working directory for a specific node.
|
||
If the directory doesn't exist, the directory is created.
|
||
|
||
:param node: Node instance
|
||
|
||
:returns: Node working directory
|
||
"""
|
||
|
||
workdir = os.path.join(self._path, "project-files", node.manager.module_name.lower(), node.id)
|
||
try:
|
||
os.makedirs(workdir, exist_ok=True)
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not create the node working directory: {}".format(e))
|
||
return workdir
|
||
|
||
def tmp_working_directory(self):
|
||
"""
|
||
A temporary directory. Will be clean at project open and close
|
||
"""
|
||
return os.path.join(self._path, "tmp")
|
||
|
||
def capture_working_directory(self):
|
||
"""
|
||
Returns a working directory where to temporary store packet capture files.
|
||
|
||
:returns: path to the directory
|
||
"""
|
||
|
||
workdir = os.path.join(self._path, "tmp", "captures")
|
||
try:
|
||
os.makedirs(workdir, exist_ok=True)
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not create the capture working directory: {}".format(e))
|
||
return workdir
|
||
|
||
def mark_node_for_destruction(self, node):
|
||
"""
|
||
:param node: An instance of Node
|
||
"""
|
||
|
||
self.remove_node(node)
|
||
self._nodes_to_destroy.add(node)
|
||
|
||
def add_node(self, node):
|
||
"""
|
||
Adds a node to the project.
|
||
In theory this should be called by the node manager.
|
||
|
||
:param node: Node instance
|
||
"""
|
||
|
||
self._nodes.add(node)
|
||
|
||
def remove_node(self, node):
|
||
"""
|
||
Removes a node from the project.
|
||
In theory this should be called by the node manager.
|
||
|
||
:param node: Node instance
|
||
"""
|
||
|
||
if node in self._nodes:
|
||
self._nodes.remove(node)
|
||
|
||
@asyncio.coroutine
|
||
def close(self):
|
||
"""
|
||
Closes the project, but keep information on disk
|
||
"""
|
||
|
||
for module in self.compute():
|
||
yield from module.instance().project_closing(self)
|
||
yield from self._close_and_clean(self._temporary)
|
||
for module in self.compute():
|
||
yield from module.instance().project_closed(self)
|
||
|
||
try:
|
||
if os.path.exists(self.tmp_working_directory()):
|
||
shutil.rmtree(self.tmp_working_directory())
|
||
except OSError:
|
||
pass
|
||
|
||
@asyncio.coroutine
|
||
def _close_and_clean(self, cleanup):
|
||
"""
|
||
Closes the project, and cleanup the disk if cleanup is True
|
||
|
||
:param cleanup: Whether to delete the project directory
|
||
"""
|
||
|
||
tasks = []
|
||
for node in self._nodes:
|
||
tasks.append(asyncio.async(node.manager.close_node(node.id)))
|
||
|
||
if tasks:
|
||
done, _ = yield from asyncio.wait(tasks)
|
||
for future in done:
|
||
try:
|
||
future.result()
|
||
except (Exception, GeneratorExit) as e:
|
||
log.error("Could not close node {}".format(e), exc_info=1)
|
||
|
||
if cleanup and os.path.exists(self.path):
|
||
try:
|
||
yield from wait_run_in_executor(shutil.rmtree, self.path)
|
||
log.info("Project {id} with path '{path}' deleted".format(path=self._path, id=self._id))
|
||
except OSError as e:
|
||
raise aiohttp.web.HTTPInternalServerError(text="Could not delete the project directory: {}".format(e))
|
||
else:
|
||
log.info("Project {id} with path '{path}' closed".format(path=self._path, id=self._id))
|
||
|
||
if self._used_tcp_ports:
|
||
log.warning("Project {} has TCP ports still in use: {}".format(self.id, self._used_tcp_ports))
|
||
if self._used_udp_ports:
|
||
log.warning("Project {} has UDP ports still in use: {}".format(self.id, self._used_udp_ports))
|
||
|
||
# clean the remaining ports that have not been cleaned by their respective node.
|
||
port_manager = PortManager.instance()
|
||
for port in self._used_tcp_ports.copy():
|
||
port_manager.release_tcp_port(port, self)
|
||
for port in self._used_udp_ports.copy():
|
||
port_manager.release_udp_port(port, self)
|
||
|
||
@asyncio.coroutine
|
||
def commit(self):
|
||
"""
|
||
Writes project changes on disk
|
||
"""
|
||
|
||
while self._nodes_to_destroy:
|
||
node = self._nodes_to_destroy.pop()
|
||
yield from node.delete()
|
||
self.remove_node(node)
|
||
for module in self.compute():
|
||
yield from module.instance().project_committed(self)
|
||
|
||
@asyncio.coroutine
|
||
def delete(self):
|
||
"""
|
||
Removes project from disk
|
||
"""
|
||
|
||
for module in self.compute():
|
||
yield from module.instance().project_closing(self)
|
||
yield from self._close_and_clean(True)
|
||
for module in self.compute():
|
||
yield from module.instance().project_closed(self)
|
||
|
||
@classmethod
|
||
def clean_project_directory(cls):
|
||
"""
|
||
At startup drop old temporary project. After a crash for example
|
||
"""
|
||
|
||
directory = get_default_project_directory()
|
||
if os.path.exists(directory):
|
||
for project in os.listdir(directory):
|
||
path = os.path.join(directory, project)
|
||
if os.path.exists(os.path.join(path, ".gns3_temporary")):
|
||
log.warning("Purge old temporary project {}".format(project))
|
||
shutil.rmtree(path)
|
||
|
||
def compute(self):
|
||
"""
|
||
Returns all loaded modules from compute.
|
||
"""
|
||
|
||
# We import it at the last time to avoid circular dependencies
|
||
from ..compute import MODULES
|
||
return MODULES
|
||
|
||
def emit(self, action, event):
|
||
"""
|
||
Send an event to all the client listening for notifications
|
||
|
||
:param action: Action name
|
||
:param event: Event to send
|
||
"""
|
||
NotificationManager.instance().emit(action, event, project_id=self.id)
|
||
|
||
@asyncio.coroutine
|
||
def list_files(self):
|
||
"""
|
||
:returns: Array of files in project without temporary files. The files are dictionary {"path": "test.bin", "md5sum": "aaaaa"}
|
||
"""
|
||
|
||
files = []
|
||
for (dirpath, dirnames, filenames) in os.walk(self.path):
|
||
for filename in filenames:
|
||
if not filename.endswith(".ghost"):
|
||
path = os.path.relpath(dirpath, self.path)
|
||
path = os.path.join(path, filename)
|
||
path = os.path.normpath(path)
|
||
file_info = {"path": path}
|
||
|
||
try:
|
||
file_info["md5sum"] = yield from wait_run_in_executor(self._hash_file, os.path.join(dirpath, filename))
|
||
except OSError:
|
||
continue
|
||
files.append(file_info)
|
||
|
||
return files
|
||
|
||
def _hash_file(self, path):
|
||
"""
|
||
Compute and md5 hash for file
|
||
|
||
:returns: hexadecimal md5
|
||
"""
|
||
|
||
m = hashlib.md5()
|
||
with open(path, "rb") as f:
|
||
while True:
|
||
buf = f.read(128)
|
||
if not buf:
|
||
break
|
||
m.update(buf)
|
||
return m.hexdigest()
|
||
|
||
def export(self, include_images=False):
|
||
"""
|
||
Export the project as zip. It's a ZipStream object.
|
||
The file will be read chunk by chunk when you iterate on
|
||
the zip.
|
||
|
||
It will ignore some files like snapshots and
|
||
|
||
:returns: ZipStream object
|
||
"""
|
||
|
||
z = zipstream.ZipFile()
|
||
# topdown allows to modify the list of directory in order to ignore the directory
|
||
for root, dirs, files in os.walk(self._path, topdown=True):
|
||
# Remove snapshots and capture
|
||
if os.path.split(root)[-1:][0] == "project-files":
|
||
dirs[:] = [d for d in dirs if d not in ("snapshots", "tmp")]
|
||
|
||
# Ignore log files and OS noise
|
||
files = [f for f in files if not f.endswith('_log.txt') and not f.endswith('.log') and f != '.DS_Store']
|
||
|
||
for file in files:
|
||
path = os.path.join(root, file)
|
||
# We rename the .gns3 project.gns3 to avoid the task to the client to guess the file name
|
||
if file.endswith(".gns3"):
|
||
self._export_project_file(path, z, include_images)
|
||
else:
|
||
# We merge the data from all server in the same project-files directory
|
||
node_directory = os.path.join(self._path, "servers", "vm")
|
||
if os.path.commonprefix([root, node_directory]) == node_directory:
|
||
z.write(path, os.path.relpath(path, node_directory))
|
||
else:
|
||
z.write(path, os.path.relpath(path, self._path))
|
||
return z
|
||
|
||
def _export_images(self, image, type, z):
|
||
"""
|
||
Take a project file (.gns3) and export images to the zip
|
||
|
||
:param image: Image path
|
||
:param type: Type of image
|
||
:param z: Zipfile instance for the export
|
||
"""
|
||
from . import MODULES
|
||
|
||
for module in MODULES:
|
||
try:
|
||
img_directory = module.instance().get_images_directory()
|
||
except NotImplementedError:
|
||
# Some modules don't have images
|
||
continue
|
||
|
||
directory = os.path.split(img_directory)[-1:][0]
|
||
|
||
if os.path.exists(image):
|
||
path = image
|
||
else:
|
||
path = os.path.join(img_directory, image)
|
||
|
||
if os.path.exists(path):
|
||
arcname = os.path.join("images", directory, os.path.basename(image))
|
||
z.write(path, arcname)
|
||
break
|
||
|
||
def _export_project_file(self, path, z, include_images):
|
||
"""
|
||
Take a project file (.gns3) and patch it for the export
|
||
|
||
:param path: Path of the .gns3
|
||
"""
|
||
|
||
with open(path) as f:
|
||
topology = json.load(f)
|
||
if "topology" in topology and "nodes" in topology["topology"]:
|
||
for node in topology["topology"]["nodes"]:
|
||
if "properties" in node and node["type"] != "DockerVM":
|
||
for prop, value in node["properties"].items():
|
||
if prop.endswith("image"):
|
||
node["properties"][prop] = os.path.basename(value)
|
||
if include_images is True:
|
||
self._export_images(value, node["type"], z)
|
||
z.writestr("project.gns3", json.dumps(topology).encode())
|
||
|
||
def import_zip(self, stream, gns3vm=True):
|
||
"""
|
||
Import a project contain in a zip file
|
||
|
||
:param stream: A io.BytesIO of the zipfile
|
||
:param gns3vm: True move Docker, IOU and Qemu to the GNS3 VM
|
||
"""
|
||
|
||
with zipfile.ZipFile(stream) as myzip:
|
||
myzip.extractall(self.path)
|
||
|
||
project_file = os.path.join(self.path, "project.gns3")
|
||
if os.path.exists(project_file):
|
||
with open(project_file) as f:
|
||
topology = json.load(f)
|
||
topology["project_id"] = self.id
|
||
topology["name"] = self.name
|
||
topology.setdefault("topology", {})
|
||
topology["topology"].setdefault("nodes", [])
|
||
topology["topology"]["servers"] = [
|
||
{
|
||
"id": 1,
|
||
"local": True,
|
||
"vm": False
|
||
}
|
||
]
|
||
|
||
# By default all node run on local server
|
||
for node in topology["topology"]["nodes"]:
|
||
node["server_id"] = 1
|
||
|
||
if gns3vm:
|
||
# Move to servers/vm directory the data that should be import on remote server
|
||
modules_to_vm = {
|
||
"qemu": "QemuVM",
|
||
"iou": "IOUDevice",
|
||
"docker": "DockerVM"
|
||
}
|
||
|
||
node_directory = os.path.join(self.path, "servers", "vm", "project-files")
|
||
vm_server_use = False
|
||
|
||
for module, vm_type in modules_to_vm.items():
|
||
module_directory = os.path.join(self.path, "project-files", module)
|
||
if os.path.exists(module_directory):
|
||
os.makedirs(node_directory, exist_ok=True)
|
||
shutil.move(module_directory, os.path.join(node_directory, module))
|
||
|
||
# Patch node to use the GNS3 VM
|
||
for node in topology["topology"]["nodes"]:
|
||
if node["type"] == vm_type:
|
||
node["server_id"] = 2
|
||
vm_server_use = True
|
||
|
||
# We use the GNS3 VM. We need to add the server to the list
|
||
if vm_server_use:
|
||
topology["topology"]["servers"].append({
|
||
"id": 2,
|
||
"vm": True,
|
||
"local": False
|
||
})
|
||
|
||
# Write the modified topology
|
||
with open(project_file, "w") as f:
|
||
json.dump(topology, f, indent=4)
|
||
|
||
# Rename to a human distinctive name
|
||
shutil.move(project_file, os.path.join(self.path, self.name + ".gns3"))
|
||
if os.path.exists(os.path.join(self.path, "images")):
|
||
self._import_images()
|
||
|
||
def _import_images(self):
|
||
"""
|
||
Copy images to the images directory or delete them if they
|
||
already exists.
|
||
"""
|
||
image_dir = self._config().get("images_path")
|
||
|
||
root = os.path.join(self.path, "images")
|
||
for (dirpath, dirnames, filenames) in os.walk(root):
|
||
for filename in filenames:
|
||
path = os.path.join(dirpath, filename)
|
||
dst = os.path.join(image_dir, os.path.relpath(path, root))
|
||
os.makedirs(os.path.dirname(dst), exist_ok=True)
|
||
shutil.move(path, dst)
|
||
|
||
# Cleanup the project
|
||
shutil.rmtree(root, ignore_errors=True)
|