gns3-server/gns3server/controller/link.py

471 lines
15 KiB
Python
Raw Normal View History

2016-03-11 15:51:35 +00:00
#!/usr/bin/env python
#
# Copyright (C) 2016 GNS3 Technologies Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import os
import re
2016-03-11 15:51:35 +00:00
import uuid
import html
import aiohttp
2016-03-11 15:51:35 +00:00
import logging
log = logging.getLogger(__name__)
FILTERS = [
{
"type": "frequency_drop",
"name": "Frequency drop",
"description": "It will drop everything with a -1 frequency, drop every Nth packet with a positive frequency, or drop nothing",
"parameters": [
{
"name": "Frequency",
"minimum": -1,
"maximum": 32767,
"type": "int",
"unit": "th packet"
}
]
},
{
"type": "packet_loss",
"name": "Packet loss",
"description": "The percentage represents the chance for a packet to be lost",
"parameters": [
{
2017-07-06 09:53:05 +00:00
"name": "Chance",
"minimum": 0,
"maximum": 100,
"type": "int",
"unit": "%"
}
]
},
{
"type": "delay",
"name": "Delay",
"description": "Delay packets in milliseconds. You can add jitter in milliseconds (+/-) of the delay",
"parameters": [
{
2017-07-06 09:53:05 +00:00
"name": "Latency",
"minimum": 0,
"maximum": 32767,
"unit": "ms",
"type": "int"
},
{
2017-07-06 09:53:05 +00:00
"name": "Jitter (-/+)",
"minimum": 0,
"maximum": 32767,
"unit": "ms",
"type": "int"
}
]
},
{
"type": "corrupt",
"name": "Corrupt",
2017-07-05 14:36:39 +00:00
"description": "The percentage represents the chance for a packet to be corrupted",
"parameters": [
{
2017-07-06 09:53:05 +00:00
"name": "Chance",
"minimum": 0,
"maximum": 100,
"unit": "%",
"type": "int"
}
]
},
{
"type": "bpf",
2017-07-12 09:21:11 +00:00
"name": "Berkeley Packet Filter (BPF)",
"description": "This filter will drop any packet matching a BPF expression. Put one expression per line",
"parameters": [
{
2017-07-12 09:21:11 +00:00
"name": "Filters",
"type": "text"
}
]
}
]
2016-03-11 15:51:35 +00:00
class Link:
2016-06-03 03:32:46 +00:00
"""
Base class for links.
"""
2016-03-11 16:02:50 +00:00
def __init__(self, project, link_id=None):
2016-06-03 03:32:46 +00:00
if link_id:
self._id = link_id
else:
self._id = str(uuid.uuid4())
self._nodes = []
self._project = project
self._capturing = False
self._capture_node = None
self._capture_file_name = None
2016-05-14 00:48:10 +00:00
self._streaming_pcap = None
self._created = False
self._link_type = "ethernet"
self._suspended = False
self._filters = {}
@property
def filters(self):
"""
Get an array of filters
"""
return self._filters
@property
def nodes(self):
"""
Get the current nodes attached to this link
"""
return self._nodes
@property
def project(self):
"""
Get the project this link belongs to.
:return: Project instance.
"""
return self._project
@property
def capture_node(self):
"""
Get the capturing node
:return: Node instance.
"""
return self._capture_node
@property
def compute(self):
"""
Get the capturing node
:return: Node instance.
"""
assert self.capture_node
return self.capture_node["node"].compute
2017-07-19 15:30:25 +00:00
def get_active_filters(self):
"""
Return the active filters.
Filters are overridden if the link is suspended.
2017-07-19 15:30:25 +00:00
"""
if self._suspended:
# this is to allow all node types to support suspend link
2017-07-19 15:30:25 +00:00
return {"frequency_drop": [-1]}
return self._filters
async def update_filters(self, filters):
"""
Modify the filters list.
Filter with value 0 will be dropped because not active
"""
new_filters = {}
for (filter, values) in filters.items():
new_values = []
for value in values:
if isinstance(value, str):
new_values.append(value.strip("\n "))
else:
new_values.append(int(value))
values = new_values
if len(values) != 0 and values[0] != 0 and values[0] != '':
new_filters[filter] = values
if new_filters != self.filters:
self._filters = new_filters
if self._created:
await self.update()
self._project.emit_notification("link.updated", self.__json__())
2018-03-12 06:38:50 +00:00
self._project.dump()
async def update_suspend(self, value):
if value != self._suspended:
self._suspended = value
await self.update()
self._project.emit_notification("link.updated", self.__json__())
2018-03-12 06:38:50 +00:00
self._project.dump()
2017-07-19 15:30:25 +00:00
@property
def created(self):
"""
:returns: True the link has been created on the computes
"""
return self._created
2016-03-11 15:51:35 +00:00
async def add_node(self, node, adapter_number, port_number, label=None, dump=True):
2016-03-11 15:51:35 +00:00
"""
Add a node to the link
:param dump: Dump project on disk
2016-03-11 15:51:35 +00:00
"""
port = node.get_port(adapter_number, port_number)
2018-03-12 06:38:50 +00:00
if port is None:
raise aiohttp.web.HTTPNotFound(text="Port {}/{} for {} not found".format(adapter_number, port_number, node.name))
if port.link is not None:
raise aiohttp.web.HTTPConflict(text="Port is already used")
self._link_type = port.link_type
for other_node in self._nodes:
if other_node["node"] == node:
raise aiohttp.web.HTTPConflict(text="Cannot connect to itself")
if node.node_type in ["nat", "cloud"]:
if other_node["node"].node_type in ["nat", "cloud"]:
raise aiohttp.web.HTTPConflict(text="Connecting a {} to a {} is not allowed".format(other_node["node"].node_type, node.node_type))
# Check if user is not connecting serial => ethernet
other_port = other_node["node"].get_port(other_node["adapter_number"], other_node["port_number"])
2018-03-12 06:38:50 +00:00
if other_port is None:
raise aiohttp.web.HTTPNotFound(text="Port {}/{} for {} not found".format(other_node["adapter_number"], other_node["port_number"], other_node["node"].name))
if port.link_type != other_port.link_type:
raise aiohttp.web.HTTPConflict(text="Connecting a {} interface to a {} interface is not allowed".format(other_port.link_type, port.link_type))
if label is None:
label = {
"text": html.escape("{}/{}".format(adapter_number, port_number)),
"style": "font-family: TypeWriter;font-size: 10.0;font-weight: bold;fill: #000000;fill-opacity: 1.0;"
}
self._nodes.append({
"node": node,
2016-03-11 15:51:35 +00:00
"adapter_number": adapter_number,
"port_number": port_number,
"port": port,
"label": label
2016-03-11 15:51:35 +00:00
})
2016-06-03 03:32:46 +00:00
if len(self._nodes) == 2:
await self.create()
for n in self._nodes:
n["node"].add_link(self)
n["port"].link = self
self._created = True
self._project.emit_notification("link.created", self.__json__())
if dump:
self._project.dump()
async def update_nodes(self, nodes):
for node_data in nodes:
node = self._project.get_node(node_data["node_id"])
for port in self._nodes:
if port["node"] == node:
label = node_data.get("label")
if label:
port["label"] = label
self._project.emit_notification("link.updated", self.__json__())
2016-06-15 13:12:38 +00:00
self._project.dump()
2016-03-11 15:51:35 +00:00
async def create(self):
2016-03-14 16:40:27 +00:00
"""
Create the link
"""
2016-06-03 03:32:46 +00:00
raise NotImplementedError
2016-03-14 16:40:27 +00:00
async def update(self):
"""
Update a link
"""
raise NotImplementedError
async def delete(self):
2016-03-14 16:40:27 +00:00
"""
Delete the link
"""
for n in self._nodes:
# It could be different of self if we rollback an already existing link
if n["port"].link == self:
n["port"].link = None
n["node"].remove_link(self)
2016-03-14 16:40:27 +00:00
async def start_capture(self, data_link_type="DLT_EN10MB", capture_file_name=None):
"""
Start capture on the link
:returns: Capture object
"""
2016-06-03 03:32:46 +00:00
self._capturing = True
self._capture_file_name = capture_file_name
self._project.emit_notification("link.updated", self.__json__())
async def stop_capture(self):
"""
Stop capture on the link
"""
2016-06-03 03:32:46 +00:00
self._capturing = False
self._project.emit_notification("link.updated", self.__json__())
def pcap_streaming_url(self):
"""
Get the PCAP streaming URL on compute
:returns: URL
"""
2016-06-03 03:32:46 +00:00
assert self.capture_node
compute = self.capture_node["node"].compute
node_type = self.capture_node["node"].node_type
node_id = self.capture_node["node"].id
adapter_number = self.capture_node["adapter_number"]
port_number = self.capture_node["port_number"]
url = "/projects/{project_id}/{node_type}/nodes/{node_id}/adapters/{adapter_number}/ports/{port_number}/pcap".format(project_id=self.project.id,
node_type=node_type,
node_id=node_id,
adapter_number=adapter_number,
port_number=port_number)
return compute._getUrl(url)
async def node_updated(self, node):
"""
Called when a node member of the link is updated
"""
raise NotImplementedError
def default_capture_file_name(self):
"""
:returns: File name for a capture on this link
"""
2016-06-03 03:32:46 +00:00
2016-05-14 00:48:10 +00:00
capture_file_name = "{}_{}-{}_to_{}_{}-{}".format(self._nodes[0]["node"].name,
self._nodes[0]["adapter_number"],
self._nodes[0]["port_number"],
self._nodes[1]["node"].name,
self._nodes[1]["adapter_number"],
self._nodes[1]["port_number"])
return re.sub(r"[^0-9A-Za-z_-]", "", capture_file_name) + ".pcap"
2016-03-11 15:51:35 +00:00
@property
def id(self):
return self._id
@property
def nodes(self):
return [node['node'] for node in self._nodes]
@property
def capturing(self):
return self._capturing
@property
def capture_file_path(self):
"""
Get the path of the capture
"""
2016-06-03 03:32:46 +00:00
if self._capture_file_name:
return os.path.join(self._project.captures_directory, self._capture_file_name)
else:
return None
@property
def capture_compute_id(self):
"""
Get the capture compute ID.
"""
if self._capture_node:
return self.capture_node["node"].compute.id
else:
return None
def available_filters(self):
"""
Return the list of filters compatible with this link
:returns: Array of filters
"""
filter_node = self._get_filter_node()
if filter_node:
return FILTERS
return []
def _get_filter_node(self):
"""
Return the node where the filter will run
:returns: None if no node support filtering else the node
"""
for node in self._nodes:
2017-07-17 09:21:54 +00:00
if node["node"].node_type in ('vpcs',
2018-03-12 10:57:13 +00:00
'traceng',
2017-07-18 16:04:03 +00:00
'vmware',
2017-07-17 09:21:54 +00:00
'dynamips',
'qemu',
2017-07-18 07:24:36 +00:00
'iou',
'cloud',
'nat',
'virtualbox',
'docker'):
return node["node"]
return None
def __eq__(self, other):
if not isinstance(other, Link):
return False
return self.id == other.id
def __hash__(self):
return hash(self._id)
2017-07-18 07:24:36 +00:00
def __json__(self, topology_dump=False):
2016-06-15 13:12:38 +00:00
"""
:param topology_dump: Filter to keep only properties require for saving on disk
"""
2017-07-18 07:24:36 +00:00
res = []
for side in self._nodes:
2016-03-11 15:51:35 +00:00
res.append({
"node_id": side["node"].id,
2016-03-11 15:51:35 +00:00
"adapter_number": side["adapter_number"],
"port_number": side["port_number"],
"label": side["label"]
2016-03-11 15:51:35 +00:00
})
2016-06-15 13:12:38 +00:00
if topology_dump:
return {
"nodes": res,
"link_id": self._id,
2017-07-19 15:30:25 +00:00
"filters": self._filters,
"suspend": self._suspended
2016-06-15 13:12:38 +00:00
}
return {
2016-06-03 03:32:46 +00:00
"nodes": res,
"link_id": self._id,
"project_id": self._project.id,
"capturing": self._capturing,
2016-04-26 15:36:24 +00:00
"capture_file_name": self._capture_file_name,
"capture_file_path": self.capture_file_path,
"capture_compute_id": self.capture_compute_id,
"link_type": self._link_type,
2017-07-19 15:30:25 +00:00
"filters": self._filters,
"suspend": self._suspended
}