/* Serval Distributed Numbering Architecture (DNA) Copyright (C) 2010 Paul Gardner-Stephen This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. */ /* Serval Overlay Mesh Network. Basically we use UDP broadcast to send link-local, and then implement a BATMAN-like protocol over the top of that. Each overlay packet can contain one or more encapsulated packets each addressed using Serval DNA SIDs, with source, destination and next-hop addresses. The use of an overlay also lets us be a bit clever about using irregular transports, such as an ISM915 modem attached via ethernet (which we are planning to build in coming months), by paring off the IP and UDP headers that would otherwise dominate. Even on regular WiFi and ethernet we can aggregate packets in a way similar to IAX, but not just for voice frames. The use of long (relative to IPv4 or even IPv6) 256 bit Curve25519 addresses means that it is a really good idea to have neighbouring nodes exchange lists of peer aliases so that addresses can be summarised, possibly using less space than IPv4 would have. One approach to handle address shortening is to have the periodic TTL=255 BATMAN-style hello packets include an epoch number. This epoch number can be used by immediate neighbours of the originator to reference the neighbours listed in that packet by their ordinal position in the packet instead of by their full address. This gets us address shortening to 1 byte in most cases in return for no new packets, but the periodic hello packets will now be larger. We might deal with this issue by having these hello packets reference the previous epoch for common neighbours. Unresolved neighbour addresses could be resolved by a simple DNA request, which should only need to occur ocassionally, and other link-local neighbours could sniff and cache the responses to avoid duplicated traffic. Indeed, during quiet times nodes could preemptively advertise address resolutions if they wished, or similarly advertise the full address of a few (possibly randomly selected) neighbours in each epoch. Byzantine Robustness is a goal, so we have to think about all sorts of malicious failure modes. One approach to help byzantine robustness is to have multiple signature shells for each hop for mesh topology packets. Thus forging a report of closeness requires forging a signature. As such frames are forwarded, the outermost signature shell is removed. This is really only needed for more paranoid uses. We want to have different traffic classes for voice/video calls versus regular traffic, e.g., MeshMS frames. Thus we need to have separate traffic queues for these items. Aside from allowing us to prioritise isochronous data, it also allows us to expire old isochronous frames that are in-queue once there is no longer any point delivering them (e.g after holding them more than 200ms). We can also be clever about round-robin fair-sharing or even prioritising among isochronous streams. Since we also know about the DNA isochronous protocols and the forward error correction and other redundancy measures we also get smart about dropping, say, 1 in 3 frames from every call if we know that this can be safely done. That is, when traffic is low, we maximise redundancy, and when we start to hit the limit of traffic, we start to throw away some of the redundancy. This of course relies on us knowing when the network channel is getting too full. Smart-flooding of broadcast information is also a requirement. The long addresses help here, as we can make any address that begins with the first 192 bits all ones be broadcast, and use the remaining 64 bits as a "broadcast packet identifier" (BPI). Nodes can remember recently seen BPIs and not forward broadcast frames that have been seen recently. This should get us smart flooding of the majority of a mesh (with some node mobility issues being a factor). We could refine this later, but it will do for now, especially since for things like number resolution we are happy to send repeat requests. This file currently seems to exist solely to contain this introduction, which is fine with me. Functions land in here until their proper place becomes apparent. */ #include "serval.h" #include "conf.h" #include "rhizome.h" #include "httpd.h" #include "strbuf.h" #include "keyring.h" #include "overlay_interface.h" int overlayMode=0; keyring_file *keyring=NULL; int overlayServerMode(const struct cli_parsed *parsed) { IN(); /* Setup up client API sockets before writing our PID file We want clients to be able to connect to our sockets as soon as servald start has returned. But we don't want servald start to take very long. Try to perform only minimal CPU or IO processing here. */ overlay_mdp_setup_sockets(); monitor_setup_sockets(); // start the HTTP server if enabled httpd_server_start(HTTPD_PORT, HTTPD_PORT_MAX); /* record PID file so that servald start can return */ if (server_write_pid()) RETURN(-1); /* Get keyring available for use. Required for MDP, and very soon as a complete replacement for the HLR for DNA lookups, even in non-overlay mode. */ keyring = keyring_open_instance_cli(parsed); if (!keyring) RETURN(WHY("Could not open serval keyring file.")); /* put initial identity in if we don't have any visible */ keyring_seed(keyring); overlay_queue_init(); /* Get the set of socket file descriptors we need to monitor. Note that end-of-file will trigger select(), so we cannot run select() if we have any dummy interfaces running. So we do an ugly hack of just waiting no more than 5ms between checks if we have a dummy interface running. This is a reasonable simulation of wifi latency anyway, so we'll live with it. Larger values will affect voice transport, and smaller values would affect CPU and energy use, and make the simulation less realistic. */ #define SCHEDULE(X, Y, D) { \ static struct profile_total _stats_##X={.name="" #X "",}; \ static struct sched_ent _sched_##X={\ .stats = &_stats_##X, \ .function=X,\ }; \ _sched_##X.alarm=(gettime_ms()+Y);\ _sched_##X.deadline=(gettime_ms()+Y+D);\ schedule(&_sched_##X); } /* Periodically check for server shut down */ SCHEDULE(server_shutdown_check, 0, 100); /* Periodically reload configuration */ SCHEDULE(server_config_reload, SERVER_CONFIG_RELOAD_INTERVAL_MS, SERVER_CONFIG_RELOAD_INTERVAL_MS + 100); overlay_mdp_bind_internal_services(); olsr_init_socket(); /* Get rhizome server started BEFORE populating fd list so that the server's listen socket is in the list for poll() */ if (is_rhizome_enabled()){ rhizome_opendb(); if (config.rhizome.clean_on_start && !config.rhizome.clean_on_open) rhizome_cleanup(NULL); } // start the dna helper if configured dna_helper_start(); // preload directory service information directory_service_init(); /* Periodically check for new interfaces */ SCHEDULE(overlay_interface_discover, 1, 100); /* Periodically advertise bundles */ SCHEDULE(overlay_rhizome_advertise, 1000, 10000); /* Calculate (and possibly show) CPU usage stats periodically */ SCHEDULE(fd_periodicstats, 3000, 500); #undef SCHEDULE // log message used by tests to wait for the server to start INFO("Server initialised, entering main loop"); /* Check for activitiy and respond to it */ while(fd_poll()); RETURN(0); OUT(); }