diff options
author | Julius Bünger <buenger@mytum.de> | 2018-06-06 00:33:56 +0200 |
---|---|---|
committer | Julius Bünger <buenger@mytum.de> | 2018-06-06 00:34:36 +0200 |
commit | d15b9286cbaa6f56299f9064d269d3a1aa1ddee4 (patch) | |
tree | d9798b5d8495cdd4b92c58a1015a8f83bbe827d1 | |
parent | 6ab858be606ee40db66e349c165a6862a79c65a4 (diff) |
rps: split tests and profiler, proper shutdown
-rw-r--r-- | src/rps/Makefile.am | 16 | ||||
-rw-r--r-- | src/rps/gnunet-rps-profiler.c | 2697 | ||||
-rw-r--r-- | src/rps/test_rps.c | 108 | ||||
-rw-r--r-- | src/rps/test_rps.conf | 2 |
4 files changed, 2792 insertions, 31 deletions
diff --git a/src/rps/Makefile.am b/src/rps/Makefile.am index b5f7122683..aa354dff77 100644 --- a/src/rps/Makefile.am +++ b/src/rps/Makefile.am @@ -83,17 +83,17 @@ check_PROGRAMS = \ test_rps_churn endif +rps_test_src = \ + test_rps.c \ + rps-test_util.h rps-test_util.c \ + gnunet-service-rps_sampler_elem.h gnunet-service-rps_sampler_elem.c + ld_rps_test_lib = \ libgnunetrps.la \ $(top_builddir)/src/util/libgnunetutil.la \ $(top_builddir)/src/testbed/libgnunettestbed.la \ -lm -rps_test_src = \ - test_rps.c \ - rps-test_util.h rps-test_util.c \ - gnunet-service-rps_sampler_elem.h gnunet-service-rps_sampler_elem.c - if ENABLE_TEST_RUN AM_TESTS_ENVIRONMENT=export GNUNET_PREFIX=$${GNUNET_PREFIX:-@libdir@};export PATH=$${GNUNET_PREFIX:-@prefix@}/bin:$$PATH;unset XDG_DATA_HOME;unset XDG_CONFIG_HOME; TESTS = $(check_PROGRAMS) @@ -143,7 +143,11 @@ test_rps_churn_LDADD = $(ld_rps_test_lib) gnunet_rps_profiler_SOURCES = $(rps_test_src) gnunet_rps_profiler_LDADD = \ $(top_builddir)/src/statistics/libgnunetstatistics.la \ - $(ld_rps_test_lib) + libgnunetrps.la \ + $(top_builddir)/src/util/libgnunetutil.la \ + $(top_builddir)/src/testbed/libgnunettestbed.la \ + -lm + EXTRA_DIST = \ test_rps.conf diff --git a/src/rps/gnunet-rps-profiler.c b/src/rps/gnunet-rps-profiler.c new file mode 100644 index 0000000000..1ad9deabb2 --- /dev/null +++ b/src/rps/gnunet-rps-profiler.c @@ -0,0 +1,2697 @@ +/* + This file is part of GNUnet. + Copyright (C) 2009, 2012 GNUnet e.V. + + GNUnet is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published + by the Free Software Foundation; either version 3, or (at your + option) any later version. + + GNUnet is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + General Public License for more details. + + You should have received a copy of the GNU General Public License + along with GNUnet; see the file COPYING. If not, write to the + Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, + Boston, MA 02110-1301, USA. +*/ +/** + * @file rps/test_rps.c + * @brief Testcase for the random peer sampling service. Starts + * a peergroup with a given number of peers, then waits to + * receive size pushes/pulls from each peer. Expects to wait + * for one message from each peer. + */ +#include "platform.h" +//#include "rps_test_lib.h" +#include "gnunet_util_lib.h" +#include "gnunet_testbed_service.h" + +#include "gnunet_rps_service.h" +#include "rps-test_util.h" +#include "gnunet-service-rps_sampler_elem.h" + +#include <inttypes.h> + + +/** + * How many peers do we start? + */ +static uint32_t num_peers; + +/** + * How long do we run the test? + * In seconds. + */ +static uint32_t timeout_s; + +/** + * How long do we run the test? + */ +//#define TIMEOUT GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 30) +static struct GNUNET_TIME_Relative timeout; + + +/** + * Portion of malicious peers + */ +static double portion = .1; + +/** + * Type of malicious peer to test + */ +static unsigned int mal_type = 0; + +/** + * Handles to all of the running peers + */ +static struct GNUNET_TESTBED_Peer **testbed_peers; + +/** + * @brief Indicates whether peer should go off- or online + */ +enum PEER_ONLINE_DELTA { + /** + * @brief Indicates peer going online + */ + PEER_GO_ONLINE = 1, + /** + * @brief Indicates peer going offline + */ + PEER_GO_OFFLINE = -1, +}; + +/** + * Operation map entry + */ +struct OpListEntry +{ + /** + * DLL next ptr + */ + struct OpListEntry *next; + + /** + * DLL prev ptr + */ + struct OpListEntry *prev; + + /** + * The testbed operation + */ + struct GNUNET_TESTBED_Operation *op; + + /** + * Depending on whether we start or stop RPS service at the peer, set this to + * #PEER_GO_ONLINE (1) or #PEER_GO_OFFLINE (-1) + */ + enum PEER_ONLINE_DELTA delta; + + /** + * Index of the regarding peer + */ + unsigned int index; +}; + +/** + * OpList DLL head + */ +static struct OpListEntry *oplist_head; + +/** + * OpList DLL tail + */ +static struct OpListEntry *oplist_tail; + + +/** + * A pending reply: A request was sent and the reply is pending. + */ +struct PendingReply +{ + /** + * DLL next,prev ptr + */ + struct PendingReply *next; + struct PendingReply *prev; + + /** + * Handle to the request we are waiting for + */ + struct GNUNET_RPS_Request_Handle *req_handle; + + /** + * The peer that requested + */ + struct RPSPeer *rps_peer; +}; + + +/** + * A pending request: A request was not made yet but is scheduled for later. + */ +struct PendingRequest +{ + /** + * DLL next,prev ptr + */ + struct PendingRequest *next; + struct PendingRequest *prev; + + /** + * Handle to the request we are waiting for + */ + struct GNUNET_SCHEDULER_Task *request_task; + + /** + * The peer that requested + */ + struct RPSPeer *rps_peer; +}; + + +/** + * Information we track for each peer. + */ +struct RPSPeer +{ + /** + * Index of the peer. + */ + unsigned int index; + + /** + * Handle for RPS connect operation. + */ + struct GNUNET_TESTBED_Operation *op; + + /** + * Handle to RPS service. + */ + struct GNUNET_RPS_Handle *rps_handle; + + /** + * ID of the peer. + */ + struct GNUNET_PeerIdentity *peer_id; + + /** + * A request handle to check for an request + */ + //struct GNUNET_RPS_Request_Handle *req_handle; + + /** + * Peer on- or offline? + */ + int online; + + /** + * Number of Peer IDs to request during the whole test + */ + unsigned int num_ids_to_request; + + /** + * Pending requests DLL + */ + struct PendingRequest *pending_req_head; + struct PendingRequest *pending_req_tail; + + /** + * Number of pending requests + */ + unsigned int num_pending_reqs; + + /** + * Pending replies DLL + */ + struct PendingReply *pending_rep_head; + struct PendingReply *pending_rep_tail; + + /** + * Number of pending replies + */ + unsigned int num_pending_reps; + + /** + * Number of received PeerIDs + */ + unsigned int num_recv_ids; + + /** + * Pending operation on that peer + */ + const struct OpListEntry *entry_op_manage; + + /** + * Testbed operation to connect to statistics service + */ + struct GNUNET_TESTBED_Operation *stat_op; + + /** + * Handle to the statistics service + */ + struct GNUNET_STATISTICS_Handle *stats_h; + + /** + * @brief flags to indicate which statistics values have been already + * collected from the statistics service. + * Used to check whether we are able to shutdown. + */ + uint32_t stat_collected_flags; + + /** + * @brief File name of the file the stats are finally written to + */ + const char *file_name_stats; + + /** + * @brief File name of the file the stats are finally written to + */ + const char *file_name_probs; + + /** + * @brief The current view + */ + struct GNUNET_PeerIdentity *cur_view; + + /** + * @brief Number of peers in the #cur_view. + */ + uint32_t cur_view_count; + + /** + * @brief Number of occurrences in other peer's view + */ + uint32_t count_in_views; + + /** + * @brief statistics values + */ + uint64_t num_rounds; + uint64_t num_blocks; + uint64_t num_blocks_many_push; + uint64_t num_blocks_no_push; + uint64_t num_blocks_no_pull; + uint64_t num_blocks_many_push_no_pull; + uint64_t num_blocks_no_push_no_pull; + uint64_t num_issued_push; + uint64_t num_issued_pull_req; + uint64_t num_issued_pull_rep; + uint64_t num_sent_push; + uint64_t num_sent_pull_req; + uint64_t num_sent_pull_rep; + uint64_t num_recv_push; + uint64_t num_recv_pull_req; + uint64_t num_recv_pull_rep; +}; + +enum STAT_TYPE +{ + STAT_TYPE_ROUNDS = 0x1, /* 1 */ + STAT_TYPE_BLOCKS = 0x2, /* 2 */ + STAT_TYPE_BLOCKS_MANY_PUSH = 0x4, /* 3 */ + STAT_TYPE_BLOCKS_NO_PUSH = 0x8, /* 4 */ + STAT_TYPE_BLOCKS_NO_PULL = 0x10, /* 5 */ + STAT_TYPE_BLOCKS_MANY_PUSH_NO_PULL = 0x20, /* 6 */ + STAT_TYPE_BLOCKS_NO_PUSH_NO_PULL = 0x40, /* 7 */ + STAT_TYPE_ISSUED_PUSH_SEND = 0x80, /* 8 */ + STAT_TYPE_ISSUED_PULL_REQ = 0x100, /* 9 */ + STAT_TYPE_ISSUED_PULL_REP = 0x200, /* 10 */ + STAT_TYPE_SENT_PUSH_SEND = 0x400, /* 11 */ + STAT_TYPE_SENT_PULL_REQ = 0x800, /* 12 */ + STAT_TYPE_SENT_PULL_REP = 0x1000, /* 13 */ + STAT_TYPE_RECV_PUSH_SEND = 0x2000, /* 14 */ + STAT_TYPE_RECV_PULL_REQ = 0x4000, /* 15 */ + STAT_TYPE_RECV_PULL_REP = 0x8000, /* 16 */ + STAT_TYPE_MAX = 0x80000000, /* 32 */ +}; + +struct STATcls +{ + struct RPSPeer *rps_peer; + enum STAT_TYPE stat_type; +}; + + +/** + * Information for all the peers. + */ +static struct RPSPeer *rps_peers; + +/** + * Peermap to get the index of a given peer ID quick. + */ +static struct GNUNET_CONTAINER_MultiPeerMap *peer_map; + +/** + * IDs of the peers. + */ +static struct GNUNET_PeerIdentity *rps_peer_ids; + +/** + * ID of the targeted peer. + */ +static struct GNUNET_PeerIdentity *target_peer; + +/** + * ID of the peer that requests for the evaluation. + */ +static struct RPSPeer *eval_peer; + +/** + * Number of online peers. + */ +static unsigned int num_peers_online; + +/** + * @brief The added sizes of the peer's views + */ +static unsigned int view_sizes; + +/** + * Return value from 'main'. + */ +static int ok; + +/** + * Identifier for the churn task that runs periodically + */ +static struct GNUNET_SCHEDULER_Task *post_test_task; + +/** + * Identifier for the churn task that runs periodically + */ +static struct GNUNET_SCHEDULER_Task *shutdown_task; + +/** + * Identifier for the churn task that runs periodically + */ +static struct GNUNET_SCHEDULER_Task *churn_task; + +/** + * Called to initialise the given RPSPeer + */ +typedef void (*InitPeer) (struct RPSPeer *rps_peer); + +/** + * @brief Called directly after connecting to the service + * + * @param rps_peer Specific peer the function is called on + * @param h the handle to the rps service + */ +typedef void (*PreTest) (struct RPSPeer *rps_peer, struct GNUNET_RPS_Handle *h); + +/** + * @brief Executes functions to test the api/service for a given peer + * + * Called from within #rps_connect_complete_cb () + * Implemented by #churn_test_cb, #profiler_cb, #mal_cb, #single_req_cb, + * #delay_req_cb, #seed_big_cb, #single_peer_seed_cb, #seed_cb, #req_cancel_cb + * + * @param rps_peer the peer the task runs on + */ +typedef void (*MainTest) (struct RPSPeer *rps_peer); + +/** + * Callback called once the requested random peers are available + */ +typedef void (*ReplyHandle) (void *cls, + uint64_t n, + const struct GNUNET_PeerIdentity *recv_peers); + +/** + * Called directly before disconnecting from the service + */ +typedef void (*PostTest) (struct RPSPeer *peer); + +/** + * Function called after disconnect to evaluate test success + */ +typedef int (*EvaluationCallback) (void); + +/** + * @brief Do we have Churn? + */ +enum OPTION_CHURN { + /** + * @brief If we have churn this is set + */ + HAVE_CHURN, + /** + * @brief If we have no churn this is set + */ + HAVE_NO_CHURN, +}; + +/** + * @brief Is it ok to quit the test before the timeout? + */ +enum OPTION_QUICK_QUIT { + /** + * @brief It is ok for the test to quit before the timeout triggers + */ + HAVE_QUICK_QUIT, + + /** + * @brief It is NOT ok for the test to quit before the timeout triggers + */ + HAVE_NO_QUICK_QUIT, +}; + +/** + * @brief Do we collect statistics at the end? + */ +enum OPTION_COLLECT_STATISTICS { + /** + * @brief We collect statistics at the end + */ + COLLECT_STATISTICS, + + /** + * @brief We do not collect statistics at the end + */ + NO_COLLECT_STATISTICS, +}; + +/** + * @brief Do we collect views during run? + */ +enum OPTION_COLLECT_VIEW { + /** + * @brief We collect view during run + */ + COLLECT_VIEW, + + /** + * @brief We do not collect the view during run + */ + NO_COLLECT_VIEW, +}; + +/** + * Structure to define a single test + */ +struct SingleTestRun +{ + /** + * Name of the test + */ + char *name; + + /** + * Called with a single peer in order to initialise that peer + */ + InitPeer init_peer; + + /** + * Called directly after connecting to the service + */ + PreTest pre_test; + + /** + * Main function for each peer + */ + MainTest main_test; + + /** + * Callback called once the requested peers are available + */ + ReplyHandle reply_handle; + + /** + * Called directly before disconnecting from the service + */ + PostTest post_test; + + /** + * Function to evaluate the test results + */ + EvaluationCallback eval_cb; + + /** + * Request interval + */ + uint32_t request_interval; + + /** + * Number of Requests to make. + */ + uint32_t num_requests; + + /** + * Run with (-out) churn + */ + enum OPTION_CHURN have_churn; + + /** + * Quit test before timeout? + */ + enum OPTION_QUICK_QUIT have_quick_quit; + + /** + * Collect statistics at the end? + */ + enum OPTION_COLLECT_STATISTICS have_collect_statistics; + + /** + * Collect view during run? + */ + enum OPTION_COLLECT_VIEW have_collect_view; + + /** + * @brief Mark which values from the statistics service to collect at the end + * of the run + */ + uint32_t stat_collect_flags; +} cur_test_run; + +/** + * Did we finish the test? + */ +static int post_test; + +/** + * Are we shutting down? + */ +static int in_shutdown; + +/** + * Append arguments to file + */ +static void +tofile_ (const char *file_name, const char *line) +{ + struct GNUNET_DISK_FileHandle *f; + /* char output_buffer[512]; */ + size_t size; + /* int size; */ + size_t size2; + + if (NULL == (f = GNUNET_DISK_file_open (file_name, + GNUNET_DISK_OPEN_APPEND | + GNUNET_DISK_OPEN_WRITE | + GNUNET_DISK_OPEN_CREATE, + GNUNET_DISK_PERM_USER_READ | + GNUNET_DISK_PERM_USER_WRITE | + GNUNET_DISK_PERM_GROUP_READ | + GNUNET_DISK_PERM_OTHER_READ))) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Not able to open file %s\n", + file_name); + return; + } + /* size = GNUNET_snprintf (output_buffer, + sizeof (output_buffer), + "%llu %s\n", + GNUNET_TIME_absolute_get ().abs_value_us, + line); + if (0 > size) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Failed to write string to buffer (size: %i)\n", + size); + return; + } */ + + size = strlen (line) * sizeof (char); + + size2 = GNUNET_DISK_file_write (f, line, size); + if (size != size2) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Unable to write to file! (Size: %lu, size2: %lu)\n", + size, + size2); + if (GNUNET_YES != GNUNET_DISK_file_close (f)) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Unable to close file\n"); + } + return; + } + + if (GNUNET_YES != GNUNET_DISK_file_close (f)) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Unable to close file\n"); + } +} + +/** + * This function is used to facilitate writing important information to disk + */ +#define tofile(file_name, ...) do {\ + char tmp_buf[512];\ + int size;\ + size = GNUNET_snprintf(tmp_buf,sizeof(tmp_buf),__VA_ARGS__);\ + if (0 > size)\ + GNUNET_log (GNUNET_ERROR_TYPE_WARNING,\ + "Failed to create tmp_buf\n");\ + else\ + tofile_(file_name,tmp_buf);\ + } while (0); + + +/** + * Write the ids and their according index in the given array to a file + * Unused + */ +/* static void +ids_to_file (char *file_name, + struct GNUNET_PeerIdentity *peer_ids, + unsigned int num_peer_ids) +{ + unsigned int i; + + for (i=0 ; i < num_peer_ids ; i++) + { + to_file (file_name, + "%u\t%s", + i, + GNUNET_i2s_full (&peer_ids[i])); + } +} */ + +/** + * Test the success of a single test + */ +static int +evaluate (void) +{ + unsigned int i; + int tmp_ok; + + tmp_ok = 1; + + for (i = 0; i < num_peers; i++) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "%u. peer [%s] received %u of %u expected peer_ids: %i\n", + i, + GNUNET_i2s (rps_peers[i].peer_id), + rps_peers[i].num_recv_ids, + rps_peers[i].num_ids_to_request, + (rps_peers[i].num_ids_to_request == rps_peers[i].num_recv_ids)); + tmp_ok &= (rps_peers[i].num_ids_to_request == rps_peers[i].num_recv_ids); + } + return tmp_ok? 0 : 1; +} + + +/** + * Creates an oplist entry and adds it to the oplist DLL + */ +static struct OpListEntry * +make_oplist_entry () +{ + struct OpListEntry *entry; + + entry = GNUNET_new (struct OpListEntry); + GNUNET_CONTAINER_DLL_insert_tail (oplist_head, oplist_tail, entry); + return entry; +} + + +/** + * @brief Checks if given peer already received its statistics value from the + * statistics service. + * + * @param rps_peer the peer to check for + * + * @return #GNUNET_YES if so + * #GNUNET_NO otherwise + */ +static int check_statistics_collect_completed_single_peer ( + const struct RPSPeer *rps_peer) +{ + if (cur_test_run.stat_collect_flags != + (cur_test_run.stat_collect_flags & + rps_peer->stat_collected_flags)) + { + return GNUNET_NO; + } + return GNUNET_YES; +} +/** + * @brief Checks if all peers already received their statistics value from the + * statistics service. + * + * @return #GNUNET_YES if so + * #GNUNET_NO otherwise + */ +static int check_statistics_collect_completed () +{ + uint32_t i; + + for (i = 0; i < num_peers; i++) + { + if (GNUNET_NO == check_statistics_collect_completed_single_peer (&rps_peers[i])) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "At least Peer %" PRIu32 " did not yet receive all statistics values\n", + i); + return GNUNET_NO; + } + } + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "All peers received their statistics values\n"); + return GNUNET_YES; +} + +/** + * Task run on timeout to shut everything down. + */ +static void +shutdown_op (void *cls) +{ + unsigned int i; + + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Shutdown task scheduled, going down.\n"); + in_shutdown = GNUNET_YES; + if (NULL != post_test_task) + { + GNUNET_SCHEDULER_cancel (post_test_task); + } + if (NULL != churn_task) + { + GNUNET_SCHEDULER_cancel (churn_task); + churn_task = NULL; + } + for (i = 0; i < num_peers; i++) + { + if (NULL != rps_peers[i].rps_handle) + { + GNUNET_RPS_disconnect (rps_peers[i].rps_handle); + } + if (NULL != rps_peers[i].op) + { + GNUNET_TESTBED_operation_done (rps_peers[i].op); + } + } +} + + +/** + * Task run on timeout to collect statistics and potentially shut down. + */ +static void +post_test_op (void *cls) +{ + unsigned int i; + + post_test_task = NULL; + post_test = GNUNET_YES; + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Post test task scheduled, going down.\n"); + if (NULL != churn_task) + { + GNUNET_SCHEDULER_cancel (churn_task); + churn_task = NULL; + } + for (i = 0; i < num_peers; i++) + { + if (NULL != rps_peers[i].op) + { + GNUNET_TESTBED_operation_done (rps_peers[i].op); + rps_peers[i].op = NULL; + } + if (NULL != cur_test_run.post_test) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Executing post_test for peer %u\n", i); + cur_test_run.post_test (&rps_peers[i]); + } + } + /* If we do not collect statistics, shut down directly */ + if (NO_COLLECT_STATISTICS == cur_test_run.have_collect_statistics || + GNUNET_YES == check_statistics_collect_completed()) + { + GNUNET_SCHEDULER_shutdown (); + } +} + + +/** + * Seed peers. + */ +static void +seed_peers (void *cls) +{ + struct RPSPeer *peer = cls; + unsigned int amount; + unsigned int i; + + // TODO if malicious don't seed mal peers + amount = round (.5 * num_peers); + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Seeding peers:\n"); + for (i = 0 ; i < amount ; i++) + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Seeding %u. peer: %s\n", + i, + GNUNET_i2s (&rps_peer_ids[i])); + + GNUNET_RPS_seed_ids (peer->rps_handle, amount, rps_peer_ids); +} + + +/** + * Get the id of peer i. + */ + void +info_cb (void *cb_cls, + struct GNUNET_TESTBED_Operation *op, + const struct GNUNET_TESTBED_PeerInformation *pinfo, + const char *emsg) +{ + struct OpListEntry *entry = (struct OpListEntry *) cb_cls; + + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + + if (NULL == pinfo || NULL != emsg) + { + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, "Got Error: %s\n", emsg); + GNUNET_TESTBED_operation_done (entry->op); + return; + } + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Peer %u is %s\n", + entry->index, + GNUNET_i2s (pinfo->result.id)); + + rps_peer_ids[entry->index] = *(pinfo->result.id); + rps_peers[entry->index].peer_id = &rps_peer_ids[entry->index]; + + GNUNET_assert (GNUNET_OK == + GNUNET_CONTAINER_multipeermap_put (peer_map, + &rps_peer_ids[entry->index], + &rps_peers[entry->index], + GNUNET_CONTAINER_MULTIHASHMAPOPTION_UNIQUE_ONLY)); + tofile ("/tmp/rps/peer_ids", + "%u\t%s\n", + entry->index, + GNUNET_i2s_full (&rps_peer_ids[entry->index])); + + GNUNET_CONTAINER_DLL_remove (oplist_head, oplist_tail, entry); + GNUNET_TESTBED_operation_done (entry->op); + GNUNET_free (entry); +} + + +/** + * Callback to be called when RPS service connect operation is completed + * + * @param cls the callback closure from functions generating an operation + * @param op the operation that has been finished + * @param ca_result the RPS service handle returned from rps_connect_adapter + * @param emsg error message in case the operation has failed; will be NULL if + * operation has executed successfully. + */ +static void +rps_connect_complete_cb (void *cls, + struct GNUNET_TESTBED_Operation *op, + void *ca_result, + const char *emsg) +{ + struct RPSPeer *rps_peer = cls; + struct GNUNET_RPS_Handle *rps = ca_result; + + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + + rps_peer->rps_handle = rps; + rps_peer->online = GNUNET_YES; + num_peers_online++; + + GNUNET_assert (op == rps_peer->op); + if (NULL != emsg) + { + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, + "Failed to connect to RPS service: %s\n", + emsg); + ok = 1; + GNUNET_SCHEDULER_shutdown (); + return; + } + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Started client successfully\n"); + + cur_test_run.main_test (rps_peer); +} + + +/** + * Adapter function called to establish a connection to + * the RPS service. + * + * @param cls closure + * @param cfg configuration of the peer to connect to; will be available until + * GNUNET_TESTBED_operation_done() is called on the operation returned + * from GNUNET_TESTBED_service_connect() + * @return service handle to return in 'op_result', NULL on error + */ +static void * +rps_connect_adapter (void *cls, + const struct GNUNET_CONFIGURATION_Handle *cfg) +{ + struct GNUNET_RPS_Handle *h; + + h = GNUNET_RPS_connect (cfg); + + if (NULL != cur_test_run.pre_test) + cur_test_run.pre_test (cls, h); + + return h; +} + +/** + * Called to open a connection to the peer's statistics + * + * @param cls peer context + * @param cfg configuration of the peer to connect to; will be available until + * GNUNET_TESTBED_operation_done() is called on the operation returned + * from GNUNET_TESTBED_service_connect() + * @return service handle to return in 'op_result', NULL on error + */ +static void * +stat_connect_adapter (void *cls, + const struct GNUNET_CONFIGURATION_Handle *cfg) +{ + struct RPSPeer *peer = cls; + + peer->stats_h = GNUNET_STATISTICS_create ("rps-profiler", cfg); + return peer->stats_h; +} + +/** + * Called to disconnect from peer's statistics service + * + * @param cls peer context + * @param op_result service handle returned from the connect adapter + */ +static void +stat_disconnect_adapter (void *cls, void *op_result) +{ + struct RPSPeer *peer = cls; + + //GNUNET_break (GNUNET_OK == GNUNET_STATISTICS_watch_cancel + // (peer->stats_h, "core", "# peers connected", + // stat_iterator, peer)); + //GNUNET_break (GNUNET_OK == GNUNET_STATISTICS_watch_cancel + // (peer->stats_h, "nse", "# peers connected", + // stat_iterator, peer)); + GNUNET_STATISTICS_destroy (op_result, GNUNET_NO); + peer->stats_h = NULL; +} + +/** + * Called after successfully opening a connection to a peer's statistics + * service; we register statistics monitoring for CORE and NSE here. + * + * @param cls the callback closure from functions generating an operation + * @param op the operation that has been finished + * @param ca_result the service handle returned from GNUNET_TESTBED_ConnectAdapter() + * @param emsg error message in case the operation has failed; will be NULL if + * operation has executed successfully. + */ +static void +stat_complete_cb (void *cls, struct GNUNET_TESTBED_Operation *op, + void *ca_result, const char *emsg ) +{ + //struct GNUNET_STATISTICS_Handle *sh = ca_result; + //struct RPSPeer *peer = (struct RPSPeer *) cls; + + if (NULL != emsg) + { + GNUNET_break (0); + return; + } + //GNUNET_break (GNUNET_OK == GNUNET_STATISTICS_watch + // (sh, "core", "# peers connected", + // stat_iterator, peer)); + //GNUNET_break (GNUNET_OK == GNUNET_STATISTICS_watch + // (sh, "nse", "# peers connected", + // stat_iterator, peer)); +} + + +/** + * Adapter function called to destroy connection to + * RPS service. + * + * @param cls closure + * @param op_result service handle returned from the connect adapter + */ +static void +rps_disconnect_adapter (void *cls, + void *op_result) +{ + struct RPSPeer *peer = cls; + struct GNUNET_RPS_Handle *h = op_result; + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "disconnect_adapter()\n"); + GNUNET_assert (NULL != peer); + GNUNET_RPS_disconnect (h); + peer->rps_handle = NULL; +} + + +/*********************************************************************** + * Definition of tests +***********************************************************************/ + +/** + * Callback to call on receipt of a reply + * + * @param cls closure + * @param n number of peers + * @param recv_peers the received peers + */ +static void +default_reply_handle (void *cls, + uint64_t n, + const struct GNUNET_PeerIdentity *recv_peers) +{ + struct RPSPeer *rps_peer; + struct PendingReply *pending_rep = (struct PendingReply *) cls; + unsigned int i; + + rps_peer = pending_rep->rps_peer; + GNUNET_CONTAINER_DLL_remove (rps_peer->pending_rep_head, + rps_peer->pending_rep_tail, + pending_rep); + rps_peer->num_pending_reps--; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "[%s] got %" PRIu64 " peers:\n", + GNUNET_i2s (rps_peer->peer_id), + n); + + for (i = 0; i < n; i++) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "%u: %s\n", + i, + GNUNET_i2s (&recv_peers[i])); + + rps_peer->num_recv_ids++; + } + + if (0 == evaluate () && HAVE_QUICK_QUIT == cur_test_run.have_quick_quit) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Test succeeded before timeout\n"); + GNUNET_assert (NULL != post_test_task); + GNUNET_SCHEDULER_cancel (post_test_task); + post_test_task = GNUNET_SCHEDULER_add_now (&post_test_op, NULL); + GNUNET_assert (NULL!= post_test_task); + } +} + +/** + * Request random peers. + */ +static void +request_peers (void *cls) +{ + struct PendingRequest *pending_req = cls; + struct RPSPeer *rps_peer; + struct PendingReply *pending_rep; + + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + return; + rps_peer = pending_req->rps_peer; + GNUNET_assert (1 <= rps_peer->num_pending_reqs); + GNUNET_CONTAINER_DLL_remove (rps_peer->pending_req_head, + rps_peer->pending_req_tail, + pending_req); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Requesting one peer\n"); + pending_rep = GNUNET_new (struct PendingReply); + pending_rep->rps_peer = rps_peer; + pending_rep->req_handle = GNUNET_RPS_request_peers (rps_peer->rps_handle, + 1, + cur_test_run.reply_handle, + pending_rep); + GNUNET_CONTAINER_DLL_insert_tail (rps_peer->pending_rep_head, + rps_peer->pending_rep_tail, + pending_rep); + rps_peer->num_pending_reps++; + rps_peer->num_pending_reqs--; +} + +static void +cancel_pending_req (struct PendingRequest *pending_req) +{ + struct RPSPeer *rps_peer; + + rps_peer = pending_req->rps_peer; + GNUNET_CONTAINER_DLL_remove (rps_peer->pending_req_head, + rps_peer->pending_req_tail, + pending_req); + rps_peer->num_pending_reqs--; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Cancelling pending request\n"); + GNUNET_SCHEDULER_cancel (pending_req->request_task); + GNUNET_free (pending_req); +} + +static void +cancel_request (struct PendingReply *pending_rep) +{ + struct RPSPeer *rps_peer; + + rps_peer = pending_rep->rps_peer; + GNUNET_CONTAINER_DLL_remove (rps_peer->pending_rep_head, + rps_peer->pending_rep_tail, + pending_rep); + rps_peer->num_pending_reps--; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Cancelling request\n"); + GNUNET_RPS_request_cancel (pending_rep->req_handle); + GNUNET_free (pending_rep); +} + + +/** + * Schedule requests for peer @a rps_peer that have neither been scheduled, nor + * issued, nor replied + */ +void +schedule_missing_requests (struct RPSPeer *rps_peer) +{ + unsigned int i; + struct PendingRequest *pending_req; + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Scheduling %u - %u missing requests\n", + rps_peer->num_ids_to_request, + rps_peer->num_pending_reqs + rps_peer->num_pending_reps); + GNUNET_assert (rps_peer->num_pending_reqs + rps_peer->num_pending_reps <= + rps_peer->num_ids_to_request); + for (i = rps_peer->num_pending_reqs + rps_peer->num_pending_reps; + i < rps_peer->num_ids_to_request; i++) + { + pending_req = GNUNET_new (struct PendingRequest); + pending_req->rps_peer = rps_peer; + pending_req->request_task = GNUNET_SCHEDULER_add_delayed ( + GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, + cur_test_run.request_interval * i), + request_peers, + pending_req); + GNUNET_CONTAINER_DLL_insert_tail (rps_peer->pending_req_head, + rps_peer->pending_req_tail, + pending_req); + rps_peer->num_pending_reqs++; + } +} + +void +cancel_pending_req_rep (struct RPSPeer *rps_peer) +{ + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Cancelling all (pending) requests.\n"); + while (NULL != rps_peer->pending_req_head) + cancel_pending_req (rps_peer->pending_req_head); + GNUNET_assert (0 == rps_peer->num_pending_reqs); + while (NULL != rps_peer->pending_rep_head) + cancel_request (rps_peer->pending_rep_head); + GNUNET_assert (0 == rps_peer->num_pending_reps); +} + +/*********************************** + * MALICIOUS +***********************************/ + +/** + * Initialise only non-mal RPSPeers + */ +static void mal_init_peer (struct RPSPeer *rps_peer) +{ + if (rps_peer->index >= round (portion * num_peers)) + rps_peer->num_ids_to_request = 1; +} + + +/** + * @brief Set peers to (non-)malicious before execution + * + * Of signature #PreTest + * + * @param rps_peer the peer to set (non-) malicious + * @param h the handle to the service + */ +static void +mal_pre (struct RPSPeer *rps_peer, struct GNUNET_RPS_Handle *h) +{ + #ifdef ENABLE_MALICIOUS + uint32_t num_mal_peers; + + GNUNET_assert ( (1 >= portion) && + (0 < portion) ); + num_mal_peers = round (portion * num_peers); + + if (rps_peer->index < num_mal_peers) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "%u. peer [%s] of %" PRIu32 " malicious peers turning malicious\n", + rps_peer->index, + GNUNET_i2s (rps_peer->peer_id), + num_mal_peers); + + GNUNET_RPS_act_malicious (h, mal_type, num_mal_peers, + rps_peer_ids, target_peer); + } + #endif /* ENABLE_MALICIOUS */ +} + +static void +mal_cb (struct RPSPeer *rps_peer) +{ + uint32_t num_mal_peers; + + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + + #ifdef ENABLE_MALICIOUS + GNUNET_assert ( (1 >= portion) && + (0 < portion) ); + num_mal_peers = round (portion * num_peers); + + if (rps_peer->index >= num_mal_peers) + { /* It's useless to ask a malicious peer about a random sample - + it's not sampling */ + GNUNET_SCHEDULER_add_delayed (GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 2), + seed_peers, rps_peer); + schedule_missing_requests (rps_peer); + } + #endif /* ENABLE_MALICIOUS */ +} + +/*********************************** + * CHURN +***********************************/ + +static void +churn (void *cls); + +/** + * @brief Starts churn + * + * Has signature of #MainTest + * + * This is not implemented too nicely as this is called for each peer, but we + * only need to call it once. (Yes we check that we only schedule the task + * once.) + * + * @param rps_peer The peer it's called for + */ +static void +churn_test_cb (struct RPSPeer *rps_peer) +{ + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + + /* Start churn */ + if (HAVE_CHURN == cur_test_run.have_churn && NULL == churn_task) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Starting churn task\n"); + churn_task = GNUNET_SCHEDULER_add_delayed ( + GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 5), + churn, + NULL); + } else { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Not starting churn task\n"); + } + + schedule_missing_requests (rps_peer); +} + +/*********************************** + * PROFILER +***********************************/ + +/** + * Callback to be called when RPS service is started or stopped at peers + * + * @param cls NULL + * @param op the operation handle + * @param emsg NULL on success; otherwise an error description + */ +static void +churn_cb (void *cls, + struct GNUNET_TESTBED_Operation *op, + const char *emsg) +{ + // FIXME + struct OpListEntry *entry = cls; + + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + + GNUNET_TESTBED_operation_done (entry->op); + if (NULL != emsg) + { + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, "Failed to start/stop RPS at a peer\n"); + GNUNET_SCHEDULER_shutdown (); + return; + } + GNUNET_assert (0 != entry->delta); + + num_peers_online += entry->delta; + + if (PEER_GO_OFFLINE == entry->delta) + { /* Peer hopefully just went offline */ + if (GNUNET_YES != rps_peers[entry->index].online) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "peer %s was expected to go offline but is still marked as online\n", + GNUNET_i2s (rps_peers[entry->index].peer_id)); + GNUNET_break (0); + } + else + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "peer %s probably went offline as expected\n", + GNUNET_i2s (rps_peers[entry->index].peer_id)); + } + rps_peers[entry->index].online = GNUNET_NO; + } + + else if (PEER_GO_ONLINE < entry->delta) + { /* Peer hopefully just went online */ + if (GNUNET_NO != rps_peers[entry->index].online) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "peer %s was expected to go online but is still marked as offline\n", + GNUNET_i2s (rps_peers[entry->index].peer_id)); + GNUNET_break (0); + } + else + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "peer %s probably went online as expected\n", + GNUNET_i2s (rps_peers[entry->index].peer_id)); + if (NULL != cur_test_run.pre_test) + { + cur_test_run.pre_test (&rps_peers[entry->index], + rps_peers[entry->index].rps_handle); + schedule_missing_requests (&rps_peers[entry->index]); + } + } + rps_peers[entry->index].online = GNUNET_YES; + } + else + { + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, + "Invalid value for delta: %i\n", entry->delta); + GNUNET_break (0); + } + + GNUNET_CONTAINER_DLL_remove (oplist_head, oplist_tail, entry); + rps_peers[entry->index].entry_op_manage = NULL; + GNUNET_free (entry); + //if (num_peers_in_round[current_round] == peers_running) + // run_round (); +} + +/** + * @brief Set the rps-service up or down for a specific peer + * + * @param i index of action + * @param j index of peer + * @param delta (#PEER_ONLINE_DELTA) down (-1) or up (1) + * @param prob_go_on_off the probability of the action + */ +static void +manage_service_wrapper (unsigned int i, unsigned int j, + enum PEER_ONLINE_DELTA delta, + double prob_go_on_off) +{ + struct OpListEntry *entry = NULL; + uint32_t prob; + + /* make sure that management operation is not already scheduled */ + if (NULL != rps_peers[j].entry_op_manage) + { + return; + } + + prob = GNUNET_CRYPTO_random_u32 (GNUNET_CRYPTO_QUALITY_WEAK, + UINT32_MAX); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "%u. selected peer (%u: %s) is %s.\n", + i, + j, + GNUNET_i2s (rps_peers[j].peer_id), + (PEER_GO_ONLINE == delta) ? "online" : "offline"); + if (prob < prob_go_on_off * UINT32_MAX) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "%s goes %s\n", + GNUNET_i2s (rps_peers[j].peer_id), + (PEER_GO_OFFLINE == delta) ? "offline" : "online"); + + if (PEER_GO_OFFLINE == delta) + cancel_pending_req_rep (&rps_peers[j]); + entry = make_oplist_entry (); + entry->delta = delta; + entry->index = j; + entry->op = GNUNET_TESTBED_peer_manage_service (NULL, + testbed_peers[j], + "rps", + &churn_cb, + entry, + (PEER_GO_OFFLINE == delta) ? 0 : 1); + rps_peers[j].entry_op_manage = entry; + } +} + + +static void +churn (void *cls) +{ + unsigned int i; + unsigned int j; + double portion_online; + unsigned int *permut; + double prob_go_offline; + double portion_go_online; + double portion_go_offline; + + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Churn function executing\n"); + + churn_task = NULL; /* Should be invalid by now */ + + /* Compute the probability for an online peer to go offline + * this round */ + portion_online = num_peers_online * 1.0 / num_peers; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Portion online: %f\n", + portion_online); + portion_go_online = ((1 - portion_online) * .5 * .66); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Portion that should go online: %f\n", + portion_go_online); + portion_go_offline = (portion_online + portion_go_online) - .75; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Portion that probably goes offline: %f\n", + portion_go_offline); + prob_go_offline = portion_go_offline / (portion_online * .5); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Probability of a selected online peer to go offline: %f\n", + prob_go_offline); + + permut = GNUNET_CRYPTO_random_permute (GNUNET_CRYPTO_QUALITY_WEAK, + (unsigned int) num_peers); + + /* Go over 50% randomly chosen peers */ + for (i = 0; i < .5 * num_peers; i++) + { + j = permut[i]; + + /* If online, shut down with certain probability */ + if (GNUNET_YES == rps_peers[j].online) + { + manage_service_wrapper (i, j, -1, prob_go_offline); + } + + /* If offline, restart with certain probability */ + else if (GNUNET_NO == rps_peers[j].online) + { + manage_service_wrapper (i, j, 1, 0.66); + } + } + + GNUNET_free (permut); + + churn_task = GNUNET_SCHEDULER_add_delayed ( + GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 2), + churn, + NULL); +} + + +/** + * Initialise given RPSPeer + */ +static void profiler_init_peer (struct RPSPeer *rps_peer) +{ + if (num_peers - 1 == rps_peer->index) + { + rps_peer->num_ids_to_request = cur_test_run.num_requests; + } else { + rps_peer->num_ids_to_request = 0; + } + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "peer shall request %i peers\n", + rps_peer->num_ids_to_request); +} + + +/** + * Callback to call on receipt of a reply + * + * @param cls closure + * @param n number of peers + * @param recv_peers the received peers + */ +static void +profiler_reply_handle (void *cls, + uint64_t n, + const struct GNUNET_PeerIdentity *recv_peers) +{ + struct RPSPeer *rps_peer; + struct RPSPeer *rcv_rps_peer; + char *file_name; + char *file_name_dh; + char *file_name_dhr; + unsigned int i; + struct PendingReply *pending_rep = (struct PendingReply *) cls; + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "profiler_reply_handle()\n"); + rps_peer = pending_rep->rps_peer; + file_name = "/tmp/rps/received_ids"; + file_name_dh = "/tmp/rps/diehard_input"; + file_name_dhr = "/tmp/rps/diehard_input_raw"; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "[%s] got %" PRIu64 " peers:\n", + GNUNET_i2s (rps_peer->peer_id), + n); + for (i = 0; i < n; i++) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "%u: %s\n", + i, + GNUNET_i2s (&recv_peers[i])); + tofile (file_name, + "%s\n", + GNUNET_i2s_full (&recv_peers[i])); + rcv_rps_peer = GNUNET_CONTAINER_multipeermap_get (peer_map, &recv_peers[i]); + GNUNET_assert (NULL != rcv_rps_peer); + tofile (file_name_dh, + "%" PRIu32 "\n", + (uint32_t) rcv_rps_peer->index); + to_file_raw (file_name_dhr, + &rcv_rps_peer->index, + sizeof (uint32_t)); + } + default_reply_handle (cls, n, recv_peers); +} + + +static void +profiler_cb (struct RPSPeer *rps_peer) +{ + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) + { + return; + } + + /* Start churn */ + if (HAVE_CHURN == cur_test_run.have_churn && NULL == churn_task) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Starting churn task\n"); + churn_task = GNUNET_SCHEDULER_add_delayed ( + GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 5), + churn, + NULL); + } else { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Not starting churn task\n"); + } + + /* Only request peer ids at one peer. + * (It's the before-last because last one is target of the focussed attack.) + */ + if (eval_peer == rps_peer) + schedule_missing_requests (rps_peer); +} + +/** + * Function called from #profiler_eval with a filename. + * + * @param cls closure + * @param filename complete filename (absolute path) + * @return #GNUNET_OK to continue to iterate, + * #GNUNET_NO to stop iteration with no error, + * #GNUNET_SYSERR to abort iteration with error! + */ +int +file_name_cb (void *cls, const char *filename) +{ + if (NULL != strstr (filename, "sampler_el")) + { + struct RPS_SamplerElement *s_elem; + struct GNUNET_CRYPTO_AuthKey auth_key; + const char *key_char; + uint32_t i; + + key_char = filename + 20; /* Length of "/tmp/rps/sampler_el-" */ + tofile (filename, "--------------------------\n"); + + auth_key = string_to_auth_key (key_char); + s_elem = RPS_sampler_elem_create (); + RPS_sampler_elem_set (s_elem, auth_key); + + for (i = 0; i < num_peers; i++) + { + RPS_sampler_elem_next (s_elem, &rps_peer_ids[i]); + } + RPS_sampler_elem_destroy (s_elem); + } + return GNUNET_OK; +} + +/** + * This is run after the test finished. + * + * Compute all perfect samples. + */ +int +profiler_eval (void) +{ + /* Compute perfect sample for each sampler element */ + if (-1 == GNUNET_DISK_directory_scan ("/tmp/rps/", file_name_cb, NULL)) + { + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, "Scan of directory failed\n"); + } + + return evaluate (); +} + +static uint32_t fac (uint32_t x) +{ + if (1 >= x) + { + return x; + } + return x * fac (x - 1); +} + +static uint32_t binom (uint32_t n, uint32_t k) +{ + //GNUNET_assert (n >= k); + if (k > n) return 0; + if (0 > n) return 0; + if (0 > k) return 0; + if (0 == k) return 1; + return fac (n) + / + fac(k) * fac(n - k); +} + +/** + * @brief is b in view of a? + * + * @param a + * @param b + * + * @return + */ +static int is_in_view (uint32_t a, uint32_t b) +{ + uint32_t i; + for (i = 0; i < rps_peers[a].cur_view_count; i++) + { + if (0 == memcmp (rps_peers[b].peer_id, + &rps_peers[a].cur_view[i], + sizeof (struct GNUNET_PeerIdentity))) + { + return GNUNET_YES; + } + } + return GNUNET_NO; +} + +static uint32_t get_idx_of_pid (const struct GNUNET_PeerIdentity *pid) +{ + uint32_t i; + + for (i = 0; i < num_peers; i++) + { + if (0 == memcmp (pid, + rps_peers[i].peer_id, + sizeof (struct GNUNET_PeerIdentity))) + { + return i; + } + } + //return 0; /* Should not happen - make compiler happy */ + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, + "No known _PeerIdentity %s!\n", + GNUNET_i2s_full (pid)); + GNUNET_assert (0); +} + +/** + * @brief Counts number of peers in view of a that have b in their view + * + * @param a + * @param uint32_tb + * + * @return + */ +static uint32_t count_containing_views (uint32_t a, uint32_t b) +{ + uint32_t i; + uint32_t peer_idx; + uint32_t count = 0; + + for (i = 0; i < rps_peers[a].cur_view_count; i++) + { + peer_idx = get_idx_of_pid (&rps_peers[a].cur_view[i]); + if (GNUNET_YES == is_in_view (peer_idx, b)) + { + count++; + } + } + return count; +} + +/** + * @brief Computes the probability for each other peer to be selected by the + * sampling process based on the views of all peers + * + * @param peer_idx index of the peer that is about to sample + */ +static void compute_probabilities (uint32_t peer_idx) +{ + //double probs[num_peers] = { 0 }; + double probs[num_peers]; + size_t probs_as_str_size = (num_peers * 10 + 1) * sizeof (char); + char *probs_as_str = GNUNET_malloc (probs_as_str_size); + char *probs_as_str_cpy; + uint32_t i; + double prob_push; + double prob_pull; + uint32_t view_size; + uint32_t cont_views; + uint32_t number_of_being_in_pull_events; + int tmp; + uint32_t count_non_zero_prob = 0; + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Computing probabilities for peer %" PRIu32 "\n", peer_idx); + /* Firstly without knowledge of old views */ + for (i = 0; i < num_peers; i++) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\tfor peer %" PRIu32 ":\n", i); + view_size = rps_peers[i].cur_view_count; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\tview_size: %" PRIu32 "\n", view_size); + /* For peer i the probability of being sampled is + * evenly distributed among all possibly observed peers. */ + /* We could have observed a peer in three cases: + * 1. peer sent a push + * 2. peer was contained in a pull reply + * 3. peer was in history (sampler) - ignored for now */ + /* 1. Probability of having received a push from peer i */ + if ((GNUNET_YES == is_in_view (i, peer_idx)) && + (1 <= (0.45 * view_size))) + { + prob_push = 1.0 * binom (0.45 * view_size, 1) + / + binom (view_size, 0.45 * view_size); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\t%" PRIu32 " is in %" PRIu32 "'s view, prob: %f\n", + peer_idx, + i, + prob_push); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\tposs choices from view: %" PRIu32 ", containing i: %" PRIu32 "\n", + binom (view_size, 0.45 * view_size), + binom (0.45 * view_size, 1)); + } else { + prob_push = 0; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\t%" PRIu32 " is not in %" PRIu32 "'s view, prob: 0\n", + peer_idx, + i); + } + /* 2. Probability of peer i being contained in pulls */ + view_size = rps_peers[peer_idx].cur_view_count; + cont_views = count_containing_views (peer_idx, i); + number_of_being_in_pull_events = + (binom (view_size, 0.45 * view_size) - + binom (view_size - cont_views, 0.45 * view_size)); + if (0 != number_of_being_in_pull_events) + { + prob_pull = number_of_being_in_pull_events + / + (1.0 * binom (view_size, 0.45 * view_size)); + } else + { + prob_pull = 0; + } + probs[i] = prob_push + prob_pull - (prob_push * prob_pull); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\t%" PRIu32 " has %" PRIu32 " of %" PRIu32 + " peers in its view who know %" PRIu32 " prob: %f\n", + peer_idx, + cont_views, + view_size, + i, + prob_pull); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\tnumber of possible pull combinations: %" PRIu32 "\n", + binom (view_size, 0.45 * view_size)); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\tnumber of possible pull combinations without %" PRIu32 + ": %" PRIu32 "\n", + i, + binom (view_size - cont_views, 0.45 * view_size)); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t\tnumber of possible pull combinations with %" PRIu32 + ": %" PRIu32 "\n", + i, + number_of_being_in_pull_events); + + if (0 != probs[i]) count_non_zero_prob++; + } + /* normalize */ + if (0 != count_non_zero_prob) + { + for (i = 0; i < num_peers; i++) + { + probs[i] = probs[i] * (1.0 / count_non_zero_prob); + } + } else { + for (i = 0; i < num_peers; i++) + { + probs[i] = 0; + } + } + /* str repr */ + for (i = 0; i < num_peers; i++) + { + probs_as_str_cpy = GNUNET_strndup (probs_as_str, probs_as_str_size); + tmp = GNUNET_snprintf (probs_as_str, + probs_as_str_size, + "%s %7.6f", probs_as_str_cpy, probs[i]); + GNUNET_free (probs_as_str_cpy); + GNUNET_assert (0 <= tmp); + } + + to_file_w_len (rps_peers[peer_idx].file_name_probs, + probs_as_str_size, + probs_as_str); + GNUNET_free (probs_as_str); +} + +/** + * @brief This counts the number of peers in which views a given peer occurs. + * + * It also stores this value in the rps peer. + * + * @param peer_idx the index of the peer to count the representation + * + * @return the number of occurrences + */ +static uint32_t count_peer_in_views_2 (uint32_t peer_idx) +{ + uint32_t i, j; + uint32_t count = 0; + + for (i = 0; i < num_peers; i++) /* Peer in which view is counted */ + { + for (j = 0; j < rps_peers[i].cur_view_count; j++) /* entry in view */ + { + if (0 == memcmp (rps_peers[peer_idx].peer_id, + &rps_peers[i].cur_view[j], + sizeof (struct GNUNET_PeerIdentity))) + { + count++; + break; + } + } + } + rps_peers[peer_idx].count_in_views = count; + return count; +} + +static uint32_t cumulated_view_sizes () +{ + uint32_t i; + + view_sizes = 0; + for (i = 0; i < num_peers; i++) /* Peer in which view is counted */ + { + view_sizes += rps_peers[i].cur_view_count; + } + return view_sizes; +} + +static void count_peer_in_views (uint32_t *count_peers) +{ + uint32_t i, j; + + for (i = 0; i < num_peers; i++) /* Peer in which view is counted */ + { + for (j = 0; j < rps_peers[i].cur_view_count; j++) /* entry in view */ + { + if (0 == memcmp (rps_peers[i].peer_id, + &rps_peers[i].cur_view[j], + sizeof (struct GNUNET_PeerIdentity))) + { + count_peers[i]++; + } + } + } +} + +void compute_diversity () +{ + uint32_t i; + /* ith entry represents the numer of occurrences in other peer's views */ + uint32_t *count_peers = GNUNET_new_array (num_peers, uint32_t); + uint32_t views_total_size; + double expected; + /* deviation from expected number of peers */ + double *deviation = GNUNET_new_array (num_peers, double); + + views_total_size = 0; + expected = 0; + + /* For each peer count its representation in other peer's views*/ + for (i = 0; i < num_peers; i++) /* Peer to count */ + { + views_total_size += rps_peers[i].cur_view_count; + count_peer_in_views (count_peers); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Counted representation of %" PRIu32 "th peer [%s]: %" PRIu32"\n", + i, + GNUNET_i2s (rps_peers[i].peer_id), + count_peers[i]); + } + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "size of all views combined: %" PRIu32 "\n", + views_total_size); + expected = ((double) 1/num_peers) * views_total_size; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Expected number of occurrences of each peer in all views: %f\n", + expected); + for (i = 0; i < num_peers; i++) /* Peer to count */ + { + deviation[i] = expected - count_peers[i]; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Deviation from expectation: %f\n", deviation[i]); + } + GNUNET_free (count_peers); + GNUNET_free (deviation); +} + +void print_view_sizes() +{ + uint32_t i; + + for (i = 0; i < num_peers; i++) /* Peer to count */ + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "View size of %" PRIu32 ". [%s] is %" PRIu32 "\n", + i, + GNUNET_i2s (rps_peers[i].peer_id), + rps_peers[i].cur_view_count); + } +} + +void all_views_updated_cb() +{ + compute_diversity(); + print_view_sizes(); +} + +void view_update_cb (void *cls, + uint64_t view_size, + const struct GNUNET_PeerIdentity *peers) +{ + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "View was updated (%" PRIu64 ")\n", view_size); + struct RPSPeer *rps_peer = (struct RPSPeer *) cls; + to_file ("/tmp/rps/view_sizes.txt", + "%" PRIu64 " %" PRIu32 "", + rps_peer->index, + view_size); + for (int i = 0; i < view_size; i++) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "\t%s\n", GNUNET_i2s (&peers[i])); + } + GNUNET_array_grow (rps_peer->cur_view, + rps_peer->cur_view_count, + view_size); + //*rps_peer->cur_view = *peers; + GNUNET_memcpy (rps_peer->cur_view, + peers, + view_size * sizeof (struct GNUNET_PeerIdentity)); + to_file ("/tmp/rps/count_in_views.txt", + "%" PRIu64 " %" PRIu32 "", + rps_peer->index, + count_peer_in_views_2 (rps_peer->index)); + cumulated_view_sizes(); + if (0 != view_size) + { + to_file ("/tmp/rps/repr.txt", + "%" PRIu64 /* index */ + " %" PRIu32 /* occurrence in views */ + " %" PRIu32 /* view sizes */ + " %f" /* fraction of repr in views */ + " %f" /* average view size */ + " %f" /* prob of occurrence in view slot */ + " %f" "", /* exp frac of repr in views */ + rps_peer->index, + count_peer_in_views_2 (rps_peer->index), + view_sizes, + count_peer_in_views_2 (rps_peer->index) / (view_size * 1.0), /* fraction of representation in views */ + view_sizes / (view_size * 1.0), /* average view size */ + 1.0 /view_size, /* prob of occurrence in view slot */ + (1.0/view_size) * (view_sizes/view_size) /* expected fraction of repr in views */ + ); + } + compute_probabilities (rps_peer->index); + all_views_updated_cb(); +} + +static void +pre_profiler (struct RPSPeer *rps_peer, struct GNUNET_RPS_Handle *h) +{ + rps_peer->file_name_probs = + store_prefix_file_name (rps_peer->peer_id, "probs"); + GNUNET_RPS_view_request (h, 0, view_update_cb, rps_peer); +} + +void write_final_stats (void){ + uint32_t i; + + for (i = 0; i < num_peers; i++) + { + to_file ("/tmp/rps/final_stats.dat", + "%" PRIu32 " " /* index */ + "%s %" /* id */ + PRIu64 " %" /* rounds */ + PRIu64 " %" PRIu64 " %" PRIu64 " %" PRIu64 " %" PRIu64 " %" PRIu64 " %" /* blocking */ + PRIu64 " %" PRIu64 " %" PRIu64 " %" /* issued */ + PRIu64 " %" PRIu64 " %" PRIu64 " %" /* sent */ + PRIu64 " %" PRIu64 " %" PRIu64 /* recv */, + i, + GNUNET_i2s (rps_peers[i].peer_id), + rps_peers[i].num_rounds, + rps_peers[i].num_blocks, + rps_peers[i].num_blocks_many_push, + rps_peers[i].num_blocks_no_push, + rps_peers[i].num_blocks_no_pull, + rps_peers[i].num_blocks_many_push_no_pull, + rps_peers[i].num_blocks_no_push_no_pull, + rps_peers[i].num_issued_push, + rps_peers[i].num_issued_pull_req, + rps_peers[i].num_issued_pull_rep, + rps_peers[i].num_sent_push, + rps_peers[i].num_sent_pull_req, + rps_peers[i].num_sent_pull_rep, + rps_peers[i].num_recv_push, + rps_peers[i].num_recv_pull_req, + rps_peers[i].num_recv_pull_rep); + } +} + +/** + * Continuation called by #GNUNET_STATISTICS_get() functions. + * + * Remembers that this specific statistics value was received for this peer. + * Checks whether all peers received their statistics yet. + * Issues the shutdown. + * + * @param cls closure + * @param success #GNUNET_OK if statistics were + * successfully obtained, #GNUNET_SYSERR if not. + */ +void +post_test_shutdown_ready_cb (void *cls, + int success) +{ + struct STATcls *stat_cls = (struct STATcls *) cls; + struct RPSPeer *rps_peer = stat_cls->rps_peer; + if (GNUNET_OK == success) + { + /* set flag that we we got the value */ + rps_peer->stat_collected_flags |= stat_cls->stat_type; + } else { + GNUNET_log (GNUNET_ERROR_TYPE_ERROR, + "Peer %u did not receive statistics value\n", + rps_peer->index); + GNUNET_free (stat_cls); + GNUNET_break (0); + } + + if (NULL != rps_peer->stat_op && + GNUNET_YES == check_statistics_collect_completed_single_peer (rps_peer)) + { + GNUNET_TESTBED_operation_done (rps_peer->stat_op); + } + + write_final_stats (); + if (GNUNET_YES == check_statistics_collect_completed()) + { + //write_final_stats (); + GNUNET_free (stat_cls); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Shutting down\n"); + GNUNET_SCHEDULER_shutdown (); + } else { + GNUNET_free (stat_cls); + } +} + +/** + * @brief Converts string representation to the corresponding #STAT_TYPE enum. + * + * @param stat_str string representation of statistics specifier + * + * @return corresponding enum + */ +enum STAT_TYPE stat_str_2_type (const char *stat_str) +{ + if (0 == strncmp ("# rounds blocked - no pull replies", stat_str, strlen ("# rounds blocked - no pull replies"))) + { + return STAT_TYPE_BLOCKS_NO_PULL; + } + else if (0 == strncmp ("# rounds blocked - too many pushes, no pull replies", stat_str, strlen ("# rounds blocked - too many pushes, no pull replies"))) + { + return STAT_TYPE_BLOCKS_MANY_PUSH_NO_PULL; + } + else if (0 == strncmp ("# rounds blocked - too many pushes", stat_str, strlen ("# rounds blocked - too many pushes"))) + { + return STAT_TYPE_BLOCKS_MANY_PUSH; + } + else if (0 == strncmp ("# rounds blocked - no pushes, no pull replies", stat_str, strlen ("# rounds blocked - no pushes, no pull replies"))) + { + return STAT_TYPE_BLOCKS_NO_PUSH_NO_PULL; + } + else if (0 == strncmp ("# rounds blocked - no pushes", stat_str, strlen ("# rounds blocked - no pushes"))) + { + return STAT_TYPE_BLOCKS_NO_PUSH; + } + else if (0 == strncmp ("# rounds blocked", stat_str, strlen ("# rounds blocked"))) + { + return STAT_TYPE_BLOCKS; + } + else if (0 == strncmp ("# rounds", stat_str, strlen ("# rounds"))) + { + return STAT_TYPE_ROUNDS; + } + else if (0 == strncmp ("# push send issued", stat_str, strlen ("# push send issued"))) + { + return STAT_TYPE_ISSUED_PUSH_SEND; + } + else if (0 == strncmp ("# pull request send issued", stat_str, strlen ("# pull request send issued"))) + { + return STAT_TYPE_ISSUED_PULL_REQ; + } + else if (0 == strncmp ("# pull reply send issued", stat_str, strlen ("# pull reply send issued"))) + { + return STAT_TYPE_ISSUED_PULL_REP; + } + else if (0 == strncmp ("# pushes sent", stat_str, strlen ("# pushes sent"))) + { + return STAT_TYPE_SENT_PUSH_SEND; + } + else if (0 == strncmp ("# pull requests sent", stat_str, strlen ("# pull requests sent"))) + { + return STAT_TYPE_SENT_PULL_REQ; + } + else if (0 == strncmp ("# pull replys sent", stat_str, strlen ("# pull replys sent"))) + { + return STAT_TYPE_SENT_PULL_REP; + } + else if (0 == strncmp ("# push message received", stat_str, strlen ("# push message received"))) + { + return STAT_TYPE_RECV_PUSH_SEND; + } + else if (0 == strncmp ("# pull request message received", stat_str, strlen ("# pull request message received"))) + { + return STAT_TYPE_RECV_PULL_REQ; + } + else if (0 == strncmp ("# pull reply messages received", stat_str, strlen ("# pull reply messages received"))) + { + return STAT_TYPE_RECV_PULL_REP; + } + return STAT_TYPE_MAX; +} + + +/** + * @brief Converts #STAT_TYPE enum to the equivalent string representation that + * is stored with the statistics service. + * + * @param stat_type #STAT_TYPE enum + * + * @return string representation that matches statistics value + */ +char* stat_type_2_str (enum STAT_TYPE stat_type) +{ + switch (stat_type) + { + case STAT_TYPE_ROUNDS: + return "# rounds"; + case STAT_TYPE_BLOCKS: + return "# rounds blocked"; + case STAT_TYPE_BLOCKS_MANY_PUSH: + return "# rounds blocked - too many pushes"; + case STAT_TYPE_BLOCKS_NO_PUSH: + return "# rounds blocked - no pushes"; + case STAT_TYPE_BLOCKS_NO_PULL: + return "# rounds blocked - no pull replies"; + case STAT_TYPE_BLOCKS_MANY_PUSH_NO_PULL: + return "# rounds blocked - too many pushes, no pull replies"; + case STAT_TYPE_BLOCKS_NO_PUSH_NO_PULL: + return "# rounds blocked - no pushes, no pull replies"; + case STAT_TYPE_ISSUED_PUSH_SEND: + return "# push send issued"; + case STAT_TYPE_ISSUED_PULL_REQ: + return "# pull request send issued"; + case STAT_TYPE_ISSUED_PULL_REP: + return "# pull reply send issued"; + case STAT_TYPE_SENT_PUSH_SEND: + return "# pushes sent"; + case STAT_TYPE_SENT_PULL_REQ: + return "# pull requests sent"; + case STAT_TYPE_SENT_PULL_REP: + return "# pull replys sent"; + case STAT_TYPE_RECV_PUSH_SEND: + return "# push message received"; + case STAT_TYPE_RECV_PULL_REQ: + return "# pull request message received"; + case STAT_TYPE_RECV_PULL_REP: + return "# pull reply messages received"; + case STAT_TYPE_MAX: + default: + return "ERROR"; + ; + } +} + +/** + * Callback function to process statistic values. + * + * @param cls closure + * @param subsystem name of subsystem that created the statistic + * @param name the name of the datum + * @param value the current value + * @param is_persistent #GNUNET_YES if the value is persistent, #GNUNET_NO if not + * @return #GNUNET_OK to continue, #GNUNET_SYSERR to abort iteration + */ +int +stat_iterator (void *cls, + const char *subsystem, + const char *name, + uint64_t value, + int is_persistent) +{ + const struct STATcls *stat_cls = (const struct STATcls *) cls; + struct RPSPeer *rps_peer = (struct RPSPeer *) stat_cls->rps_peer; + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Got stat value: %s - %" PRIu64 "\n", + //stat_type_2_str (stat_cls->stat_type), + name, + value); + to_file (rps_peer->file_name_stats, + "%s: %" PRIu64 "\n", + name, + value); + switch (stat_str_2_type (name)) + { + case STAT_TYPE_ROUNDS: + rps_peer->num_rounds = value; + break; + case STAT_TYPE_BLOCKS: + rps_peer->num_blocks = value; + break; + case STAT_TYPE_BLOCKS_MANY_PUSH: + rps_peer->num_blocks_many_push = value; + break; + case STAT_TYPE_BLOCKS_NO_PUSH: + rps_peer->num_blocks_no_push = value; + break; + case STAT_TYPE_BLOCKS_NO_PULL: + rps_peer->num_blocks_no_pull = value; + break; + case STAT_TYPE_BLOCKS_MANY_PUSH_NO_PULL: + rps_peer->num_blocks_many_push_no_pull = value; + break; + case STAT_TYPE_BLOCKS_NO_PUSH_NO_PULL: + rps_peer->num_blocks_no_push_no_pull = value; + break; + case STAT_TYPE_ISSUED_PUSH_SEND: + rps_peer->num_issued_push = value; + break; + case STAT_TYPE_ISSUED_PULL_REQ: + rps_peer->num_issued_pull_req = value; + break; + case STAT_TYPE_ISSUED_PULL_REP: + rps_peer->num_issued_pull_rep = value; + break; + case STAT_TYPE_SENT_PUSH_SEND: + rps_peer->num_sent_push = value; + break; + case STAT_TYPE_SENT_PULL_REQ: + rps_peer->num_sent_pull_req = value; + break; + case STAT_TYPE_SENT_PULL_REP: + rps_peer->num_sent_pull_rep = value; + break; + case STAT_TYPE_RECV_PUSH_SEND: + rps_peer->num_recv_push = value; + break; + case STAT_TYPE_RECV_PULL_REQ: + rps_peer->num_recv_pull_req = value; + break; + case STAT_TYPE_RECV_PULL_REP: + rps_peer->num_recv_pull_rep = value; + break; + case STAT_TYPE_MAX: + default: + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Unknown statistics string: %s\n", + name); + break; + } + return GNUNET_OK; +} + +void post_profiler (struct RPSPeer *rps_peer) +{ + if (COLLECT_STATISTICS != cur_test_run.have_collect_statistics) + { + return; + } + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Going to request statistic values with mask 0x%" PRIx32 "\n", + cur_test_run.stat_collect_flags); + + struct STATcls *stat_cls; + uint32_t stat_type; + for (stat_type = STAT_TYPE_ROUNDS; + stat_type < STAT_TYPE_MAX; + stat_type = stat_type <<1) + { + if (stat_type & cur_test_run.stat_collect_flags) + { + stat_cls = GNUNET_malloc (sizeof (struct STATcls)); + stat_cls->rps_peer = rps_peer; + stat_cls->stat_type = stat_type; + rps_peer->file_name_stats = + store_prefix_file_name (rps_peer->peer_id, "stats"); + GNUNET_STATISTICS_get (rps_peer->stats_h, + "rps", + stat_type_2_str (stat_type), + post_test_shutdown_ready_cb, + stat_iterator, + (struct STATcls *) stat_cls); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, + "Requested statistics for %s (peer %" PRIu32 ")\n", + stat_type_2_str (stat_type), + rps_peer->index); + } + } +} + + +/*********************************************************************** + * /Definition of tests +***********************************************************************/ + + +/** + * Actual "main" function for the testcase. + * + * @param cls closure + * @param h the run handle + * @param n_peers number of peers in 'peers' + * @param peers handle to peers run in the testbed + * @param links_succeeded the number of overlay link connection attempts that + * succeeded + * @param links_failed the number of overlay link connection attempts that + * failed + */ +static void +test_run (void *cls, + struct GNUNET_TESTBED_RunHandle *h, + unsigned int n_peers, + struct GNUNET_TESTBED_Peer **peers, + unsigned int links_succeeded, + unsigned int links_failed) +{ + unsigned int i; + struct OpListEntry *entry; + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "RUN was called\n"); + printf ("test 1\n"); + + /* Check whether we timed out */ + if (n_peers != num_peers || + NULL == peers || + 0 == links_succeeded) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Going down due to args (eg. timeout)\n"); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "\tn_peers: %u\n", n_peers); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "\tnum_peers: %" PRIu32 "\n", num_peers); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "\tpeers: %p\n", peers); + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "\tlinks_succeeded: %u\n", links_succeeded); + GNUNET_SCHEDULER_shutdown (); + return; + } + + + /* Initialize peers */ + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "going to initialise peers\n"); + testbed_peers = peers; + num_peers_online = 0; + for (i = 0; i < num_peers; i++) + { + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "initialising %u\n", i); + entry = make_oplist_entry (); + entry->index = i; + rps_peers[i].index = i; + if (NULL != cur_test_run.init_peer) + cur_test_run.init_peer (&rps_peers[i]); + if (NO_COLLECT_VIEW == cur_test_run.have_collect_view) + { + rps_peers->cur_view_count = 0; + rps_peers->cur_view = NULL; + } + entry->op = GNUNET_TESTBED_peer_get_information (peers[i], + GNUNET_TESTBED_PIT_IDENTITY, + &info_cb, + entry); + } + + /* Bring peers up */ + GNUNET_assert (num_peers == n_peers); + for (i = 0; i < n_peers; i++) + { + rps_peers[i].index = i; + rps_peers[i].op = + GNUNET_TESTBED_service_connect (&rps_peers[i], + peers[i], + "rps", + &rps_connect_complete_cb, + &rps_peers[i], + &rps_connect_adapter, + &rps_disconnect_adapter, + &rps_peers[i]); + /* Connect all peers to statistics service */ + if (COLLECT_STATISTICS == cur_test_run.have_collect_statistics) + { + rps_peers[i].stat_op = + GNUNET_TESTBED_service_connect (NULL, + peers[i], + "statistics", + stat_complete_cb, + &rps_peers[i], + &stat_connect_adapter, + &stat_disconnect_adapter, + &rps_peers[i]); + } + } + + if (NULL != churn_task) + GNUNET_SCHEDULER_cancel (churn_task); + post_test_task = GNUNET_SCHEDULER_add_delayed (timeout, &post_test_op, NULL); + timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, + (timeout_s * 1.2) + 0.1 * num_peers); + shutdown_task = GNUNET_SCHEDULER_add_delayed (timeout, &shutdown_op, NULL); + shutdown_task = GNUNET_SCHEDULER_add_shutdown (shutdown_op, NULL); + +} + + +/** + * Entry point for the testcase, sets up the testbed. + * + * @param argc unused + * @param argv unused + * @return 0 on success + */ +static void +run (void *cls, + char *const *args, + const char *cfgfile, + const struct GNUNET_CONFIGURATION_Handle *cfg) +{ + //int ret_value; + + /* Defaults for tests */ + churn_task = NULL; + + GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "This is the profiler\n"); + cur_test_run.name = "test-rps-profiler"; + num_peers = 10; + mal_type = 3; + cur_test_run.init_peer = profiler_init_peer; + //cur_test_run.pre_test = mal_pre; + cur_test_run.pre_test = pre_profiler; + cur_test_run.main_test = profiler_cb; + cur_test_run.reply_handle = profiler_reply_handle; + cur_test_run.eval_cb = profiler_eval; + cur_test_run.post_test = post_profiler; + cur_test_run.request_interval = 2; + cur_test_run.num_requests = 5; + //cur_test_run.have_churn = HAVE_CHURN; + cur_test_run.have_churn = HAVE_NO_CHURN; + cur_test_run.have_quick_quit = HAVE_NO_QUICK_QUIT; + cur_test_run.have_collect_statistics = COLLECT_STATISTICS; + cur_test_run.stat_collect_flags = STAT_TYPE_ROUNDS | + STAT_TYPE_BLOCKS | + STAT_TYPE_BLOCKS_MANY_PUSH | + STAT_TYPE_BLOCKS_NO_PUSH | + STAT_TYPE_BLOCKS_NO_PULL | + STAT_TYPE_BLOCKS_MANY_PUSH_NO_PULL | + STAT_TYPE_BLOCKS_NO_PUSH_NO_PULL | + STAT_TYPE_ISSUED_PUSH_SEND | + STAT_TYPE_ISSUED_PULL_REQ | + STAT_TYPE_ISSUED_PULL_REP | + STAT_TYPE_SENT_PUSH_SEND | + STAT_TYPE_SENT_PULL_REQ | + STAT_TYPE_SENT_PULL_REP | + STAT_TYPE_RECV_PUSH_SEND | + STAT_TYPE_RECV_PULL_REQ | + STAT_TYPE_RECV_PULL_REP; + cur_test_run.have_collect_view = COLLECT_VIEW; + timeout_s = 300; + + /* 'Clean' directory */ + (void) GNUNET_DISK_directory_remove ("/tmp/rps/"); + GNUNET_DISK_directory_create ("/tmp/rps/"); + timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, timeout_s); + + rps_peers = GNUNET_new_array (num_peers, struct RPSPeer); + peer_map = GNUNET_CONTAINER_multipeermap_create (num_peers, GNUNET_NO); + rps_peer_ids = GNUNET_new_array (num_peers, struct GNUNET_PeerIdentity); + if ( (2 == mal_type) || + (3 == mal_type)) + target_peer = &rps_peer_ids[num_peers - 2]; + if (profiler_eval == cur_test_run.eval_cb) + eval_peer = &rps_peers[num_peers - 1]; /* FIXME: eval_peer could be a + malicious peer if not careful + with the malicious portion */ + + ok = 1; + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "before _run()\n"); + //ret_value = GNUNET_TESTBED_test_run (cur_test_run.name, + // "test_rps.conf", + // num_peers, + // 0, NULL, NULL, + // &test_run, NULL); + GNUNET_TESTBED_run (NULL, + cfg, + num_peers, + 0, /* event mask */ + NULL, + NULL, + &test_run, + NULL); + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "after _run()\n"); + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "gnunet-rps-profiler returned.\n"); +} + +/** + * Entry point for the testcase, sets up the testbed. + * + * @param argc unused + * @param argv unused + * @return 0 on success + */ +int +main (int argc, char *argv[]) +{ + int ret_value; + struct GNUNET_GETOPT_CommandLineOption options[] = { + GNUNET_GETOPT_option_uint ('n', + "peers", + "COUNT", + gettext_noop ("number of peers to start"), + &num_peers), + GNUNET_GETOPT_OPTION_END + }; + + //if (GNUNET_OK != + // GNUNET_STRINGS_get_utf8_args (argc, argv, + // &argc, &argv)) + // return 2; + ret_value = 0; + if (GNUNET_OK != + GNUNET_PROGRAM_run (argc, + argv, + "gnunet-rps-profiler", + gettext_noop ("Measure quality and performance of the RPS service."), + options, + &run, + NULL)) + { + ret_value = 1; + } + if (GNUNET_OK != ret_value) + { + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Test did not run successfully!\n"); + } + + ret_value = cur_test_run.eval_cb(); + if (NO_COLLECT_VIEW == cur_test_run.have_collect_view) + { + GNUNET_array_grow (rps_peers->cur_view, + rps_peers->cur_view_count, + 0); + } + GNUNET_free (rps_peers); + GNUNET_free (rps_peer_ids); + GNUNET_CONTAINER_multipeermap_destroy (peer_map); + printf ("test -1\n"); + return ret_value; +} + +/* end of test_rps.c */ diff --git a/src/rps/test_rps.c b/src/rps/test_rps.c index bae28428f6..39aeacfebc 100644 --- a/src/rps/test_rps.c +++ b/src/rps/test_rps.c @@ -42,6 +42,12 @@ static uint32_t num_peers; /** * How long do we run the test? + * In seconds. + */ +static uint32_t timeout_s; + +/** + * How long do we run the test? */ //#define TIMEOUT GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 30) static struct GNUNET_TIME_Relative timeout; @@ -371,6 +377,11 @@ static int ok; /** * Identifier for the churn task that runs periodically */ +static struct GNUNET_SCHEDULER_Task *post_test_task; + +/** + * Identifier for the churn task that runs periodically + */ static struct GNUNET_SCHEDULER_Task *shutdown_task; /** @@ -556,6 +567,11 @@ struct SingleTestRun } cur_test_run; /** + * Did we finish the test? + */ +static int post_test; + +/** * Are we shutting down? */ static int in_shutdown; @@ -755,6 +771,41 @@ shutdown_op (void *cls) GNUNET_log (GNUNET_ERROR_TYPE_WARNING, "Shutdown task scheduled, going down.\n"); in_shutdown = GNUNET_YES; + if (NULL != post_test_task) + { + GNUNET_SCHEDULER_cancel (post_test_task); + } + if (NULL != churn_task) + { + GNUNET_SCHEDULER_cancel (churn_task); + churn_task = NULL; + } + for (i = 0; i < num_peers; i++) + { + if (NULL != rps_peers[i].rps_handle) + { + GNUNET_RPS_disconnect (rps_peers[i].rps_handle); + } + if (NULL != rps_peers[i].op) + { + GNUNET_TESTBED_operation_done (rps_peers[i].op); + } + } +} + + +/** + * Task run on timeout to collect statistics and potentially shut down. + */ +static void +post_test_op (void *cls) +{ + unsigned int i; + + post_test_task = NULL; + post_test = GNUNET_YES; + GNUNET_log (GNUNET_ERROR_TYPE_WARNING, + "Post test task scheduled, going down.\n"); if (NULL != churn_task) { GNUNET_SCHEDULER_cancel (churn_task); @@ -763,7 +814,10 @@ shutdown_op (void *cls) for (i = 0; i < num_peers; i++) { if (NULL != rps_peers[i].op) + { GNUNET_TESTBED_operation_done (rps_peers[i].op); + rps_peers[i].op = NULL; + } if (NULL != cur_test_run.post_test) { GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Executing post_test for peer %u\n", i); @@ -847,7 +901,7 @@ info_cb (void *cb_cls, { struct OpListEntry *entry = (struct OpListEntry *) cb_cls; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -901,7 +955,7 @@ rps_connect_complete_cb (void *cls, struct RPSPeer *rps_peer = cls; struct GNUNET_RPS_Handle *rps = ca_result; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1105,10 +1159,10 @@ default_reply_handle (void *cls, if (0 == evaluate () && HAVE_QUICK_QUIT == cur_test_run.have_quick_quit) { GNUNET_log (GNUNET_ERROR_TYPE_DEBUG, "Test succeeded before timeout\n"); - GNUNET_assert (NULL != shutdown_task); - GNUNET_SCHEDULER_cancel (shutdown_task); - shutdown_task = GNUNET_SCHEDULER_add_now (&shutdown_op, NULL); - GNUNET_assert (NULL!= shutdown_task); + GNUNET_assert (NULL != post_test_task); + GNUNET_SCHEDULER_cancel (post_test_task); + post_test_task = GNUNET_SCHEDULER_add_now (&post_test_op, NULL); + GNUNET_assert (NULL!= post_test_task); } } @@ -1122,7 +1176,7 @@ request_peers (void *cls) struct RPSPeer *rps_peer; struct PendingReply *pending_rep; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) return; rps_peer = pending_req->rps_peer; GNUNET_assert (1 <= rps_peer->num_pending_reqs); @@ -1185,7 +1239,7 @@ cancel_request_cb (void *cls) struct RPSPeer *rps_peer = cls; struct PendingReply *pending_rep; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) return; pending_rep = rps_peer->pending_rep_head; GNUNET_assert (1 <= rps_peer->num_pending_reps); @@ -1290,7 +1344,7 @@ mal_cb (struct RPSPeer *rps_peer) { uint32_t num_mal_peers; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1317,7 +1371,7 @@ mal_cb (struct RPSPeer *rps_peer) static void single_req_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1331,7 +1385,7 @@ single_req_cb (struct RPSPeer *rps_peer) static void delay_req_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1345,7 +1399,7 @@ delay_req_cb (struct RPSPeer *rps_peer) static void seed_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1361,7 +1415,7 @@ seed_cb (struct RPSPeer *rps_peer) static void seed_big_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1387,7 +1441,7 @@ single_peer_seed_cb (struct RPSPeer *rps_peer) static void seed_req_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1406,7 +1460,7 @@ seed_req_cb (struct RPSPeer *rps_peer) static void req_cancel_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1439,7 +1493,7 @@ churn (void *cls); static void churn_test_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1480,7 +1534,7 @@ churn_cb (void *cls, // FIXME struct OpListEntry *entry = cls; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1615,7 +1669,7 @@ churn (void *cls) double portion_go_online; double portion_go_offline; - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -1731,7 +1785,7 @@ profiler_reply_handle (void *cls, static void profiler_cb (struct RPSPeer *rps_peer) { - if (GNUNET_YES == in_shutdown) + if (GNUNET_YES == in_shutdown || GNUNET_YES == post_test) { return; } @@ -2644,7 +2698,12 @@ run (void *cls, if (NULL != churn_task) GNUNET_SCHEDULER_cancel (churn_task); + post_test_task = GNUNET_SCHEDULER_add_delayed (timeout, &post_test_op, NULL); + timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, + (timeout_s * 1.2) + 0.1 * num_peers); shutdown_task = GNUNET_SCHEDULER_add_delayed (timeout, &shutdown_op, NULL); + shutdown_task = GNUNET_SCHEDULER_add_shutdown (shutdown_op, NULL); + } @@ -2673,7 +2732,7 @@ main (int argc, char *argv[]) cur_test_run.stat_collect_flags = 0; cur_test_run.have_collect_view = NO_COLLECT_VIEW; churn_task = NULL; - timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 30); + timeout_s = 30; if (strstr (argv[0], "malicious") != NULL) { @@ -2725,7 +2784,7 @@ main (int argc, char *argv[]) cur_test_run.main_test = seed_big_cb; cur_test_run.eval_cb = no_eval; cur_test_run.have_churn = HAVE_NO_CHURN; - timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 10); + timeout_s = 10; } else if (strstr (argv[0], "_single_peer_seed") != NULL) @@ -2761,7 +2820,7 @@ main (int argc, char *argv[]) cur_test_run.main_test = req_cancel_cb; cur_test_run.eval_cb = no_eval; cur_test_run.have_churn = HAVE_NO_CHURN; - timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 10); + timeout_s = 10; } else if (strstr (argv[0], "_churn") != NULL) @@ -2775,7 +2834,7 @@ main (int argc, char *argv[]) cur_test_run.eval_cb = default_eval_cb; cur_test_run.have_churn = HAVE_NO_CHURN; cur_test_run.have_quick_quit = HAVE_NO_QUICK_QUIT; - timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 10); + timeout_s = 10; } else if (strstr (argv[0], "profiler") != NULL) @@ -2814,12 +2873,13 @@ main (int argc, char *argv[]) STAT_TYPE_RECV_PULL_REQ | STAT_TYPE_RECV_PULL_REP; cur_test_run.have_collect_view = COLLECT_VIEW; - timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, 300); + timeout_s = 150; /* 'Clean' directory */ (void) GNUNET_DISK_directory_remove ("/tmp/rps/"); GNUNET_DISK_directory_create ("/tmp/rps/"); } + timeout = GNUNET_TIME_relative_multiply (GNUNET_TIME_UNIT_SECONDS, timeout_s); rps_peers = GNUNET_new_array (num_peers, struct RPSPeer); peer_map = GNUNET_CONTAINER_multipeermap_create (num_peers, GNUNET_NO); diff --git a/src/rps/test_rps.conf b/src/rps/test_rps.conf index cf6b3150c1..e4ed111e2a 100644 --- a/src/rps/test_rps.conf +++ b/src/rps/test_rps.conf @@ -33,7 +33,7 @@ OVERLAY_TOPOLOGY = CLIQUE # OVERLAY_RANDOM_LINKS = 25 -# SETUP_TIMEOUT = 2 m +SETUP_TIMEOUT = 2 m [nse] WORKBITS = 0 |