/* DHT.c * * An implementation of the DHT as seen in http://wiki.tox.im/index.php/DHT * * Copyright (C) 2013 Tox project All Rights Reserved. * * This file is part of Tox. * * Tox is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * Tox is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with Tox. If not, see . * */ /*----------------------------------------------------------------------------------*/ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "logger.h" #include "DHT.h" #ifdef ENABLE_ASSOC_DHT #include "assoc.h" #endif #include "ping.h" #include "network.h" #include "LAN_discovery.h" #include "misc_tools.h" #include "util.h" /* The timeout after which a node is discarded completely. */ #define KILL_NODE_TIMEOUT 300 /* Ping interval in seconds for each random sending of a get nodes request. */ #define GET_NODE_INTERVAL 20 #define MAX_PUNCHING_PORTS 48 /* Interval in seconds between punching attempts*/ #define PUNCH_INTERVAL 3 #define MAX_NORMAL_PUNCHING_TRIES 5 #define NAT_PING_REQUEST 0 #define NAT_PING_RESPONSE 1 /* Number of get node requests to send to quickly find close nodes. */ #define MAX_BOOTSTRAP_TIMES 10 /* Used in the comparison function for sorting lists of Client_data. */ typedef struct { Client_data c1; Client_data c2; } ClientPair; /* Create the declaration for a quick sort for ClientPair structures. */ declare_quick_sort(ClientPair); /* Create the quicksort function. See misc_tools.h for the definition. */ make_quick_sort(ClientPair); Client_data *DHT_get_close_list(DHT *dht) { return dht->close_clientlist; } /* Compares client_id1 and client_id2 with client_id. * * return 0 if both are same distance. * return 1 if client_id1 is closer. * return 2 if client_id2 is closer. */ int id_closest(const uint8_t *id, const uint8_t *id1, const uint8_t *id2) { size_t i; uint8_t distance1, distance2; for (i = 0; i < CLIENT_ID_SIZE; ++i) { distance1 = abs(((int8_t *)id)[i] ^ ((int8_t *)id1)[i]); distance2 = abs(((int8_t *)id)[i] ^ ((int8_t *)id2)[i]); if (distance1 < distance2) return 1; if (distance1 > distance2) return 2; } return 0; } /* Turns the result of id_closest into something quick_sort can use. * Assumes p1->c1 == p2->c1. */ static int client_id_cmp(const ClientPair p1, const ClientPair p2) { int c = id_closest(p1.c1.client_id, p1.c2.client_id, p2.c2.client_id); if (c == 2) return -1; return c; } /* Shared key generations are costly, it is therefor smart to store commonly used * ones so that they can re used later without being computed again. * * If shared key is already in shared_keys, copy it to shared_key. * else generate it into shared_key and copy it to shared_keys */ void get_shared_key(Shared_Keys *shared_keys, uint8_t *shared_key, const uint8_t *secret_key, const uint8_t *client_id) { uint32_t i, num = ~0, curr = 0; for (i = 0; i < MAX_KEYS_PER_SLOT; ++i) { int index = client_id[30] * MAX_KEYS_PER_SLOT + i; if (shared_keys->keys[index].stored) { if (memcmp(client_id, shared_keys->keys[index].client_id, CLIENT_ID_SIZE) == 0) { memcpy(shared_key, shared_keys->keys[index].shared_key, crypto_box_BEFORENMBYTES); ++shared_keys->keys[index].times_requested; shared_keys->keys[index].time_last_requested = unix_time(); return; } if (num != 0) { if (is_timeout(shared_keys->keys[index].time_last_requested, KEYS_TIMEOUT)) { num = 0; curr = index; } else if (num > shared_keys->keys[index].times_requested) { num = shared_keys->keys[index].times_requested; curr = index; } } } else { if (num != 0) { num = 0; curr = index; } } } encrypt_precompute(client_id, secret_key, shared_key); if (num != (uint32_t)~0) { shared_keys->keys[curr].stored = 1; shared_keys->keys[curr].times_requested = 1; memcpy(shared_keys->keys[curr].client_id, client_id, CLIENT_ID_SIZE); memcpy(shared_keys->keys[curr].shared_key, shared_key, crypto_box_BEFORENMBYTES); shared_keys->keys[curr].time_last_requested = unix_time(); } } /* Copy shared_key to encrypt/decrypt DHT packet from client_id into shared_key * for packets that we receive. */ void DHT_get_shared_key_recv(DHT *dht, uint8_t *shared_key, const uint8_t *client_id) { return get_shared_key(&dht->shared_keys_recv, shared_key, dht->self_secret_key, client_id); } /* Copy shared_key to encrypt/decrypt DHT packet from client_id into shared_key * for packets that we send. */ void DHT_get_shared_key_sent(DHT *dht, uint8_t *shared_key, const uint8_t *client_id) { return get_shared_key(&dht->shared_keys_sent, shared_key, dht->self_secret_key, client_id); } void to_net_family(IP *ip) { if (ip->family == AF_INET) ip->family = TOX_AF_INET; else if (ip->family == AF_INET6) ip->family = TOX_AF_INET6; } void to_host_family(IP *ip) { if (ip->family == TOX_AF_INET) ip->family = AF_INET; else if (ip->family == TOX_AF_INET6) ip->family = AF_INET6; } /* Pack number of nodes into data of maxlength length. * * return length of packed nodes on success. * return -1 on failure. */ int pack_nodes(uint8_t *data, uint16_t length, const Node_format *nodes, uint16_t number) { uint32_t i, packed_length = 0; for (i = 0; i < number; ++i) { int ipv6 = -1; uint8_t net_family; if (nodes[i].ip_port.ip.family == AF_INET) { ipv6 = 0; net_family = TOX_AF_INET; } else if (nodes[i].ip_port.ip.family == TCP_INET) { ipv6 = 0; net_family = TOX_TCP_INET; } else if (nodes[i].ip_port.ip.family == AF_INET6) { ipv6 = 1; net_family = TOX_AF_INET6; } else if (nodes[i].ip_port.ip.family == TCP_INET6) { ipv6 = 1; net_family = TOX_TCP_INET6; } else { return -1; } if (ipv6 == 0) { uint32_t size = 1 + sizeof(IP4) + sizeof(uint16_t) + CLIENT_ID_SIZE; if (packed_length + size > length) return -1; data[packed_length] = net_family; memcpy(data + packed_length + 1, &nodes[i].ip_port.ip.ip4, sizeof(IP4)); memcpy(data + packed_length + 1 + sizeof(IP4), &nodes[i].ip_port.port, sizeof(uint16_t)); memcpy(data + packed_length + 1 + sizeof(IP4) + sizeof(uint16_t), nodes[i].client_id, CLIENT_ID_SIZE); packed_length += size; } else if (ipv6 == 1) { uint32_t size = 1 + sizeof(IP6) + sizeof(uint16_t) + CLIENT_ID_SIZE; if (packed_length + size > length) return -1; data[packed_length] = net_family; memcpy(data + packed_length + 1, &nodes[i].ip_port.ip.ip6, sizeof(IP6)); memcpy(data + packed_length + 1 + sizeof(IP6), &nodes[i].ip_port.port, sizeof(uint16_t)); memcpy(data + packed_length + 1 + sizeof(IP6) + sizeof(uint16_t), nodes[i].client_id, CLIENT_ID_SIZE); packed_length += size; } else { return -1; } } return packed_length; } /* Unpack data of length into nodes of size max_num_nodes. * Put the length of the data processed in processed_data_len. * tcp_enabled sets if TCP nodes are expected (true) or not (false). * * return number of unpacked nodes on success. * return -1 on failure. */ int unpack_nodes(Node_format *nodes, uint16_t max_num_nodes, uint16_t *processed_data_len, const uint8_t *data, uint16_t length, uint8_t tcp_enabled) { uint32_t num = 0, len_processed = 0; while (num < max_num_nodes && len_processed < length) { int ipv6 = -1; uint8_t host_family; if (data[len_processed] == TOX_AF_INET) { ipv6 = 0; host_family = AF_INET; } else if (data[len_processed] == TOX_TCP_INET) { if (!tcp_enabled) return -1; ipv6 = 0; host_family = TCP_INET; } else if (data[len_processed] == TOX_AF_INET6) { ipv6 = 1; host_family = AF_INET6; } else if (data[len_processed] == TOX_TCP_INET6) { if (!tcp_enabled) return -1; ipv6 = 1; host_family = TCP_INET6; } else { return -1; } if (ipv6 == 0) { uint32_t size = 1 + sizeof(IP4) + sizeof(uint16_t) + CLIENT_ID_SIZE; if (len_processed + size > length) return -1; nodes[num].ip_port.ip.family = host_family; memcpy(&nodes[num].ip_port.ip.ip4, data + len_processed + 1, sizeof(IP4)); memcpy(&nodes[num].ip_port.port, data + len_processed + 1 + sizeof(IP4), sizeof(uint16_t)); memcpy(nodes[num].client_id, data + len_processed + 1 + sizeof(IP4) + sizeof(uint16_t), CLIENT_ID_SIZE); len_processed += size; ++num; } else if (ipv6 == 1) { uint32_t size = 1 + sizeof(IP6) + sizeof(uint16_t) + CLIENT_ID_SIZE; if (len_processed + size > length) return -1; nodes[num].ip_port.ip.family = host_family; memcpy(&nodes[num].ip_port.ip.ip6, data + len_processed + 1, sizeof(IP6)); memcpy(&nodes[num].ip_port.port, data + len_processed + 1 + sizeof(IP6), sizeof(uint16_t)); memcpy(nodes[num].client_id, data + len_processed + 1 + sizeof(IP6) + sizeof(uint16_t), CLIENT_ID_SIZE); len_processed += size; ++num; } else { return -1; } } if (processed_data_len) *processed_data_len = len_processed; return num; } /* Check if client with client_id is already in list of length length. * If it is then set its corresponding timestamp to current time. * If the id is already in the list with a different ip_port, update it. * TODO: Maybe optimize this. * * return True(1) or False(0) */ static int client_or_ip_port_in_list(Client_data *list, uint32_t length, const uint8_t *client_id, IP_Port ip_port) { uint32_t i; uint64_t temp_time = unix_time(); /* if client_id is in list, find it and maybe overwrite ip_port */ for (i = 0; i < length; ++i) if (id_equal(list[i].client_id, client_id)) { /* Refresh the client timestamp. */ if (ip_port.ip.family == AF_INET) { LOGGER_SCOPE( if (!ipport_equal(&list[i].assoc4.ip_port, &ip_port)) { LOGGER_INFO("coipil[%u]: switching ipv4 from %s:%u to %s:%u", i, ip_ntoa(&list[i].assoc4.ip_port.ip), ntohs(list[i].assoc4.ip_port.port), ip_ntoa(&ip_port.ip), ntohs(ip_port.port)); } ); if (LAN_ip(list[i].assoc4.ip_port.ip) != 0 && LAN_ip(ip_port.ip) == 0) return 1; list[i].assoc4.ip_port = ip_port; list[i].assoc4.timestamp = temp_time; } else if (ip_port.ip.family == AF_INET6) { LOGGER_SCOPE( if (!ipport_equal(&list[i].assoc4.ip_port, &ip_port)) { LOGGER_INFO("coipil[%u]: switching ipv6 from %s:%u to %s:%u", i, ip_ntoa(&list[i].assoc6.ip_port.ip), ntohs(list[i].assoc6.ip_port.port), ip_ntoa(&ip_port.ip), ntohs(ip_port.port)); } ); if (LAN_ip(list[i].assoc6.ip_port.ip) != 0 && LAN_ip(ip_port.ip) == 0) return 1; list[i].assoc6.ip_port = ip_port; list[i].assoc6.timestamp = temp_time; } return 1; } /* client_id not in list yet: see if we can find an identical ip_port, in * that case we kill the old client_id by overwriting it with the new one * TODO: maybe we SHOULDN'T do that if that client_id is in a friend_list * and the one who is the actual friend's client_id/address set? */ for (i = 0; i < length; ++i) { /* MAYBE: check the other address, if valid, don't nuke? */ if ((ip_port.ip.family == AF_INET) && ipport_equal(&list[i].assoc4.ip_port, &ip_port)) { /* Initialize client timestamp. */ list[i].assoc4.timestamp = temp_time; memcpy(list[i].client_id, client_id, CLIENT_ID_SIZE); LOGGER_DEBUG("coipil[%u]: switching client_id (ipv4)", i); /* kill the other address, if it was set */ memset(&list[i].assoc6, 0, sizeof(list[i].assoc6)); return 1; } else if ((ip_port.ip.family == AF_INET6) && ipport_equal(&list[i].assoc6.ip_port, &ip_port)) { /* Initialize client timestamp. */ list[i].assoc6.timestamp = temp_time; memcpy(list[i].client_id, client_id, CLIENT_ID_SIZE); LOGGER_DEBUG("coipil[%u]: switching client_id (ipv6)", i); /* kill the other address, if it was set */ memset(&list[i].assoc4, 0, sizeof(list[i].assoc4)); return 1; } } return 0; } /* Check if client with client_id is already in node format list of length length. * * return 1 if true. * return 0 if false. */ static int client_in_nodelist(const Node_format *list, uint32_t length, const uint8_t *client_id) { uint32_t i; for (i = 0; i < length; ++i) { if (id_equal(list[i].client_id, client_id)) return 1; } return 0; } /* return friend number from the client_id. * return -1 if a failure occurs. */ static int friend_number(const DHT *dht, const uint8_t *client_id) { uint32_t i; for (i = 0; i < dht->num_friends; ++i) { if (id_equal(dht->friends_list[i].client_id, client_id)) return i; } return -1; } /*TODO: change this to 7 when done*/ #define HARDENING_ALL_OK 2 /* return 0 if not. * return 1 if route request are ok * return 2 if it responds to send node packets correctly * return 4 if it can test other nodes correctly * return HARDENING_ALL_OK if all ok. */ static uint8_t hardening_correct(const Hardening *h) { return h->routes_requests_ok + (h->send_nodes_ok << 1) + (h->testing_requests << 2); } /* * helper for get_close_nodes(). argument list is a monster :D */ static void get_close_nodes_inner(const uint8_t *client_id, Node_format *nodes_list, sa_family_t sa_family, const Client_data *client_list, uint32_t client_list_length, uint32_t *num_nodes_ptr, uint8_t is_LAN, uint8_t want_good) { if ((sa_family != AF_INET) && (sa_family != AF_INET6) && (sa_family != 0)) return; uint32_t num_nodes = *num_nodes_ptr; int j, closest; uint32_t i; for (i = 0; i < client_list_length; i++) { const Client_data *client = &client_list[i]; /* node already in list? */ if (client_in_nodelist(nodes_list, MAX_SENT_NODES, client->client_id)) continue; const IPPTsPng *ipptp = NULL; if (sa_family == AF_INET) { ipptp = &client->assoc4; } else if (sa_family == AF_INET6) { ipptp = &client->assoc6; } else { if (client->assoc4.timestamp >= client->assoc6.timestamp) { ipptp = &client->assoc4; } else { ipptp = &client->assoc6; } } /* node not in a good condition? */ if (is_timeout(ipptp->timestamp, BAD_NODE_TIMEOUT)) continue; /* don't send LAN ips to non LAN peers */ if (LAN_ip(ipptp->ip_port.ip) == 0 && !is_LAN) continue; if (LAN_ip(ipptp->ip_port.ip) != 0 && want_good && hardening_correct(&ipptp->hardening) != HARDENING_ALL_OK && !id_equal(client_id, client->client_id)) continue; if (num_nodes < MAX_SENT_NODES) { memcpy(nodes_list[num_nodes].client_id, client->client_id, CLIENT_ID_SIZE ); nodes_list[num_nodes].ip_port = ipptp->ip_port; num_nodes++; } else { /* see if node_list contains a client_id that's "further away" * compared to the one we're looking at at the moment, if there * is, replace it */ for (j = 0; j < MAX_SENT_NODES; ++j) { closest = id_closest( client_id, nodes_list[j].client_id, client->client_id ); /* second client_id is closer than current: change to it */ if (closest == 2) { memcpy( nodes_list[j].client_id, client->client_id, CLIENT_ID_SIZE); nodes_list[j].ip_port = ipptp->ip_port; break; } } } } *num_nodes_ptr = num_nodes; } /* Find MAX_SENT_NODES nodes closest to the client_id for the send nodes request: * put them in the nodes_list and return how many were found. * * TODO: For the love of based make * this function cleaner and much more efficient. * * want_good : do we want only good nodes as checked with the hardening returned or not? */ static int get_somewhat_close_nodes(const DHT *dht, const uint8_t *client_id, Node_format *nodes_list, sa_family_t sa_family, uint8_t is_LAN, uint8_t want_good) { uint32_t num_nodes = 0, i; get_close_nodes_inner(client_id, nodes_list, sa_family, dht->close_clientlist, LCLIENT_LIST, &num_nodes, is_LAN, want_good); /*TODO uncomment this when hardening is added to close friend clients for (i = 0; i < dht->num_friends; ++i) get_close_nodes_inner(dht, client_id, nodes_list, sa_family, dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, &num_nodes, is_LAN, want_good); */ for (i = 0; i < dht->num_friends; ++i) get_close_nodes_inner(client_id, nodes_list, sa_family, dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, &num_nodes, is_LAN, 0); return num_nodes; } int get_close_nodes(const DHT *dht, const uint8_t *client_id, Node_format *nodes_list, sa_family_t sa_family, uint8_t is_LAN, uint8_t want_good) { memset(nodes_list, 0, MAX_SENT_NODES * sizeof(Node_format)); #ifdef ENABLE_ASSOC_DHT if (!dht->assoc) #endif return get_somewhat_close_nodes(dht, client_id, nodes_list, sa_family, is_LAN, want_good); #ifdef ENABLE_ASSOC_DHT //TODO: assoc, sa_family 0 (don't care if ipv4 or ipv6) support. Client_data *result[MAX_SENT_NODES]; Assoc_close_entries request; memset(&request, 0, sizeof(request)); request.count = MAX_SENT_NODES; request.count_good = MAX_SENT_NODES - 2; /* allow 2 'indirect' nodes */ request.result = result; request.wanted_id = client_id; request.flags = (is_LAN ? LANOk : 0) + (sa_family == AF_INET ? ProtoIPv4 : ProtoIPv6); uint8_t num_found = Assoc_get_close_entries(dht->assoc, &request); if (!num_found) { LOGGER_DEBUG("get_close_nodes(): Assoc_get_close_entries() returned zero nodes"); return get_somewhat_close_nodes(dht, client_id, nodes_list, sa_family, is_LAN, want_good); } LOGGER_DEBUG("get_close_nodes(): Assoc_get_close_entries() returned %i 'direct' and %i 'indirect' nodes", request.count_good, num_found - request.count_good); uint8_t i, num_returned = 0; for (i = 0; i < num_found; i++) { Client_data *client = result[i]; if (client) { id_copy(nodes_list[num_returned].client_id, client->client_id); if (sa_family == AF_INET) if (ipport_isset(&client->assoc4.ip_port)) { nodes_list[num_returned].ip_port = client->assoc4.ip_port; num_returned++; continue; } if (sa_family == AF_INET6) if (ipport_isset(&client->assoc6.ip_port)) { nodes_list[num_returned].ip_port = client->assoc6.ip_port; num_returned++; continue; } } } return num_returned; #endif } /* Replace first bad (or empty) node with this one. * * return 0 if successful. * return 1 if not (list contains no bad nodes). */ static int replace_bad( Client_data *list, uint32_t length, const uint8_t *client_id, IP_Port ip_port ) { if ((ip_port.ip.family != AF_INET) && (ip_port.ip.family != AF_INET6)) return 1; uint32_t i; for (i = 0; i < length; ++i) { /* If node is bad */ Client_data *client = &list[i]; if (is_timeout(client->assoc4.timestamp, BAD_NODE_TIMEOUT) && is_timeout(client->assoc6.timestamp, BAD_NODE_TIMEOUT)) { IPPTsPng *ipptp_write = NULL; IPPTsPng *ipptp_clear = NULL; if (ip_port.ip.family == AF_INET) { ipptp_write = &client->assoc4; ipptp_clear = &client->assoc6; } else { ipptp_write = &client->assoc6; ipptp_clear = &client->assoc4; } memcpy(client->client_id, client_id, CLIENT_ID_SIZE); ipptp_write->ip_port = ip_port; ipptp_write->timestamp = unix_time(); ip_reset(&ipptp_write->ret_ip_port.ip); ipptp_write->ret_ip_port.port = 0; ipptp_write->ret_timestamp = 0; /* zero out other address */ memset(ipptp_clear, 0, sizeof(*ipptp_clear)); return 0; } } return 1; } /* Sort the list. It will be sorted from furthest to closest. * Turns list into data that quick sort can use and reverts it back. */ static void sort_list(Client_data *list, uint32_t length, const uint8_t *comp_client_id) { Client_data cd; ClientPair pairs[length]; uint32_t i; memcpy(cd.client_id, comp_client_id, CLIENT_ID_SIZE); for (i = 0; i < length; ++i) { pairs[i].c1 = cd; pairs[i].c2 = list[i]; } ClientPair_quick_sort(pairs, length, client_id_cmp); for (i = 0; i < length; ++i) list[i] = pairs[i].c2; } /* Replace first node that is possibly bad (tests failed or not done yet.) with this one. * * return 0 if successful. * return 1 if not (list contains no bad nodes). */ static int replace_possible_bad( Client_data *list, uint32_t length, const uint8_t *client_id, IP_Port ip_port, const uint8_t *comp_client_id ) { if ((ip_port.ip.family != AF_INET) && (ip_port.ip.family != AF_INET6)) return 1; sort_list(list, length, comp_client_id); /* TODO: decide if the following lines should stay commented or not. if (id_closest(comp_client_id, list[0].client_id, client_id) == 1) return 0;*/ uint32_t i; for (i = 0; i < length; ++i) { /* If node is bad */ Client_data *client = &list[i]; if (hardening_correct(&client->assoc4.hardening) != HARDENING_ALL_OK && hardening_correct(&client->assoc6.hardening) != HARDENING_ALL_OK) { IPPTsPng *ipptp_write = NULL; IPPTsPng *ipptp_clear = NULL; if (ip_port.ip.family == AF_INET) { ipptp_write = &client->assoc4; ipptp_clear = &client->assoc6; } else { ipptp_write = &client->assoc6; ipptp_clear = &client->assoc4; } memcpy(client->client_id, client_id, CLIENT_ID_SIZE); ipptp_write->ip_port = ip_port; ipptp_write->timestamp = unix_time(); ip_reset(&ipptp_write->ret_ip_port.ip); ipptp_write->ret_ip_port.port = 0; ipptp_write->ret_timestamp = 0; /* zero out other address */ memset(ipptp_clear, 0, sizeof(*ipptp_clear)); return 0; } } return 1; } /* Replace the first good node that is further to the comp_client_id than that of the client_id in the list * * returns 0 when the item was stored, 1 otherwise */ static int replace_good( Client_data *list, uint32_t length, const uint8_t *client_id, IP_Port ip_port, const uint8_t *comp_client_id ) { if ((ip_port.ip.family != AF_INET) && (ip_port.ip.family != AF_INET6)) return 1; /* TODO: eventually remove this.*/ if (length != LCLIENT_LIST) sort_list(list, length, comp_client_id); int8_t replace = -1; /* Because the list is sorted, we can simply check the client_id at the * border, either it is closer, then every other one is as well, or it is * further, then it gets pushed out in favor of the new address, which * will with the next sort() move to its "rightful" position * * CAVEAT: weirdly enough, the list is sorted DESCENDING in distance * so the furthest element is the first, NOT the last (at least that's * what the comment above sort_list() claims) */ if (id_closest(comp_client_id, list[0].client_id, client_id) == 2) replace = 0; if (replace != -1) { #ifdef DEBUG assert(replace >= 0 && replace < length); #endif Client_data *client = &list[replace]; IPPTsPng *ipptp_write = NULL; IPPTsPng *ipptp_clear = NULL; if (ip_port.ip.family == AF_INET) { ipptp_write = &client->assoc4; ipptp_clear = &client->assoc6; } else { ipptp_write = &client->assoc6; ipptp_clear = &client->assoc4; } memcpy(client->client_id, client_id, CLIENT_ID_SIZE); ipptp_write->ip_port = ip_port; ipptp_write->timestamp = unix_time(); ip_reset(&ipptp_write->ret_ip_port.ip); ipptp_write->ret_ip_port.port = 0; ipptp_write->ret_timestamp = 0; /* zero out other address */ memset(ipptp_clear, 0, sizeof(*ipptp_clear)); return 0; } return 1; } /* Attempt to add client with ip_port and client_id to the friends client list * and close_clientlist. * * returns 1+ if the item is used in any list, 0 else */ int addto_lists(DHT *dht, IP_Port ip_port, const uint8_t *client_id) { uint32_t i, used = 0; /* convert IPv4-in-IPv6 to IPv4 */ if ((ip_port.ip.family == AF_INET6) && IPV6_IPV4_IN_V6(ip_port.ip.ip6)) { ip_port.ip.family = AF_INET; ip_port.ip.ip4.uint32 = ip_port.ip.ip6.uint32[3]; } /* NOTE: Current behavior if there are two clients with the same id is * to replace the first ip by the second. */ if (!client_or_ip_port_in_list(dht->close_clientlist, LCLIENT_LIST, client_id, ip_port)) { if (replace_bad(dht->close_clientlist, LCLIENT_LIST, client_id, ip_port)) { if (replace_possible_bad(dht->close_clientlist, LCLIENT_LIST, client_id, ip_port, dht->self_public_key)) { /* If we can't replace bad nodes we try replacing good ones. */ if (!replace_good(dht->close_clientlist, LCLIENT_LIST, client_id, ip_port, dht->self_public_key)) used++; } else used++; } else used++; } else used++; for (i = 0; i < dht->num_friends; ++i) { if (!client_or_ip_port_in_list(dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, client_id, ip_port)) { if (replace_bad(dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, client_id, ip_port)) { /*if (replace_possible_bad(dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, client_id, ip_port, dht->friends_list[i].client_id)) {*/ /* If we can't replace bad nodes we try replacing good ones. */ if (!replace_good(dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, client_id, ip_port, dht->friends_list[i].client_id)) used++; /*} else used++;*/ } else used++; } else used++; } #ifdef ENABLE_ASSOC_DHT if (dht->assoc) { IPPTs ippts; ippts.ip_port = ip_port; ippts.timestamp = unix_time(); Assoc_add_entry(dht->assoc, client_id, &ippts, NULL, used ? 1 : 0); } #endif return used; } /* If client_id is a friend or us, update ret_ip_port * nodeclient_id is the id of the node that sent us this info. */ static int returnedip_ports(DHT *dht, IP_Port ip_port, const uint8_t *client_id, const uint8_t *nodeclient_id) { uint32_t i, j; uint64_t temp_time = unix_time(); uint32_t used = 0; /* convert IPv4-in-IPv6 to IPv4 */ if ((ip_port.ip.family == AF_INET6) && IPV6_IPV4_IN_V6(ip_port.ip.ip6)) { ip_port.ip.family = AF_INET; ip_port.ip.ip4.uint32 = ip_port.ip.ip6.uint32[3]; } if (id_equal(client_id, dht->self_public_key)) { for (i = 0; i < LCLIENT_LIST; ++i) { if (id_equal(nodeclient_id, dht->close_clientlist[i].client_id)) { if (ip_port.ip.family == AF_INET) { dht->close_clientlist[i].assoc4.ret_ip_port = ip_port; dht->close_clientlist[i].assoc4.ret_timestamp = temp_time; } else if (ip_port.ip.family == AF_INET6) { dht->close_clientlist[i].assoc6.ret_ip_port = ip_port; dht->close_clientlist[i].assoc6.ret_timestamp = temp_time; } ++used; break; } } } else { for (i = 0; i < dht->num_friends; ++i) { if (id_equal(client_id, dht->friends_list[i].client_id)) { for (j = 0; j < MAX_FRIEND_CLIENTS; ++j) { if (id_equal(nodeclient_id, dht->friends_list[i].client_list[j].client_id)) { if (ip_port.ip.family == AF_INET) { dht->friends_list[i].client_list[j].assoc4.ret_ip_port = ip_port; dht->friends_list[i].client_list[j].assoc4.ret_timestamp = temp_time; } else if (ip_port.ip.family == AF_INET6) { dht->friends_list[i].client_list[j].assoc6.ret_ip_port = ip_port; dht->friends_list[i].client_list[j].assoc6.ret_timestamp = temp_time; } ++used; goto end; } } } } } end: #ifdef ENABLE_ASSOC_DHT if (dht->assoc) { IPPTs ippts; ippts.ip_port = ip_port; ippts.timestamp = temp_time; /* this is only a hear-say entry, so ret-ipp is NULL, but used is required * to decide how valuable it is ("used" may throw an "unused" entry out) */ Assoc_add_entry(dht->assoc, client_id, &ippts, NULL, used ? 1 : 0); } #endif return 0; } #define NODES_ENCRYPTED_MESSAGE_LENGTH (crypto_box_NONCEBYTES + sizeof(uint64_t) + sizeof(Node_format) + sizeof(Node_format) + crypto_box_MACBYTES) /* Send a getnodes request. sendback_node is the node that it will send back the response to (set to NULL to disable this) */ static int getnodes(DHT *dht, IP_Port ip_port, const uint8_t *public_key, const uint8_t *client_id, const Node_format *sendback_node) { /* Check if packet is going to be sent to ourself. */ if (id_equal(public_key, dht->self_public_key)) return -1; uint8_t plain_message[sizeof(Node_format) * 2] = {0}; Node_format receiver; memcpy(receiver.client_id, public_key, CLIENT_ID_SIZE); receiver.ip_port = ip_port; memcpy(plain_message, &receiver, sizeof(receiver)); uint64_t ping_id = 0; if (sendback_node != NULL) { memcpy(plain_message + sizeof(receiver), sendback_node, sizeof(Node_format)); ping_id = ping_array_add(&dht->dht_harden_ping_array, plain_message, sizeof(plain_message)); } else { ping_id = ping_array_add(&dht->dht_ping_array, plain_message, sizeof(receiver)); } if (ping_id == 0) return -1; uint8_t plain[CLIENT_ID_SIZE + sizeof(ping_id)]; uint8_t encrypt[sizeof(plain) + crypto_box_MACBYTES]; uint8_t data[1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES + sizeof(encrypt)]; memcpy(plain, client_id, CLIENT_ID_SIZE); memcpy(plain + CLIENT_ID_SIZE, &ping_id, sizeof(ping_id)); uint8_t shared_key[crypto_box_BEFORENMBYTES]; DHT_get_shared_key_sent(dht, shared_key, public_key); uint8_t nonce[crypto_box_NONCEBYTES]; new_nonce(nonce); int len = encrypt_data_symmetric( shared_key, nonce, plain, sizeof(plain), encrypt ); if (len != sizeof(encrypt)) return -1; data[0] = NET_PACKET_GET_NODES; memcpy(data + 1, dht->self_public_key, CLIENT_ID_SIZE); memcpy(data + 1 + CLIENT_ID_SIZE, nonce, crypto_box_NONCEBYTES); memcpy(data + 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES, encrypt, len); return sendpacket(dht->net, ip_port, data, sizeof(data)); } /* Send a send nodes response: message for IPv6 nodes */ static int sendnodes_ipv6(const DHT *dht, IP_Port ip_port, const uint8_t *public_key, const uint8_t *client_id, const uint8_t *sendback_data, uint16_t length, const uint8_t *shared_encryption_key) { /* Check if packet is going to be sent to ourself. */ if (id_equal(public_key, dht->self_public_key)) return -1; if (length > NODES_ENCRYPTED_MESSAGE_LENGTH || length == 0) return -1; size_t Node_format_size = sizeof(Node_format); uint8_t data[1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES + Node_format_size * MAX_SENT_NODES + length + crypto_box_MACBYTES]; Node_format nodes_list[MAX_SENT_NODES]; uint32_t num_nodes = get_close_nodes(dht, client_id, nodes_list, 0, LAN_ip(ip_port.ip) == 0, 1); if (num_nodes == 0) return 0; uint8_t plain[1 + Node_format_size * MAX_SENT_NODES + length]; uint8_t encrypt[sizeof(plain) + crypto_box_MACBYTES]; uint8_t nonce[crypto_box_NONCEBYTES]; new_nonce(nonce); int nodes_length = pack_nodes(plain + 1, Node_format_size * MAX_SENT_NODES, nodes_list, num_nodes); if (nodes_length <= 0) return -1; plain[0] = num_nodes; memcpy(plain + 1 + nodes_length, sendback_data, length); int len = encrypt_data_symmetric( shared_encryption_key, nonce, plain, 1 + nodes_length + length, encrypt ); if (len != 1 + nodes_length + length + crypto_box_MACBYTES) return -1; data[0] = NET_PACKET_SEND_NODES_IPV6; memcpy(data + 1, dht->self_public_key, CLIENT_ID_SIZE); memcpy(data + 1 + CLIENT_ID_SIZE, nonce, crypto_box_NONCEBYTES); memcpy(data + 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES, encrypt, len); return sendpacket(dht->net, ip_port, data, 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES + len); } static int handle_getnodes(void *object, IP_Port source, const uint8_t *packet, uint32_t length) { uint32_t cmp_len = 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES + CLIENT_ID_SIZE + crypto_box_MACBYTES; if (length <= cmp_len) return 1; if (length > cmp_len + NODES_ENCRYPTED_MESSAGE_LENGTH) return 1; uint16_t sendback_data_length = length - cmp_len; DHT *dht = object; /* Check if packet is from ourself. */ if (id_equal(packet + 1, dht->self_public_key)) return 1; uint8_t plain[CLIENT_ID_SIZE + sendback_data_length]; uint8_t shared_key[crypto_box_BEFORENMBYTES]; DHT_get_shared_key_recv(dht, shared_key, packet + 1); int len = decrypt_data_symmetric( shared_key, packet + 1 + CLIENT_ID_SIZE, packet + 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES, CLIENT_ID_SIZE + sendback_data_length + crypto_box_MACBYTES, plain ); if (len != CLIENT_ID_SIZE + sendback_data_length) return 1; sendnodes_ipv6(dht, source, packet + 1, plain, plain + CLIENT_ID_SIZE, sendback_data_length, shared_key); add_to_ping(dht->ping, packet + 1, source); return 0; } /* return 0 if no return 1 if yes */ static uint8_t sent_getnode_to_node(DHT *dht, const uint8_t *client_id, IP_Port node_ip_port, uint64_t ping_id, Node_format *sendback_node) { uint8_t data[sizeof(Node_format) * 2]; if (ping_array_check(data, sizeof(data), &dht->dht_ping_array, ping_id) == sizeof(Node_format)) { memset(sendback_node, 0, sizeof(Node_format)); } else if (ping_array_check(data, sizeof(data), &dht->dht_harden_ping_array, ping_id) == sizeof(data)) { memcpy(sendback_node, data + sizeof(Node_format), sizeof(Node_format)); } else { return 0; } Node_format test; memcpy(&test, data, sizeof(Node_format)); if (!ipport_equal(&test.ip_port, &node_ip_port) || memcmp(test.client_id, client_id, CLIENT_ID_SIZE) != 0) return 0; return 1; } /* Function is needed in following functions. */ static int send_hardening_getnode_res(const DHT *dht, const Node_format *sendto, const uint8_t *queried_client_id, const uint8_t *nodes_data, uint16_t nodes_data_length); static int handle_sendnodes_core(void *object, IP_Port source, const uint8_t *packet, uint32_t length, Node_format *plain_nodes, uint16_t size_plain_nodes, uint32_t *num_nodes_out) { DHT *dht = object; uint32_t cid_size = 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES + 1 + sizeof(uint64_t) + crypto_box_MACBYTES; if (length <= cid_size) /* too short */ return 1; uint32_t data_size = length - cid_size; if (data_size == 0) return 1; if (data_size > sizeof(Node_format) * MAX_SENT_NODES) /* invalid length */ return 1; uint8_t plain[1 + data_size + sizeof(uint64_t)]; uint8_t shared_key[crypto_box_BEFORENMBYTES]; DHT_get_shared_key_sent(dht, shared_key, packet + 1); int len = decrypt_data_symmetric( shared_key, packet + 1 + CLIENT_ID_SIZE, packet + 1 + CLIENT_ID_SIZE + crypto_box_NONCEBYTES, 1 + data_size + sizeof(uint64_t) + crypto_box_MACBYTES, plain); if ((unsigned int)len != sizeof(plain)) return 1; if (plain[0] > size_plain_nodes || plain[0] == 0) return 1; Node_format sendback_node; uint64_t ping_id; memcpy(&ping_id, plain + 1 + data_size, sizeof(ping_id)); if (!sent_getnode_to_node(dht, packet + 1, source, ping_id, &sendback_node)) return 1; uint16_t length_nodes = 0; int num_nodes = unpack_nodes(plain_nodes, plain[0], &length_nodes, plain + 1, data_size, 0); if (length_nodes != data_size) return 1; if (num_nodes != plain[0]) return 1; if (num_nodes <= 0) return 1; /* store the address the *request* was sent to */ addto_lists(dht, source, packet + 1); *num_nodes_out = num_nodes; send_hardening_getnode_res(dht, &sendback_node, packet + 1, plain + 1, data_size); return 0; } static int handle_sendnodes_ipv6(void *object, IP_Port source, const uint8_t *packet, uint32_t length) { DHT *dht = object; Node_format plain_nodes[MAX_SENT_NODES]; uint32_t num_nodes; if (handle_sendnodes_core(object, source, packet, length, plain_nodes, MAX_SENT_NODES, &num_nodes)) return 1; if (num_nodes == 0) return 0; uint32_t i; for (i = 0; i < num_nodes; i++) { if (ipport_isset(&plain_nodes[i].ip_port)) { send_ping_request(dht->ping, plain_nodes[i].ip_port, plain_nodes[i].client_id); returnedip_ports(dht, plain_nodes[i].ip_port, plain_nodes[i].client_id, packet + 1); } } return 0; } /*----------------------------------------------------------------------------------*/ /*------------------------END of packet handling functions--------------------------*/ /* * Send get nodes requests with client_id to max_num peers in list of length length */ /* static void get_bunchnodes(DHT *dht, Client_data *list, uint16_t length, uint16_t max_num, uint8_t *client_id) { uint32_t i, num = 0; for (i = 0; i < length; ++i) { IPPTsPng *assoc; uint32_t a; for (a = 0, assoc = &list[i].assoc6; a < 2; a++, assoc = &list[i].assoc4) if (ipport_isset(&(assoc->ip_port)) && !is_timeout(assoc->ret_timestamp, BAD_NODE_TIMEOUT)) { getnodes(dht, assoc->ip_port, list[i].client_id, client_id, NULL); ++num; if (num >= max_num) return; } } } */ int DHT_addfriend(DHT *dht, const uint8_t *client_id) { if (friend_number(dht, client_id) != -1) /* Is friend already in DHT? */ return 1; DHT_Friend *temp; temp = realloc(dht->friends_list, sizeof(DHT_Friend) * (dht->num_friends + 1)); if (temp == NULL) return 1; dht->friends_list = temp; memset(&dht->friends_list[dht->num_friends], 0, sizeof(DHT_Friend)); memcpy(dht->friends_list[dht->num_friends].client_id, client_id, CLIENT_ID_SIZE); dht->friends_list[dht->num_friends].nat.NATping_id = random_64b(); ++dht->num_friends; #ifdef ENABLE_ASSOC_DHT if (dht->assoc) { /* get up to MAX_FRIEND_CLIENTS connectable nodes */ DHT_Friend *friend = &dht->friends_list[dht->num_friends - 1]; Assoc_close_entries close_entries; memset(&close_entries, 0, sizeof(close_entries)); close_entries.wanted_id = client_id; close_entries.count_good = MAX_FRIEND_CLIENTS / 2; close_entries.count = MAX_FRIEND_CLIENTS; close_entries.result = calloc(MAX_FRIEND_CLIENTS, sizeof(*close_entries.result)); uint8_t i, found = Assoc_get_close_entries(dht->assoc, &close_entries); for (i = 0; i < found; i++) memcpy(&friend->client_list[i], close_entries.result[i], sizeof(*close_entries.result[i])); if (found) { /* send getnodes to the "best" entry */ Client_data *client = &friend->client_list[0]; if (ipport_isset(&client->assoc4.ip_port)) getnodes(dht, client->assoc4.ip_port, client->client_id, friend->client_id, NULL); if (ipport_isset(&client->assoc6.ip_port)) getnodes(dht, client->assoc6.ip_port, client->client_id, friend->client_id, NULL); } } #endif /*this isn't really useful anymore. get_bunchnodes(dht, dht->close_clientlist, LCLIENT_LIST, MAX_FRIEND_CLIENTS, client_id);*/ return 0; } int DHT_delfriend(DHT *dht, const uint8_t *client_id) { uint32_t i; DHT_Friend *temp; for (i = 0; i < dht->num_friends; ++i) { /* Equal */ if (id_equal(dht->friends_list[i].client_id, client_id)) { --dht->num_friends; if (dht->num_friends != i) { memcpy( &dht->friends_list[i], &dht->friends_list[dht->num_friends], sizeof(DHT_Friend) ); } if (dht->num_friends == 0) { free(dht->friends_list); dht->friends_list = NULL; return 0; } temp = realloc(dht->friends_list, sizeof(DHT_Friend) * (dht->num_friends)); if (temp == NULL) return 1; dht->friends_list = temp; return 0; } } return 1; } /* TODO: Optimize this. */ int DHT_getfriendip(const DHT *dht, const uint8_t *client_id, IP_Port *ip_port) { uint32_t i, j; ip_reset(&ip_port->ip); ip_port->port = 0; for (i = 0; i < dht->num_friends; ++i) { /* Equal */ if (id_equal(dht->friends_list[i].client_id, client_id)) { for (j = 0; j < MAX_FRIEND_CLIENTS; ++j) { Client_data *client = &dht->friends_list[i].client_list[j]; if (id_equal(client->client_id, client_id)) { IPPTsPng *assoc = NULL; uint32_t a; for (a = 0, assoc = &client->assoc6; a < 2; a++, assoc = &client->assoc4) if (!is_timeout(assoc->timestamp, BAD_NODE_TIMEOUT)) { *ip_port = assoc->ip_port; return 1; } } } return 0; } } return -1; } /* returns number of nodes not in kill-timeout */ static uint8_t do_ping_and_sendnode_requests(DHT *dht, uint64_t *lastgetnode, const uint8_t *client_id, Client_data *list, uint32_t list_count, uint32_t *bootstrap_times) { uint32_t i; uint8_t not_kill = 0; uint64_t temp_time = unix_time(); uint32_t num_nodes = 0; Client_data *client_list[list_count * 2]; IPPTsPng *assoc_list[list_count * 2]; for (i = 0; i < list_count; i++) { /* If node is not dead. */ Client_data *client = &list[i]; IPPTsPng *assoc; uint32_t a; for (a = 0, assoc = &client->assoc6; a < 2; a++, assoc = &client->assoc4) if (!is_timeout(assoc->timestamp, KILL_NODE_TIMEOUT)) { not_kill++; if (is_timeout(assoc->last_pinged, PING_INTERVAL)) { send_ping_request(dht->ping, assoc->ip_port, client->client_id ); assoc->last_pinged = temp_time; } /* If node is good. */ if (!is_timeout(assoc->timestamp, BAD_NODE_TIMEOUT)) { client_list[num_nodes] = client; assoc_list[num_nodes] = assoc; ++num_nodes; } } } if ((num_nodes != 0) && (is_timeout(*lastgetnode, GET_NODE_INTERVAL) || *bootstrap_times < MAX_BOOTSTRAP_TIMES)) { uint32_t rand_node = rand() % num_nodes; getnodes(dht, assoc_list[rand_node]->ip_port, client_list[rand_node]->client_id, client_id, NULL); *lastgetnode = temp_time; ++*bootstrap_times; } return not_kill; } /* Ping each client in the "friends" list every PING_INTERVAL seconds. Send a get nodes request * every GET_NODE_INTERVAL seconds to a random good node for each "friend" in our "friends" list. */ static void do_DHT_friends(DHT *dht) { uint32_t i; for (i = 0; i < dht->num_friends; ++i) do_ping_and_sendnode_requests(dht, &dht->friends_list[i].lastgetnode, dht->friends_list[i].client_id, dht->friends_list[i].client_list, MAX_FRIEND_CLIENTS, &dht->friends_list[i].bootstrap_times); } /* Ping each client in the close nodes list every PING_INTERVAL seconds. * Send a get nodes request every GET_NODE_INTERVAL seconds to a random good node in the list. */ static void do_Close(DHT *dht) { uint8_t not_killed = do_ping_and_sendnode_requests(dht, &dht->close_lastgetnodes, dht->self_public_key, dht->close_clientlist, LCLIENT_LIST, &dht->close_bootstrap_times); if (!not_killed) { /* all existing nodes are at least KILL_NODE_TIMEOUT, * which means we are mute, as we only send packets to * nodes NOT in KILL_NODE_TIMEOUT * * so: reset all nodes to be BAD_NODE_TIMEOUT, but not * KILL_NODE_TIMEOUT, so we at least keep trying pings */ uint64_t badonly = unix_time() - BAD_NODE_TIMEOUT; size_t i, a; for (i = 0; i < LCLIENT_LIST; i++) { Client_data *client = &dht->close_clientlist[i]; IPPTsPng *assoc; for (a = 0, assoc = &client->assoc4; a < 2; a++, assoc = &client->assoc6) if (assoc->timestamp) assoc->timestamp = badonly; } } } void DHT_getnodes(DHT *dht, const IP_Port *from_ipp, const uint8_t *from_id, const uint8_t *which_id) { getnodes(dht, *from_ipp, from_id, which_id, NULL); } void DHT_bootstrap(DHT *dht, IP_Port ip_port, const uint8_t *public_key) { /*#ifdef ENABLE_ASSOC_DHT if (dht->assoc) { IPPTs ippts; ippts.ip_port = ip_port; ippts.timestamp = 0; Assoc_add_entry(dht->assoc, public_key, &ippts, NULL, 0); } #endif*/ getnodes(dht, ip_port, public_key, dht->self_public_key, NULL); } int DHT_bootstrap_from_address(DHT *dht, const char *address, uint8_t ipv6enabled, uint16_t port, const uint8_t *public_key) { IP_Port ip_port_v64; IP *ip_extra = NULL; IP_Port ip_port_v4; ip_init(&ip_port_v64.ip, ipv6enabled); if (ipv6enabled) { /* setup for getting BOTH: an IPv6 AND an IPv4 address */ ip_port_v64.ip.family = AF_UNSPEC; ip_reset(&ip_port_v4.ip); ip_extra = &ip_port_v4.ip; } if (addr_resolve_or_parse_ip(address, &ip_port_v64.ip, ip_extra)) { ip_port_v64.port = port; DHT_bootstrap(dht, ip_port_v64, public_key); if ((ip_extra != NULL) && ip_isset(ip_extra)) { ip_port_v4.port = port; DHT_bootstrap(dht, ip_port_v4, public_key); } return 1; } else return 0; } /* Send the given packet to node with client_id * * return -1 if failure. */ int route_packet(const DHT *dht, const uint8_t *client_id, const uint8_t *packet, uint32_t length) { uint32_t i; for (i = 0; i < LCLIENT_LIST; ++i) { if (id_equal(client_id, dht->close_clientlist[i].client_id)) { const Client_data *client = &dht->close_clientlist[i]; if (ip_isset(&client->assoc6.ip_port.ip)) return sendpacket(dht->net, client->assoc6.ip_port, packet, length); else if (ip_isset(&client->assoc4.ip_port.ip)) return sendpacket(dht->net, client->assoc4.ip_port, packet, length); else break; } } return -1; } /* Puts all the different ips returned by the nodes for a friend_num into array ip_portlist. * ip_portlist must be at least MAX_FRIEND_CLIENTS big. * * return the number of ips returned. * return 0 if we are connected to friend or if no ips were found. * return -1 if no such friend. */ static int friend_iplist(const DHT *dht, IP_Port *ip_portlist, uint16_t friend_num) { if (friend_num >= dht->num_friends) return -1; DHT_Friend *friend = &dht->friends_list[friend_num]; Client_data *client; IP_Port ipv4s[MAX_FRIEND_CLIENTS]; int num_ipv4s = 0; IP_Port ipv6s[MAX_FRIEND_CLIENTS]; int num_ipv6s = 0; int i; for (i = 0; i < MAX_FRIEND_CLIENTS; ++i) { client = &(friend->client_list[i]); /* If ip is not zero and node is good. */ if (ip_isset(&client->assoc4.ret_ip_port.ip) && !is_timeout(client->assoc4.ret_timestamp, BAD_NODE_TIMEOUT)) { ipv4s[num_ipv4s] = client->assoc4.ret_ip_port; ++num_ipv4s; } if (ip_isset(&client->assoc6.ret_ip_port.ip) && !is_timeout(client->assoc6.ret_timestamp, BAD_NODE_TIMEOUT)) { ipv6s[num_ipv6s] = client->assoc6.ret_ip_port; ++num_ipv6s; } if (id_equal(client->client_id, friend->client_id)) if (!is_timeout(client->assoc6.timestamp, BAD_NODE_TIMEOUT) || !is_timeout(client->assoc4.timestamp, BAD_NODE_TIMEOUT)) return 0; /* direct connectivity */ } #ifdef FRIEND_IPLIST_PAD memcpy(ip_portlist, ipv6s, num_ipv6s * sizeof(IP_Port)); if (num_ipv6s == MAX_FRIEND_CLIENTS) return MAX_FRIEND_CLIENTS; int num_ipv4s_used = MAX_FRIEND_CLIENTS - num_ipv6s; if (num_ipv4s_used > num_ipv4s) num_ipv4s_used = num_ipv4s; memcpy(&ip_portlist[num_ipv6s], ipv4s, num_ipv4s_used * sizeof(IP_Port)); return num_ipv6s + num_ipv4s_used; #else /* !FRIEND_IPLIST_PAD */ /* there must be some secret reason why we can't pad the longer list * with the shorter one... */ if (num_ipv6s >= num_ipv4s) { memcpy(ip_portlist, ipv6s, num_ipv6s * sizeof(IP_Port)); return num_ipv6s; } memcpy(ip_portlist, ipv4s, num_ipv4s * sizeof(IP_Port)); return num_ipv4s; #endif /* !FRIEND_IPLIST_PAD */ } /* Send the following packet to everyone who tells us they are connected to friend_id. * * return ip for friend. * return number of nodes the packet was sent to. (Only works if more than (MAX_FRIEND_CLIENTS / 4). */ int route_tofriend(const DHT *dht, const uint8_t *friend_id, const uint8_t *packet, uint32_t length) { int num = friend_number(dht, friend_id); if (num == -1) return 0; uint32_t i, sent = 0; uint8_t friend_sent[MAX_FRIEND_CLIENTS] = {0}; IP_Port ip_list[MAX_FRIEND_CLIENTS]; int ip_num = friend_iplist(dht, ip_list, num); if (ip_num < (MAX_FRIEND_CLIENTS / 4)) return 0; /* Reason for that? */ DHT_Friend *friend = &dht->friends_list[num]; Client_data *client; /* extra legwork, because having the outside allocating the space for us * is *usually* good(tm) (bites us in the behind in this case though) */ uint32_t a; for (a = 0; a < 2; a++) for (i = 0; i < MAX_FRIEND_CLIENTS; ++i) { if (friend_sent[i])/* Send one packet per client.*/ continue; client = &friend->client_list[i]; IPPTsPng *assoc = NULL; if (!a) assoc = &client->assoc4; else assoc = &client->assoc6; /* If ip is not zero and node is good. */ if (ip_isset(&assoc->ret_ip_port.ip) && !is_timeout(assoc->ret_timestamp, BAD_NODE_TIMEOUT)) { int retval = sendpacket(dht->net, assoc->ip_port, packet, length); if ((unsigned int)retval == length) { ++sent; friend_sent[i] = 1; } } } return sent; } /* Send the following packet to one random person who tells us they are connected to friend_id. * * return number of nodes the packet was sent to. */ static int routeone_tofriend(DHT *dht, const uint8_t *friend_id, const uint8_t *packet, uint32_t length) { int num = friend_number(dht, friend_id); if (num == -1) return 0; DHT_Friend *friend = &dht->friends_list[num]; Client_data *client; IP_Port ip_list[MAX_FRIEND_CLIENTS * 2]; int n = 0; uint32_t i; /* extra legwork, because having the outside allocating the space for us * is *usually* good(tm) (bites us in the behind in this case though) */ uint32_t a; for (a = 0; a < 2; a++) for (i = 0; i < MAX_FRIEND_CLIENTS; ++i) { client = &friend->client_list[i]; IPPTsPng *assoc = NULL; if (!a) assoc = &client->assoc4; else assoc = &client->assoc6; /* If ip is not zero and node is good. */ if (ip_isset(&assoc->ret_ip_port.ip) && !is_timeout(assoc->ret_timestamp, BAD_NODE_TIMEOUT)) { ip_list[n] = assoc->ip_port; ++n; } } if (n < 1) return 0; int retval = sendpacket(dht->net, ip_list[rand() % n], packet, length); if ((unsigned int)retval == length) return 1; return 0; } /* Puts all the different ips returned by the nodes for a friend_id into array ip_portlist. * ip_portlist must be at least MAX_FRIEND_CLIENTS big. * * return number of ips returned. * return 0 if we are connected to friend or if no ips were found. * return -1 if no such friend. */ int friend_ips(const DHT *dht, IP_Port *ip_portlist, const uint8_t *friend_id) { uint32_t i; for (i = 0; i < dht->num_friends; ++i) { /* Equal */ if (id_equal(dht->friends_list[i].client_id, friend_id)) return friend_iplist(dht, ip_portlist, i); } return -1; } /*----------------------------------------------------------------------------------*/ /*---------------------BEGINNING OF NAT PUNCHING FUNCTIONS--------------------------*/ static int send_NATping(DHT *dht, const uint8_t *public_key, uint64_t ping_id, uint8_t type) { uint8_t data[sizeof(uint64_t) + 1]; uint8_t packet[MAX_CRYPTO_REQUEST_SIZE]; int num = 0; data[0] = type; memcpy(data + 1, &ping_id, sizeof(uint64_t)); /* 254 is NAT ping request packet id */ int len = create_request(dht->self_public_key, dht->self_secret_key, packet, public_key, data, sizeof(uint64_t) + 1, CRYPTO_PACKET_NAT_PING); if (len == -1) return -1; if (type == 0) /* If packet is request use many people to route it. */ num = route_tofriend(dht, public_key, packet, len); else if (type == 1) /* If packet is response use only one person to route it */ num = routeone_tofriend(dht, public_key, packet, len); if (num == 0) return -1; return num; } /* Handle a received ping request for. */ static int handle_NATping(void *object, IP_Port source, const uint8_t *source_pubkey, const uint8_t *packet, uint32_t length) { if (length != sizeof(uint64_t) + 1) return 1; DHT *dht = object; uint64_t ping_id; memcpy(&ping_id, packet + 1, sizeof(uint64_t)); int friendnumber = friend_number(dht, source_pubkey); if (friendnumber == -1) return 1; DHT_Friend *friend = &dht->friends_list[friendnumber]; if (packet[0] == NAT_PING_REQUEST) { /* 1 is reply */ send_NATping(dht, source_pubkey, ping_id, NAT_PING_RESPONSE); friend->nat.recvNATping_timestamp = unix_time(); return 0; } else if (packet[0] == NAT_PING_RESPONSE) { if (friend->nat.NATping_id == ping_id) { friend->nat.NATping_id = random_64b(); friend->nat.hole_punching = 1; return 0; } } return 1; } /* Get the most common ip in the ip_portlist. * Only return ip if it appears in list min_num or more. * len must not be bigger than MAX_FRIEND_CLIENTS. * * return ip of 0 if failure. */ static IP NAT_commonip(IP_Port *ip_portlist, uint16_t len, uint16_t min_num) { IP zero; ip_reset(&zero); if (len > MAX_FRIEND_CLIENTS) return zero; uint32_t i, j; uint16_t numbers[MAX_FRIEND_CLIENTS] = {0}; for (i = 0; i < len; ++i) { for (j = 0; j < len; ++j) { if (ip_equal(&ip_portlist[i].ip, &ip_portlist[j].ip)) ++numbers[i]; } if (numbers[i] >= min_num) return ip_portlist[i].ip; } return zero; } /* Return all the ports for one ip in a list. * portlist must be at least len long, * where len is the length of ip_portlist. * * return number of ports and puts the list of ports in portlist. */ static uint16_t NAT_getports(uint16_t *portlist, IP_Port *ip_portlist, uint16_t len, IP ip) { uint32_t i; uint16_t num = 0; for (i = 0; i < len; ++i) { if (ip_equal(&ip_portlist[i].ip, &ip)) { portlist[num] = ntohs(ip_portlist[i].port); ++num; } } return num; } static void punch_holes(DHT *dht, IP ip, uint16_t *port_list, uint16_t numports, uint16_t friend_num) { if (numports > MAX_FRIEND_CLIENTS || numports == 0) return; uint32_t i; uint32_t top = dht->friends_list[friend_num].nat.punching_index + MAX_PUNCHING_PORTS; uint16_t firstport = port_list[0]; for (i = 0; i < numports; ++i) { if (firstport != port_list[i]) break; } if (i == numports) { /* If all ports are the same, only try that one port. */ IP_Port pinging; ip_copy(&pinging.ip, &ip); pinging.port = htons(firstport); send_ping_request(dht->ping, pinging, dht->friends_list[friend_num].client_id); } else { for (i = dht->friends_list[friend_num].nat.punching_index; i != top; ++i) { /* TODO: Improve port guessing algorithm. */ uint16_t port = port_list[(i / 2) % numports] + (i / (2 * numports)) * ((i % 2) ? -1 : 1); IP_Port pinging; ip_copy(&pinging.ip, &ip); pinging.port = htons(port); send_ping_request(dht->ping, pinging, dht->friends_list[friend_num].client_id); } dht->friends_list[friend_num].nat.punching_index = i; } if (dht->friends_list[friend_num].nat.tries > MAX_NORMAL_PUNCHING_TRIES) { top = dht->friends_list[friend_num].nat.punching_index2 + MAX_PUNCHING_PORTS; uint16_t port = 1024; IP_Port pinging; ip_copy(&pinging.ip, &ip); for (i = dht->friends_list[friend_num].nat.punching_index2; i != top; ++i) { pinging.port = htons(port + i); send_ping_request(dht->ping, pinging, dht->friends_list[friend_num].client_id); } dht->friends_list[friend_num].nat.punching_index2 = i - (MAX_PUNCHING_PORTS / 2); } ++dht->friends_list[friend_num].nat.tries; } static void do_NAT(DHT *dht) { uint32_t i; uint64_t temp_time = unix_time(); for (i = 0; i < dht->num_friends; ++i) { IP_Port ip_list[MAX_FRIEND_CLIENTS]; int num = friend_iplist(dht, ip_list, i); /* If already connected or friend is not online don't try to hole punch. */ if (num < MAX_FRIEND_CLIENTS / 2) continue; if (dht->friends_list[i].nat.NATping_timestamp + PUNCH_INTERVAL < temp_time) { send_NATping(dht, dht->friends_list[i].client_id, dht->friends_list[i].nat.NATping_id, NAT_PING_REQUEST); dht->friends_list[i].nat.NATping_timestamp = temp_time; } if (dht->friends_list[i].nat.hole_punching == 1 && dht->friends_list[i].nat.punching_timestamp + PUNCH_INTERVAL < temp_time && dht->friends_list[i].nat.recvNATping_timestamp + PUNCH_INTERVAL * 2 >= temp_time) { IP ip = NAT_commonip(ip_list, num, MAX_FRIEND_CLIENTS / 2); if (!ip_isset(&ip)) continue; uint16_t port_list[MAX_FRIEND_CLIENTS]; uint16_t numports = NAT_getports(port_list, ip_list, num, ip); punch_holes(dht, ip, port_list, numports, i); dht->friends_list[i].nat.punching_timestamp = temp_time; dht->friends_list[i].nat.hole_punching = 0; } } } /*----------------------------------------------------------------------------------*/ /*-----------------------END OF NAT PUNCHING FUNCTIONS------------------------------*/ #define HARDREQ_DATA_SIZE 384 /* Attempt to prevent amplification/other attacks*/ #define CHECK_TYPE_ROUTE_REQ 0 #define CHECK_TYPE_ROUTE_RES 1 #define CHECK_TYPE_GETNODE_REQ 2 #define CHECK_TYPE_GETNODE_RES 3 #define CHECK_TYPE_TEST_REQ 4 #define CHECK_TYPE_TEST_RES 5 static int send_hardening_req(DHT *dht, Node_format *sendto, uint8_t type, uint8_t *contents, uint16_t length) { if (length > HARDREQ_DATA_SIZE - 1) return -1; uint8_t packet[MAX_CRYPTO_REQUEST_SIZE]; uint8_t data[HARDREQ_DATA_SIZE] = {0}; data[0] = type; memcpy(data + 1, contents, length); int len = create_request(dht->self_public_key, dht->self_secret_key, packet, sendto->client_id, data, sizeof(data), CRYPTO_PACKET_HARDENING); if (len == -1) return -1; return sendpacket(dht->net, sendto->ip_port, packet, len); } /* Send a get node hardening request */ static int send_hardening_getnode_req(DHT *dht, Node_format *dest, Node_format *node_totest, uint8_t *search_id) { uint8_t data[sizeof(Node_format) + CLIENT_ID_SIZE]; memcpy(data, node_totest, sizeof(Node_format)); memcpy(data + sizeof(Node_format), search_id, CLIENT_ID_SIZE); return send_hardening_req(dht, dest, CHECK_TYPE_GETNODE_REQ, data, sizeof(Node_format) + CLIENT_ID_SIZE); } /* Send a get node hardening response */ static int send_hardening_getnode_res(const DHT *dht, const Node_format *sendto, const uint8_t *queried_client_id, const uint8_t *nodes_data, uint16_t nodes_data_length) { if (!ip_isset(&sendto->ip_port.ip)) return -1; uint8_t packet[MAX_CRYPTO_REQUEST_SIZE]; uint8_t data[1 + CLIENT_ID_SIZE + nodes_data_length]; data[0] = CHECK_TYPE_GETNODE_RES; memcpy(data + 1, queried_client_id, CLIENT_ID_SIZE); memcpy(data + 1 + CLIENT_ID_SIZE, nodes_data, nodes_data_length); int len = create_request(dht->self_public_key, dht->self_secret_key, packet, sendto->client_id, data, sizeof(data), CRYPTO_PACKET_HARDENING); if (len == -1) return -1; return sendpacket(dht->net, sendto->ip_port, packet, len); } /* TODO: improve */ static IPPTsPng *get_closelist_IPPTsPng(DHT *dht, const uint8_t *client_id, sa_family_t sa_family) { uint32_t i; for (i = 0; i < LCLIENT_LIST; ++i) { if (memcmp(dht->close_clientlist[i].client_id, client_id, CLIENT_ID_SIZE) != 0) continue; if (sa_family == AF_INET) return &dht->close_clientlist[i].assoc4; else if (sa_family == AF_INET6) return &dht->close_clientlist[i].assoc6; } return NULL; } /* * check how many nodes in nodes are also present in the closelist. * TODO: make this function better. */ static uint32_t have_nodes_closelist(DHT *dht, Node_format *nodes, uint16_t num) { uint32_t counter = 0; uint32_t i; for (i = 0; i < num; ++i) { if (id_equal(nodes[i].client_id, dht->self_public_key)) { ++counter; continue; } IPPTsPng *temp = get_closelist_IPPTsPng(dht, nodes[i].client_id, nodes[i].ip_port.ip.family); if (temp) { if (!is_timeout(temp->timestamp, BAD_NODE_TIMEOUT)) { ++counter; } } } return counter; } /* Interval in seconds between hardening checks */ #define HARDENING_INTERVAL 120 #define HARDEN_TIMEOUT 1200 /* Handle a received hardening packet */ static int handle_hardening(void *object, IP_Port source, const uint8_t *source_pubkey, const uint8_t *packet, uint32_t length) { DHT *dht = object; if (length < 2) { return 1; } switch (packet[0]) { case CHECK_TYPE_GETNODE_REQ: { if (length != HARDREQ_DATA_SIZE) return 1; Node_format node, tocheck_node; node.ip_port = source; memcpy(node.client_id, source_pubkey, CLIENT_ID_SIZE); memcpy(&tocheck_node, packet + 1, sizeof(Node_format)); if (getnodes(dht, tocheck_node.ip_port, tocheck_node.client_id, packet + 1 + sizeof(Node_format), &node) == -1) return 1; return 0; } case CHECK_TYPE_GETNODE_RES: { if (length <= CLIENT_ID_SIZE + 1) return 1; if (length > 1 + CLIENT_ID_SIZE + sizeof(Node_format) * MAX_SENT_NODES) return 1; uint16_t length_nodes = length - 1 - CLIENT_ID_SIZE; Node_format nodes[MAX_SENT_NODES]; int num_nodes = unpack_nodes(nodes, MAX_SENT_NODES, 0, packet + 1 + CLIENT_ID_SIZE, length_nodes, 0); /* TODO: MAX_SENT_NODES nodes should be returned at all times (right now we have a small network size so it could cause problems for testing and etc..) */ if (num_nodes <= 0) return 1; /* NOTE: This should work for now but should be changed to something better. */ if (have_nodes_closelist(dht, nodes, num_nodes) < (uint32_t)((num_nodes + 2) / 2)) return 1; IPPTsPng *temp = get_closelist_IPPTsPng(dht, packet + 1, nodes[0].ip_port.ip.family); if (temp == NULL) return 1; if (is_timeout(temp->hardening.send_nodes_timestamp, HARDENING_INTERVAL)) return 1; if (memcmp(temp->hardening.send_nodes_pingedid, source_pubkey, CLIENT_ID_SIZE) != 0) return 1; /* If Nodes look good and the request checks out */ temp->hardening.send_nodes_ok = 1; return 0;/* success*/ } } return 1; } /* Return a random node from all the nodes we are connected to. * TODO: improve this function. */ Node_format random_node(DHT *dht, sa_family_t sa_family) { uint8_t id[CLIENT_ID_SIZE]; uint32_t i; for (i = 0; i < CLIENT_ID_SIZE / 4; ++i) { /* populate the id with pseudorandom bytes.*/ uint32_t t = rand(); memcpy(id + i * sizeof(t), &t, sizeof(t)); } Node_format nodes_list[MAX_SENT_NODES]; memset(nodes_list, 0, sizeof(nodes_list)); uint32_t num_nodes = get_close_nodes(dht, id, nodes_list, sa_family, 1, 0); if (num_nodes == 0) return nodes_list[0]; else return nodes_list[rand() % num_nodes]; } /* Put up to max_num nodes in nodes from the closelist. * * return the number of nodes. */ uint16_t closelist_nodes(DHT *dht, Node_format *nodes, uint16_t max_num) { if (max_num == 0) return 0; uint16_t count = 0; Client_data *list = dht->close_clientlist; uint32_t i; for (i = LCLIENT_LIST; i != 0; --i) { IPPTsPng *assoc = NULL; if (!is_timeout(list[i - 1].assoc4.timestamp, BAD_NODE_TIMEOUT)) assoc = &list[i - 1].assoc4; if (!is_timeout(list[i - 1].assoc6.timestamp, BAD_NODE_TIMEOUT)) { if (assoc == NULL) assoc = &list[i - 1].assoc6; else if (rand() % 2) assoc = &list[i - 1].assoc6; } if (assoc != NULL) { memcpy(nodes[count].client_id, list[i - 1].client_id, CLIENT_ID_SIZE); nodes[count].ip_port = assoc->ip_port; ++count; if (count >= max_num) return count; } } return count; } /* Put a random node from list of list_size in node. LAN_ok is 1 if LAN ips are ok, 0 if we don't want them. */ static int random_node_fromlist(Client_data *list, uint16_t list_size, Node_format *node, uint8_t LAN_ok) { uint32_t i; uint32_t num_nodes = 0; Client_data *client_list[list_size * 2]; IPPTsPng *assoc_list[list_size * 2]; for (i = 0; i < list_size; i++) { /* If node is not dead. */ Client_data *client = &list[i]; IPPTsPng *assoc; uint32_t a; for (a = 0, assoc = &client->assoc6; a < 2; a++, assoc = &client->assoc4) { /* If node is good. */ if (!is_timeout(assoc->timestamp, BAD_NODE_TIMEOUT)) { if (!LAN_ok) { if (LAN_ip(assoc->ip_port.ip) == 0) continue; } client_list[num_nodes] = client; assoc_list[num_nodes] = assoc; ++num_nodes; } } } if (num_nodes == 0) return -1; uint32_t rand_node = rand() % num_nodes; node->ip_port = assoc_list[rand_node]->ip_port; memcpy(node->client_id, client_list[rand_node]->client_id, CLIENT_ID_SIZE); return 0; } /* Put up to max_num random nodes in nodes. * * return the number of nodes. * * NOTE:this is used to pick nodes for paths. * * TODO: remove the LAN stuff from this. */ uint16_t random_nodes_path(DHT *dht, Node_format *nodes, uint16_t max_num) { if (max_num == 0) return 0; if (dht->num_friends == 0) return 0; uint16_t count = 0; Client_data *list = NULL; uint16_t list_size = 0; uint32_t i; for (i = 0; i < max_num; ++i) { uint16_t rand_num = rand() % (dht->num_friends); list = dht->friends_list[rand_num].client_list; list_size = MAX_FRIEND_CLIENTS; uint8_t LAN_ok = 1; if (count != 0 && LAN_ip(nodes[0].ip_port.ip) != 0) LAN_ok = 0; if (random_node_fromlist(list, list_size, &nodes[count], LAN_ok) == 0) ++count; } return count; } void do_hardening(DHT *dht) { uint32_t i; for (i = 0; i < LCLIENT_LIST * 2; ++i) { IPPTsPng *cur_iptspng; sa_family_t sa_family; uint8_t *client_id = dht->close_clientlist[i / 2].client_id; if (i % 2 == 0) { cur_iptspng = &dht->close_clientlist[i / 2].assoc4; sa_family = AF_INET; } else { cur_iptspng = &dht->close_clientlist[i / 2].assoc6; sa_family = AF_INET6; } if (is_timeout(cur_iptspng->timestamp, BAD_NODE_TIMEOUT)) continue; if (cur_iptspng->hardening.send_nodes_ok == 0) { if (is_timeout(cur_iptspng->hardening.send_nodes_timestamp, HARDENING_INTERVAL)) { Node_format rand_node = random_node(dht, sa_family); if (!ipport_isset(&rand_node.ip_port)) continue; if (id_equal(client_id, rand_node.client_id)) continue; Node_format to_test; to_test.ip_port = cur_iptspng->ip_port; memcpy(to_test.client_id, client_id, CLIENT_ID_SIZE); //TODO: The search id should maybe not be ours? if (send_hardening_getnode_req(dht, &rand_node, &to_test, dht->self_public_key) > 0) { memcpy(cur_iptspng->hardening.send_nodes_pingedid, rand_node.client_id, CLIENT_ID_SIZE); cur_iptspng->hardening.send_nodes_timestamp = unix_time(); } } } else { if (is_timeout(cur_iptspng->hardening.send_nodes_timestamp, HARDEN_TIMEOUT)) { cur_iptspng->hardening.send_nodes_ok = 0; } } //TODO: add the 2 other testers. } } /*----------------------------------------------------------------------------------*/ void cryptopacket_registerhandler(DHT *dht, uint8_t byte, cryptopacket_handler_callback cb, void *object) { dht->cryptopackethandlers[byte].function = cb; dht->cryptopackethandlers[byte].object = object; } static int cryptopacket_handle(void *object, IP_Port source, const uint8_t *packet, uint32_t length) { DHT *dht = object; if (packet[0] == NET_PACKET_CRYPTO) { if (length <= crypto_box_PUBLICKEYBYTES * 2 + crypto_box_NONCEBYTES + 1 + crypto_box_MACBYTES || length > MAX_CRYPTO_REQUEST_SIZE + crypto_box_MACBYTES) return 1; if (memcmp(packet + 1, dht->self_public_key, crypto_box_PUBLICKEYBYTES) == 0) { // Check if request is for us. uint8_t public_key[crypto_box_PUBLICKEYBYTES]; uint8_t data[MAX_CRYPTO_REQUEST_SIZE]; uint8_t number; int len = handle_request(dht->self_public_key, dht->self_secret_key, public_key, data, &number, packet, length); if (len == -1 || len == 0) return 1; if (!dht->cryptopackethandlers[number].function) return 1; return dht->cryptopackethandlers[number].function(dht->cryptopackethandlers[number].object, source, public_key, data, len); } else { /* If request is not for us, try routing it. */ int retval = route_packet(dht, packet + 1, packet, length); if ((unsigned int)retval == length) return 0; } } return 1; } /*----------------------------------------------------------------------------------*/ DHT *new_DHT(Networking_Core *net) { /* init time */ unix_time_update(); if (net == NULL) return NULL; DHT *dht = calloc(1, sizeof(DHT)); if (dht == NULL) return NULL; dht->net = net; dht->ping = new_ping(dht); if (dht->ping == NULL) { kill_DHT(dht); return NULL; } networking_registerhandler(dht->net, NET_PACKET_GET_NODES, &handle_getnodes, dht); networking_registerhandler(dht->net, NET_PACKET_SEND_NODES_IPV6, &handle_sendnodes_ipv6, dht); networking_registerhandler(dht->net, NET_PACKET_CRYPTO, &cryptopacket_handle, dht); cryptopacket_registerhandler(dht, CRYPTO_PACKET_NAT_PING, &handle_NATping, dht); cryptopacket_registerhandler(dht, CRYPTO_PACKET_HARDENING, &handle_hardening, dht); new_symmetric_key(dht->secret_symmetric_key); crypto_box_keypair(dht->self_public_key, dht->self_secret_key); ping_array_init(&dht->dht_ping_array, DHT_PING_ARRAY_SIZE, PING_TIMEOUT); ping_array_init(&dht->dht_harden_ping_array, DHT_PING_ARRAY_SIZE, PING_TIMEOUT); #ifdef ENABLE_ASSOC_DHT dht->assoc = new_Assoc_default(dht->self_public_key); #endif uint32_t i; for (i = 0; i < DHT_FAKE_FRIEND_NUMBER; ++i) { uint8_t random_key_bytes[CLIENT_ID_SIZE]; randombytes(random_key_bytes, sizeof(random_key_bytes)); DHT_addfriend(dht, random_key_bytes); } return dht; } void do_DHT(DHT *dht) { unix_time_update(); if (dht->last_run == unix_time()) { return; } do_Close(dht); do_DHT_friends(dht); do_NAT(dht); do_to_ping(dht->ping); do_hardening(dht); #ifdef ENABLE_ASSOC_DHT if (dht->assoc) do_Assoc(dht->assoc, dht); #endif dht->last_run = unix_time(); } void kill_DHT(DHT *dht) { #ifdef ENABLE_ASSOC_DHT kill_Assoc(dht->assoc); #endif networking_registerhandler(dht->net, NET_PACKET_GET_NODES, NULL, NULL); networking_registerhandler(dht->net, NET_PACKET_SEND_NODES, NULL, NULL); networking_registerhandler(dht->net, NET_PACKET_SEND_NODES_IPV6, NULL, NULL); cryptopacket_registerhandler(dht, CRYPTO_PACKET_NAT_PING, NULL, NULL); cryptopacket_registerhandler(dht, CRYPTO_PACKET_HARDENING, NULL, NULL); ping_array_free_all(&dht->dht_ping_array); ping_array_free_all(&dht->dht_harden_ping_array); kill_ping(dht->ping); free(dht->friends_list); free(dht); } /* new DHT format for load/save, more robust and forward compatible */ #define DHT_STATE_COOKIE_GLOBAL 0x159000d #define DHT_STATE_COOKIE_TYPE 0x11ce #define DHT_STATE_TYPE_FRIENDS_ASSOC46 3 #define DHT_STATE_TYPE_CLIENTS_ASSOC46 4 /* Get the size of the DHT (for saving). */ uint32_t DHT_size(const DHT *dht) { uint32_t num = 0, i; for (i = 0; i < LCLIENT_LIST; ++i) if ((dht->close_clientlist[i].assoc4.timestamp != 0) || (dht->close_clientlist[i].assoc6.timestamp != 0)) num++; uint32_t size32 = sizeof(uint32_t), sizesubhead = size32 * 2; return size32 + sizesubhead + sizeof(DHT_Friend) * dht->num_friends + sizesubhead + sizeof(Client_data) * num; } static uint8_t *z_state_save_subheader(uint8_t *data, uint32_t len, uint16_t type) { uint32_t *data32 = (uint32_t *)data; data32[0] = len; data32[1] = (DHT_STATE_COOKIE_TYPE << 16) | type; data += sizeof(uint32_t) * 2; return data; } /* Save the DHT in data where data is an array of size DHT_size(). */ void DHT_save(DHT *dht, uint8_t *data) { uint32_t len; uint16_t type; *(uint32_t *)data = DHT_STATE_COOKIE_GLOBAL; data += sizeof(uint32_t); len = sizeof(DHT_Friend) * dht->num_friends; type = DHT_STATE_TYPE_FRIENDS_ASSOC46; data = z_state_save_subheader(data, len, type); memcpy(data, dht->friends_list, len); data += len; uint32_t num = 0, i; for (i = 0; i < LCLIENT_LIST; ++i) if ((dht->close_clientlist[i].assoc4.timestamp != 0) || (dht->close_clientlist[i].assoc6.timestamp != 0)) num++; len = num * sizeof(Client_data); type = DHT_STATE_TYPE_CLIENTS_ASSOC46; data = z_state_save_subheader(data, len, type); if (num) { Client_data *clients = (Client_data *)data; for (num = 0, i = 0; i < LCLIENT_LIST; ++i) if ((dht->close_clientlist[i].assoc4.timestamp != 0) || (dht->close_clientlist[i].assoc6.timestamp != 0)) memcpy(&clients[num++], &dht->close_clientlist[i], sizeof(Client_data)); } } static int dht_load_state_callback(void *outer, const uint8_t *data, uint32_t length, uint16_t type) { DHT *dht = outer; uint32_t num, i, j; switch (type) { case DHT_STATE_TYPE_FRIENDS_ASSOC46: if (length % sizeof(DHT_Friend) != 0) break; { /* localize declarations */ DHT_Friend *friend_list = (DHT_Friend *)data; num = length / sizeof(DHT_Friend); for (i = 0; i < num; ++i) { for (j = 0; j < MAX_FRIEND_CLIENTS; ++j) { Client_data *client = &friend_list[i].client_list[j]; if (client->assoc4.timestamp != 0) getnodes(dht, client->assoc4.ip_port, client->client_id, friend_list[i].client_id, NULL); if (client->assoc6.timestamp != 0) getnodes(dht, client->assoc6.ip_port, client->client_id, friend_list[i].client_id, NULL); } } } /* localize declarations */ break; case DHT_STATE_TYPE_CLIENTS_ASSOC46: if ((length % sizeof(Client_data)) != 0) break; { /* localize declarations */ num = length / sizeof(Client_data); Client_data *client_list = (Client_data *)data; for (i = 0; i < num; ++i) { if (client_list[i].assoc4.timestamp != 0) DHT_bootstrap(dht, client_list[i].assoc4.ip_port, client_list[i].client_id); if (client_list[i].assoc6.timestamp != 0) DHT_bootstrap(dht, client_list[i].assoc6.ip_port, client_list[i].client_id); } } /* localize declarations */ break; #ifdef DEBUG default: fprintf(stderr, "Load state (DHT): contains unrecognized part (len %u, type %u)\n", length, type); break; #endif } return 0; } /* Load the DHT from data of size size. * * return -1 if failure. * return 0 if success. */ int DHT_load(DHT *dht, const uint8_t *data, uint32_t length) { uint32_t cookie_len = sizeof(uint32_t); if (length > cookie_len) { uint32_t *data32 = (uint32_t *)data; if (data32[0] == DHT_STATE_COOKIE_GLOBAL) return load_state(dht_load_state_callback, dht, data + cookie_len, length - cookie_len, DHT_STATE_COOKIE_TYPE); } return -1; } /* return 0 if we are not connected to the DHT. * return 1 if we are. */ int DHT_isconnected(const DHT *dht) { uint32_t i; unix_time_update(); for (i = 0; i < LCLIENT_LIST; ++i) { const Client_data *client = &dht->close_clientlist[i]; if (!is_timeout(client->assoc4.timestamp, BAD_NODE_TIMEOUT) || !is_timeout(client->assoc6.timestamp, BAD_NODE_TIMEOUT)) return 1; } return 0; }