transmission/libtransmission/tracker.c

1127 lines
30 KiB
C

/******************************************************************************
* $Id$
*
* Copyright (c) 2005-2006 Transmission authors and contributors
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*****************************************************************************/
#include "transmission.h"
#include "shared.h"
struct tclist
{
tr_tracker_info_t * tl_inf;
int tl_badscrape;
SLIST_ENTRY( tclist ) next;
};
SLIST_HEAD( tchead, tclist );
struct tr_tracker_s
{
tr_torrent_t * tor;
char * id;
char * trackerid;
struct tchead * tiers;
size_t tierCount;
size_t tierIndex;
struct tclist * tcCur;
#define TC_CHANGE_NO 0
#define TC_CHANGE_NEXT 1
#define TC_CHANGE_NONEXT 2
#define TC_CHANGE_REDIRECT 4
int shouldChangeAnnounce;
char * redirectAddress;
int redirectAddressLen;
char * redirectScrapeAddress;
int redirectScrapeAddressLen;
char started;
char completed;
char stopped;
int interval;
int minInterval;
int scrapeInterval;
int seeders;
int leechers;
int hasManyPeers;
int complete;
int randOffset;
int completelyUnconnectable;
int allUnreachIfError;
int lastError;
uint64_t dateTry;
uint64_t dateOk;
uint64_t dateScrape;
int lastScrapeFailed;
int scrapeNeeded;
tr_http_t * http;
tr_http_t * httpScrape;
int publicPort;
};
static void setAnnounce ( tr_tracker_t * tc, struct tclist * new );
static void failureAnnouncing( tr_tracker_t * tc );
static tr_http_t * getQuery ( tr_tracker_t * tc );
static tr_http_t * getScrapeQuery ( tr_tracker_t * tc );
static void readAnswer ( tr_tracker_t * tc, const char *, int,
int * peerCount, uint8_t ** peerCompact );
static void readScrapeAnswer ( tr_tracker_t * tc, const char *, int );
static void killHttp ( tr_http_t ** http );
static int shouldChangePort( tr_tracker_t * tc );
static uint8_t * parseOriginalPeers( benc_val_t * bePeers, int * peerCount );
tr_tracker_t * tr_trackerInit( tr_torrent_t * tor )
{
tr_info_t * inf = &tor->info;
tr_tracker_t * tc;
struct tclist * item;
size_t ii, jj;
tc = calloc( 1, sizeof *tc );
if( NULL == tc )
{
return NULL;
}
tc->tor = tor;
tc->id = tor->id;
tc->started = 1;
tc->shouldChangeAnnounce = TC_CHANGE_NO;
tc->redirectAddress = NULL;
tc->interval = 300;
tc->scrapeInterval = 1200;
tc->lastError = 0;
tc->allUnreachIfError = 1;
tc->publicPort = tor->publicPort;
assert( 0 <= inf->trackerTiers );
assert( sizeof( struct tchead ) == sizeof *tc->tiers );
tc->tiers = calloc( inf->trackerTiers, sizeof *tc->tiers );
tc->tierCount = inf->trackerTiers;
for( ii = 0; tc->tierCount > ii; ii++ )
{
assert( 0 <= inf->trackerList[ii].count );
SLIST_INIT( &tc->tiers[ii] );
for( jj = inf->trackerList[ii].count; 0 < jj; jj-- )
{
item = calloc( 1, sizeof *item );
if( NULL == item )
{
tr_trackerClose( tc );
return NULL;
}
item->tl_inf = &inf->trackerList[ii].list[jj-1];
SLIST_INSERT_HEAD( &tc->tiers[ii], item, next );
}
}
setAnnounce( tc, SLIST_FIRST( &tc->tiers[0] ) );
return tc;
}
static void setAnnounce( tr_tracker_t * tc, struct tclist * new )
{
tc->tcCur = new;
/* Needs a new scrape */
tc->seeders = -1;
tc->leechers = -1;
tc->complete = -1;
tc->dateScrape = 0;
}
static void failureAnnouncing( tr_tracker_t * tc )
{
if( NULL != SLIST_NEXT( tc->tcCur, next ) ||
tc->tierIndex + 1 < tc->tierCount )
{
tc->shouldChangeAnnounce = TC_CHANGE_NEXT;
}
else
{
tc->shouldChangeAnnounce = TC_CHANGE_NONEXT;
tc->completelyUnconnectable = 1;
}
}
static int shouldConnect( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
uint64_t now;
/* Last tracker failed, try next */
if( tc->shouldChangeAnnounce == TC_CHANGE_NEXT
|| tc->shouldChangeAnnounce == TC_CHANGE_REDIRECT )
{
return 1;
}
now = tr_date();
/* If last attempt was an error and it did not change trackers,
then all must have been errors */
if( tc->lastError )
{
/* Unreachable trackers, wait 10 seconds + random value before
trying again */
if( tc->allUnreachIfError )
{
if( now < tc->dateTry + tc->randOffset + 10000 )
{
return 0;
}
}
/* The tracker rejected us (like 4XX code, unauthorized
IP...), don't hammer it - we'll probably get the same
answer next time anyway */
else
{
if( now < tc->dateTry + 1000 * tc->interval + tc->randOffset )
{
return 0;
}
else
{
/* since starting at the top of the list, reset if any
were reached previously */
tc->allUnreachIfError = 1;
}
}
}
/* Do we need to send an event? */
if( tc->started || tc->completed || tc->stopped || shouldChangePort( tc ) )
{
return 1;
}
/* Should we try and get more peers? */
if( now > tc->dateOk + 1000 * tc->interval + tc->randOffset )
{
return 1;
}
/* If there is quite a lot of people on this torrent, stress
the tracker a bit until we get a decent number of peers */
if( tc->hasManyPeers && !tr_cpIsSeeding( tor->completion ) )
{
/* reannounce in 10 seconds if we have less than 5 peers */
if( tor->peerCount < 5 )
{
if( now > tc->dateOk + 1000 * MAX( 10, tc->minInterval ) )
{
return 1;
}
}
/* reannounce in 20 seconds if we have less than 10 peers */
else if( tor->peerCount < 10 )
{
if( now > tc->dateOk + 1000 * MAX( 20, tc->minInterval ) )
{
return 1;
}
}
/* reannounce in 30 seconds if we have less than 15 peers */
else if( tor->peerCount < 15 )
{
if( now > tc->dateOk + 1000 * MAX( 30, tc->minInterval ) )
{
return 1;
}
}
}
return 0;
}
static int shouldScrape( tr_tracker_t * tc )
{
uint64_t now, interval;
/* in process of changing tracker or scrape not supported */
if( tc->shouldChangeAnnounce != TC_CHANGE_NO ||
NULL == tc->tcCur->tl_inf->scrape || tc->tcCur->tl_badscrape ||
tc->stopped )
{
return 0;
}
now = tr_date();
interval = 1000 * MAX( tc->scrapeInterval, 600 );
/* scrape more often if needed */
if( tc->scrapeNeeded || tc->lastScrapeFailed )
{
interval /= 2;
}
return now > tc->dateScrape + interval;
}
void tr_trackerAnnouncePulse( tr_tracker_t * tc, int * peerCount,
uint8_t ** peerCompact, int manual )
{
const char * data;
char * address, * announce;
int len, port;
struct tclist * next;
struct tchead * tier;
*peerCount = 0;
*peerCompact = NULL;
if( ( NULL == tc->http ) && ( manual || shouldConnect( tc ) ) )
{
/* if announcing manually, don't consider not reaching a
tracker an error */
if( manual )
{
tc->allUnreachIfError = 0;
}
tc->completelyUnconnectable = 0;
tc->randOffset = tr_rand( 60000 );
tc->dateTry = tr_date();
/* Use redirected address */
if( tc->shouldChangeAnnounce == TC_CHANGE_REDIRECT )
{
if( !tr_httpParseUrl( tc->redirectAddress, tc->redirectAddressLen,
&address, &port, &announce ) )
{
tr_err( "Tracker: redirected URL: %s:%d", address, port );
tc->http = tr_httpClient( TR_HTTP_GET, address, port,
"%s", announce );
free( address );
free( announce );
}
free( tc->redirectAddress );
tc->redirectAddress = NULL;
}
else
{
/* Need to change to next address in list */
if( tc->shouldChangeAnnounce == TC_CHANGE_NEXT )
{
tr_inf( "Tracker: failed to connect to %s:%i, trying next",
tc->tcCur->tl_inf->address, tc->tcCur->tl_inf->port );
next = SLIST_NEXT( tc->tcCur, next );
if( NULL == next )
{
assert( tc->tierCount > tc->tierIndex + 1 );
tc->tierIndex++;
next = SLIST_FIRST( &tc->tiers[tc->tierIndex] );
/* XXX will there always be at least one tracker
in a tier? */
}
tr_inf( "Tracker: switching to tracker http://%s:%i%s",
next->tl_inf->address, next->tl_inf->port,
next->tl_inf->announce );
setAnnounce( tc, next );
}
/* Need to change to first in list */
else if( SLIST_FIRST( &tc->tiers[0] ) != tc->tcCur )
{
tier = &tc->tiers[tc->tierIndex];
/* Check if the last announce was successful and
wasn't the first in the sublist */
if( tc->shouldChangeAnnounce == TC_CHANGE_NO &&
SLIST_FIRST( tier ) != tc->tcCur )
{
SLIST_REMOVE( tier, tc->tcCur, tclist, next );
SLIST_INSERT_HEAD( tier, tc->tcCur, next );
}
setAnnounce( tc, SLIST_FIRST( tier ) );
}
tc->http = getQuery( tc );
tr_inf( "Tracker: connecting to %s:%d (%s)",
tc->tcCur->tl_inf->address, tc->tcCur->tl_inf->port,
tc->started ? "sending 'started'" :
( tc->completed ? "sending 'completed'" :
( tc->stopped ? "sending 'stopped'" :
( shouldChangePort( tc ) ?
"sending 'stopped' to change port" :
"getting peers" ) ) ) );
}
tc->shouldChangeAnnounce = TC_CHANGE_NO;
}
if( NULL != tc->http )
{
switch( tr_httpPulse( tc->http, &data, &len ) )
{
case TR_NET_WAIT:
break;
case TR_NET_ERROR:
killHttp( &tc->http );
tc->dateTry = tr_date();
failureAnnouncing( tc );
tc->lastError = 1;
break;
case TR_NET_OK:
readAnswer( tc, data, len, peerCount, peerCompact );
killHttp( &tc->http );
break;
}
}
if( ( NULL == tc->httpScrape ) && shouldScrape( tc ) )
{
tc->dateScrape = tr_date();
if ( tc->redirectScrapeAddress != NULL )
{
/* Use redirected address */
if( !tr_httpParseUrl( tc->redirectScrapeAddress,
tc->redirectScrapeAddressLen,
&address, &port, &announce ) )
{
tr_err( "Scrape: redirected URL: %s:%d", address, port );
tc->httpScrape = tr_httpClient( TR_HTTP_GET, address, port,
"%s", announce );
free( address );
free( announce );
}
free( tc->redirectScrapeAddress );
tc->redirectScrapeAddress = NULL;
}
else
{
tc->httpScrape = getScrapeQuery( tc );
tr_inf( "Scrape: sent HTTP request for http://%s:%d%s",
tc->tcCur->tl_inf->address,
tc->tcCur->tl_inf->port,
tc->tcCur->tl_inf->scrape );
}
}
if( NULL != tc->httpScrape )
{
switch( tr_httpPulse( tc->httpScrape, &data, &len ) )
{
case TR_NET_WAIT:
break;
case TR_NET_ERROR:
killHttp( &tc->httpScrape );
tc->lastScrapeFailed = 1;
break;
case TR_NET_OK:
readScrapeAnswer( tc, data, len );
killHttp( &tc->httpScrape );
break;
}
}
return;
}
void tr_trackerCompleted( tr_tracker_t * tc )
{
tc->started = 0;
tc->completed = 1;
tc->stopped = 0;
}
void tr_trackerStopped( tr_tracker_t * tc )
{
/* If we are already sending a query at the moment, we need to
reconnect */
killHttp( &tc->http );
tc->started = 0;
tc->completed = 0;
tc->stopped = 1;
/* Even if we have connected recently, reconnect right now */
tc->dateTry = 0;
}
void tr_trackerClose( tr_tracker_t * tc )
{
size_t ii;
struct tclist * dead;
killHttp( &tc->http );
killHttp( &tc->httpScrape );
for( ii = 0; tc->tierCount > ii; ii++ )
{
while( !SLIST_EMPTY( &tc->tiers[ii] ) )
{
dead = SLIST_FIRST( &tc->tiers[ii] );
SLIST_REMOVE_HEAD( &tc->tiers[ii], next );
free( dead );
}
}
free( tc->tiers );
free( tc->trackerid );
free( tc );
}
static tr_http_t * getQuery( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
tr_tracker_info_t * tcInf = tc->tcCur->tl_inf;
char * event, * trackerid, * idparam;
uint64_t left;
char start;
int numwant = 50;
if( tc->started )
{
event = "&event=started";
tor->downloadedPrev += tor->downloadedCur;
tor->downloadedCur = 0;
tor->uploadedPrev += tor->uploadedCur;
tor->uploadedCur = 0;
if( shouldChangePort( tc ) )
{
tc->publicPort = tor->publicPort;
}
}
else if( tc->completed )
{
event = "&event=completed";
}
else if( tc->stopped || shouldChangePort( tc ) )
{
event = "&event=stopped";
numwant = 0;
}
else
{
event = "";
}
if( NULL == tc->trackerid )
{
trackerid = "";
idparam = "";
}
else
{
trackerid = tc->trackerid;
idparam = "&trackerid=";
}
start = ( strchr( tcInf->announce, '?' ) ? '&' : '?' );
left = tr_cpLeftBytes( tor->completion );
return tr_httpClient( TR_HTTP_GET, tcInf->address, tcInf->port,
"%s%c"
"info_hash=%s&"
"peer_id=%s&"
"port=%d&"
"uploaded=%"PRIu64"&"
"downloaded=%"PRIu64"&"
"left=%"PRIu64"&"
"compact=1&"
"numwant=%d&"
"key=%s"
"%s%s"
"%s",
tcInf->announce, start, tor->escapedHashString,
tc->id, tc->publicPort, tor->uploadedCur, tor->downloadedCur,
left, numwant, tor->key, idparam, trackerid, event );
}
static tr_http_t * getScrapeQuery( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
tr_tracker_info_t * tcInf = tc->tcCur->tl_inf;
char start;
start = ( strchr( tcInf->scrape, '?' ) ? '&' : '?' );
return tr_httpClient( TR_HTTP_GET, tcInf->address, tcInf->port,
"%s%c"
"info_hash=%s",
tcInf->scrape, start, tor->escapedHashString );
}
static void readAnswer( tr_tracker_t * tc, const char * data, int len,
int * _peerCount, uint8_t ** _peerCompact )
{
tr_torrent_t * tor = tc->tor;
int i;
int code;
benc_val_t beAll;
benc_val_t * bePeers, * beFoo;
const uint8_t * body;
int bodylen, shouldfree, scrapeNeeded;
char * address;
int peerCount;
uint8_t * peerCompact;
*_peerCount = peerCount = 0;
*_peerCompact = peerCompact = NULL;
tc->dateTry = tr_date();
code = tr_httpResponseCode( data, len );
if( 0 > code )
{
/* We don't have a valid HTTP status line */
tr_inf( "Tracker: invalid HTTP status line" );
tc->lastError = 1;
failureAnnouncing( tc );
return;
}
if( code == 301 || code == 302 )
{
tr_http_header_t hdr[] = { { "Location", NULL, 0 }, { NULL, NULL, 0 } };
tr_err( "Tracker: HTTP status code: %i", code );
tr_httpParse( data, len, hdr );
address = calloc( sizeof( char ), hdr->len+1 );
snprintf( address, hdr->len+1, "%s", hdr->data );
tc->shouldChangeAnnounce = TC_CHANGE_REDIRECT;
tc->redirectAddress = address;
tc->redirectAddressLen = hdr->len;
return;
}
if( !TR_HTTP_STATUS_OK( code ) )
{
/* we didn't get a 2xx status code */
tr_err( "Tracker: invalid HTTP status code: %i", code );
tc->lastError = 1;
tc->allUnreachIfError = 0;
failureAnnouncing( tc );
return;
}
/* find the end of the http headers */
body = (uint8_t *) tr_httpParse( data, len, NULL );
if( NULL == body )
{
tr_err( "Tracker: could not find end of HTTP headers" );
tc->lastError = 1;
failureAnnouncing( tc );
return;
}
bodylen = len - ( body - (const uint8_t*)data );
/* Find and load the dictionary */
shouldfree = 0;
for( i = 0; i < bodylen; i++ )
{
if( !tr_bencLoad( &body[i], bodylen - i, &beAll, NULL ) )
{
shouldfree = 1;
break;
}
}
if( i >= bodylen )
{
if( tc->stopped || shouldChangePort( tc ) )
{
tc->lastError = 0;
goto nodict;
}
tr_err( "Tracker: no valid dictionary found in answer" );
tc->lastError = 1;
tc->allUnreachIfError = 0;
failureAnnouncing( tc );
return;
}
/* tr_bencPrint( &beAll ); */
if( ( bePeers = tr_bencDictFind( &beAll, "failure reason" ) ) )
{
tr_err( "Tracker: Error - %s", bePeers->val.s.s );
tor->error = TR_ERROR_TC_ERROR;
snprintf( tor->errorString, sizeof( tor->errorString ),
"%s", bePeers->val.s.s );
tc->lastError = 1;
tc->allUnreachIfError = 0;
failureAnnouncing( tc );
goto cleanup;
}
else if( ( bePeers = tr_bencDictFind( &beAll, "warning message" ) ) )
{
tr_err( "Tracker: Warning - %s", bePeers->val.s.s );
tor->error = TR_ERROR_TC_WARNING;
snprintf( tor->errorString, sizeof( tor->errorString ),
"%s", bePeers->val.s.s );
}
else if( tor->error & TR_ERROR_TC_MASK )
{
tor->error = TR_OK;
}
tc->lastError = 0;
tc->allUnreachIfError = 0;
/* Get the tracker interval */
beFoo = tr_bencDictFind( &beAll, "interval" );
if( !beFoo || TYPE_INT != beFoo->type )
{
tr_err( "Tracker: no 'interval' field" );
goto cleanup;
}
tc->interval = beFoo->val.i;
tr_inf( "Tracker: interval = %d seconds", tc->interval );
tc->interval = MAX( 10, tc->interval );
/* Get the tracker minimum interval */
beFoo = tr_bencDictFind( &beAll, "min interval" );
if( beFoo && TYPE_INT == beFoo->type )
{
tc->minInterval = beFoo->val.i;
tr_inf( "Tracker: min interval = %d seconds", tc->minInterval );
tc->minInterval = MAX( 10, tc->minInterval );
if( tc->interval < tc->minInterval )
{
tc->interval = tc->minInterval;
tr_inf( "Tracker: 'interval' less than 'min interval', "
"using 'min interval'" );
}
}
else
{
tc->minInterval = 0;
tr_inf( "Tracker: no 'min interval' field" );
}
scrapeNeeded = 0;
beFoo = tr_bencDictFind( &beAll, "complete" );
if( beFoo && TYPE_INT == beFoo->type )
{
tc->seeders = beFoo->val.i;
}
else
{
scrapeNeeded = 1;
}
beFoo = tr_bencDictFind( &beAll, "incomplete" );
if( beFoo && TYPE_INT == beFoo->type )
{
tc->leechers = beFoo->val.i;
}
else
{
scrapeNeeded = 1;
}
tc->scrapeNeeded = scrapeNeeded;
if( !scrapeNeeded )
{
tc->hasManyPeers = ( tc->seeders + tc->leechers >= 50 );
}
beFoo = tr_bencDictFind( &beAll, "tracker id" );
if( beFoo )
{
free( tc->trackerid );
tc->trackerid = strdup( beFoo->val.s.s );
tr_inf( "Tracker: tracker id = %s", tc->trackerid );
}
bePeers = tr_bencDictFind( &beAll, "peers" );
if( !bePeers )
{
if( tc->stopped || shouldChangePort( tc ) )
{
goto nodict;
}
tr_err( "Tracker: no \"peers\" field" );
failureAnnouncing( tc );
goto cleanup;
}
if( TYPE_LIST == bePeers->type )
{
/* Original protocol */
if( bePeers->val.l.count > 0 )
{
peerCompact = parseOriginalPeers( bePeers, &peerCount );
}
}
else if( TYPE_STR == bePeers->type )
{
/* "Compact" extension */
if( bePeers->val.s.i >= 6 )
{
peerCount = bePeers->val.s.i / 6;
peerCompact = malloc( bePeers->val.s.i );
memcpy( peerCompact, bePeers->val.s.s, bePeers->val.s.i );
}
}
if( peerCount > 0 )
{
tr_inf( "Tracker: got %d peers", peerCount );
if( peerCount >= 50 )
{
tc->hasManyPeers = 1;
}
*_peerCount = peerCount;
*_peerCompact = peerCompact;
}
nodict:
/* Success */
tc->started = 0;
tc->completed = 0;
tc->dateOk = tr_date();
if( tc->stopped )
{
tor->status = TR_STATUS_STOPPED;
tc->stopped = 0;
}
else if( shouldChangePort( tc ) )
{
tc->started = 1;
}
cleanup:
if( shouldfree )
{
tr_bencFree( &beAll );
}
}
static void readScrapeAnswer( tr_tracker_t * tc, const char * data, int len )
{
int code;
const uint8_t * body;
int bodylen, ii;
benc_val_t scrape, * val1, * val2;
char * address;
code = tr_httpResponseCode( data, len );
if( 0 > code )
{
/* We don't have a valid HTTP status line */
tr_inf( "Scrape: invalid HTTP status line" );
tc->lastScrapeFailed = 1;
return;
}
if( code == 301 || code == 302 )
{
tr_http_header_t hdr[] = { { "Location", NULL, 0 }, { NULL, NULL, 0 } };
tr_err( "Scrape: HTTP status code: %i", code );
tr_httpParse( data, len, hdr );
address = calloc( sizeof( char ), hdr->len+1 );
snprintf( address, hdr->len+1, "%s", hdr->data );
/* Needs a new scrape */
tc->dateScrape = 0;
tc->redirectScrapeAddress = address;
tc->redirectScrapeAddressLen = hdr->len;
return;
}
if( !TR_HTTP_STATUS_OK( code ) )
{
/* we didn't get a 2xx status code */
tr_err( "Scrape: invalid HTTP status code: %i", code );
if( TR_HTTP_STATUS_FAIL_CLIENT( code ) )
{
tc->tcCur->tl_badscrape = 1;
}
tc->lastScrapeFailed = 1;
return;
}
/* find the end of the http headers */
body = (uint8_t *) tr_httpParse( data, len, NULL );
if( NULL == body )
{
tr_err( "Scrape: could not find end of HTTP headers" );
tc->lastScrapeFailed = 1;
return;
}
tc->lastScrapeFailed = 0;
bodylen = len - ( body - (const uint8_t*)data );
for( ii = 0; ii < bodylen; ii++ )
{
if( !tr_bencLoad( body + ii, bodylen - ii, &scrape, NULL ) )
{
break;
}
}
if( ii >= bodylen )
{
return;
}
val1 = tr_bencDictFind( &scrape, "files" );
if( !val1 || val1->type != TYPE_DICT || val1->val.l.count < 1 )
{
tr_bencFree( &scrape );
return;
}
val1 = &val1->val.l.vals[1];
val2 = tr_bencDictFind( val1, "complete" );
if( !val2 )
{
tr_bencFree( &scrape );
return;
}
tc->seeders = val2->val.i;
val2 = tr_bencDictFind( val1, "incomplete" );
if( !val2 )
{
tr_bencFree( &scrape );
return;
}
tc->leechers = val2->val.i;
val2 = tr_bencDictFind( val1, "downloaded" );
if( !val2 )
{
tr_bencFree( &scrape );
return;
}
tc->complete = val2->val.i;
val2 = tr_bencDictFind( &scrape, "flags" );
if( val2 )
{
val2 = tr_bencDictFind( val2, "min_request_interval" );
if( val2 )
{
tc->scrapeInterval = val2->val.i;
tr_inf( "Scrape: min_request_interval = %d seconds", tc->scrapeInterval );
}
}
tc->hasManyPeers = ( tc->seeders + tc->leechers >= 50 );
tr_bencFree( &scrape );
}
int tr_trackerSeeders( tr_tracker_t * tc )
{
if( !tc )
{
return -1;
}
return tc->seeders;
}
int tr_trackerLeechers( tr_tracker_t * tc )
{
if( !tc )
{
return -1;
}
return tc->leechers;
}
int tr_trackerDownloaded( tr_tracker_t * tc )
{
if( !tc )
{
return -1;
}
return tc->complete;
}
tr_tracker_info_t * tr_trackerGet( tr_tracker_t * tc )
{
if( !tc )
{
return NULL;
}
return tc->tcCur->tl_inf;
}
int tr_trackerCannotConnect( tr_tracker_t * tc )
{
if( !tc )
{
return 0;
}
return tc->completelyUnconnectable;
}
/* Blocking version */
int tr_trackerScrape( tr_torrent_t * tor, int * s, int * l, int * d )
{
tr_tracker_t * tc;
tr_http_t * http;
const char * data;
int len;
int ret;
tc = tr_trackerInit( tor );
if( NULL == tc->tcCur->tl_inf->scrape || tc->tcCur->tl_badscrape )
{
return 1;
}
http = getScrapeQuery( tc );
for( data = NULL; !data; tr_wait( 10 ) )
{
switch( tr_httpPulse( http, &data, &len ) )
{
case TR_NET_WAIT:
break;
case TR_NET_ERROR:
goto scrapeDone;
case TR_NET_OK:
readScrapeAnswer( tc, data, len );
goto scrapeDone;
}
}
scrapeDone:
tr_httpClose( http );
ret = 1;
if( tc->seeders > -1 && tc->leechers > -1 && tc->complete > -1 )
{
*s = tc->seeders;
*l = tc->leechers;
*d = tc->complete;
ret = 0;
}
tr_trackerClose( tc );
return ret;
}
static void killHttp( tr_http_t ** http )
{
if( NULL != *http )
{
tr_httpClose( *http );
*http = NULL;
}
}
static int shouldChangePort( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
return ( tor->publicPort != tc->publicPort );
}
/* Convert to compact form */
static uint8_t *
parseOriginalPeers( benc_val_t * bePeers, int * peerCount )
{
struct in_addr addr;
in_port_t port;
uint8_t * peerCompact;
benc_val_t * peer, * addrval, * portval;
int ii, count;
assert( TYPE_LIST == bePeers->type );
count = 0;
peerCompact = malloc( 6 * bePeers->val.l.count );
if( NULL == peerCompact )
{
return NULL;
}
for( ii = 0; bePeers->val.l.count > ii; ii++ )
{
peer = &bePeers->val.l.vals[ii];
addrval = tr_bencDictFind( peer, "ip" );
if( NULL == addrval || TYPE_STR != addrval->type ||
tr_netResolve( addrval->val.s.s, &addr ) )
{
continue;
}
memcpy( &peerCompact[6 * count], &addr, 4 );
portval = tr_bencDictFind( peer, "port" );
if( NULL == portval || TYPE_INT != portval->type ||
0 > portval->val.i || 0xffff < portval->val.i )
{
continue;
}
port = htons( portval->val.i );
memcpy( &peerCompact[6 * count + 4], &port, 2 );
count++;
}
*peerCount = count;
return peerCompact;
}