mirror of
https://github.com/transmission/transmission
synced 2024-12-25 17:17:31 +00:00
1018 lines
23 KiB
C
1018 lines
23 KiB
C
/*
|
|
* This file Copyright (C) 2007-2008 Charles Kerr <charles@transmissionbt.com>
|
|
*
|
|
* This file is licensed by the GPL version 2. Works owned by the
|
|
* Transmission project are granted a special exemption to clause 2(b)
|
|
* so that the bulk of its code can remain under the MIT license.
|
|
* This exemption does not extend to derived works not owned by
|
|
* the Transmission project.
|
|
*
|
|
* $Id$
|
|
*/
|
|
|
|
#include <assert.h>
|
|
#include <limits.h> /* INT_MAX */
|
|
#include <string.h>
|
|
#include <stdio.h>
|
|
#include <unistd.h>
|
|
|
|
#ifdef WIN32
|
|
#include <winsock2.h>
|
|
#else
|
|
#include <arpa/inet.h> /* inet_ntoa */
|
|
#endif
|
|
|
|
#include <event.h>
|
|
|
|
#include "transmission.h"
|
|
#include "session.h"
|
|
#include "bandwidth.h"
|
|
#include "crypto.h"
|
|
#include "list.h"
|
|
#include "net.h"
|
|
#include "peer-io.h"
|
|
#include "trevent.h"
|
|
#include "utils.h"
|
|
|
|
#define MAGIC_NUMBER 206745
|
|
|
|
static size_t
|
|
getPacketOverhead( size_t d )
|
|
{
|
|
/**
|
|
* http://sd.wareonearth.com/~phil/net/overhead/
|
|
*
|
|
* TCP over Ethernet:
|
|
* Assuming no header compression (e.g. not PPP)
|
|
* Add 20 IPv4 header or 40 IPv6 header (no options)
|
|
* Add 20 TCP header
|
|
* Add 12 bytes optional TCP timestamps
|
|
* Max TCP Payload data rates over ethernet are thus:
|
|
* (1500-40)/(38+1500) = 94.9285 % IPv4, minimal headers
|
|
* (1500-52)/(38+1500) = 94.1482 % IPv4, TCP timestamps
|
|
* (1500-52)/(42+1500) = 93.9040 % 802.1q, IPv4, TCP timestamps
|
|
* (1500-60)/(38+1500) = 93.6281 % IPv6, minimal headers
|
|
* (1500-72)/(38+1500) = 92.8479 % IPv6, TCP timestamps
|
|
* (1500-72)/(42+1500) = 92.6070 % 802.1q, IPv6, ICP timestamps
|
|
*/
|
|
static const double assumed_payload_data_rate = 94.0;
|
|
|
|
return (size_t)( d * ( 100.0 / assumed_payload_data_rate ) - d );
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
#define dbgmsg( io, ... ) \
|
|
do { \
|
|
if( tr_deepLoggingIsActive( ) ) \
|
|
tr_deepLog( __FILE__, __LINE__, tr_peerIoGetAddrStr( io ), __VA_ARGS__ ); \
|
|
} while( 0 )
|
|
|
|
struct tr_datatype
|
|
{
|
|
tr_bool isPieceData;
|
|
size_t length;
|
|
struct __tr_list head;
|
|
};
|
|
|
|
struct tr_peerIo
|
|
{
|
|
tr_bool isEncrypted;
|
|
tr_bool isIncoming;
|
|
tr_bool peerIdIsSet;
|
|
tr_bool extendedProtocolSupported;
|
|
tr_bool fastExtensionSupported;
|
|
|
|
int magicNumber;
|
|
|
|
uint8_t encryptionMode;
|
|
tr_port port;
|
|
int socket;
|
|
|
|
uint8_t peerId[20];
|
|
time_t timeCreated;
|
|
|
|
tr_session * session;
|
|
|
|
tr_address addr;
|
|
|
|
tr_can_read_cb canRead;
|
|
tr_did_write_cb didWrite;
|
|
tr_net_error_cb gotError;
|
|
void * userData;
|
|
|
|
tr_bandwidth * bandwidth;
|
|
tr_crypto * crypto;
|
|
|
|
struct evbuffer * inbuf;
|
|
struct evbuffer * outbuf;
|
|
struct __tr_list outbuf_datatypes; /* struct tr_datatype */
|
|
|
|
struct event event_read;
|
|
struct event event_write;
|
|
};
|
|
|
|
/***
|
|
****
|
|
***/
|
|
|
|
static void
|
|
didWriteWrapper( tr_peerIo * io, size_t bytes_transferred )
|
|
{
|
|
while( bytes_transferred )
|
|
{
|
|
struct tr_datatype * next = __tr_list_entry( io->outbuf_datatypes.next,
|
|
struct tr_datatype, head );
|
|
const size_t payload = MIN( next->length, bytes_transferred );
|
|
const size_t overhead = getPacketOverhead( payload );
|
|
|
|
tr_bandwidthUsed( io->bandwidth, TR_UP, payload, next->isPieceData );
|
|
|
|
if( overhead > 0 )
|
|
tr_bandwidthUsed( io->bandwidth, TR_UP, overhead, FALSE );
|
|
|
|
if( io->didWrite )
|
|
io->didWrite( io, payload, next->isPieceData, io->userData );
|
|
|
|
bytes_transferred -= payload;
|
|
next->length -= payload;
|
|
if( !next->length ) {
|
|
__tr_list_remove( io->outbuf_datatypes.next );
|
|
tr_free( next );
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
canReadWrapper( tr_peerIo * io )
|
|
{
|
|
tr_bool done = 0;
|
|
tr_bool err = 0;
|
|
tr_session * session = io->session;
|
|
|
|
dbgmsg( io, "canRead" );
|
|
|
|
/* try to consume the input buffer */
|
|
if( io->canRead )
|
|
{
|
|
tr_globalLock( session );
|
|
|
|
while( !done && !err )
|
|
{
|
|
size_t piece = 0;
|
|
const size_t oldLen = EVBUFFER_LENGTH( io->inbuf );
|
|
const int ret = io->canRead( io, io->userData, &piece );
|
|
|
|
const size_t used = oldLen - EVBUFFER_LENGTH( io->inbuf );
|
|
|
|
if( piece )
|
|
tr_bandwidthUsed( io->bandwidth, TR_DOWN, piece, TRUE );
|
|
|
|
if( used != piece )
|
|
tr_bandwidthUsed( io->bandwidth, TR_DOWN, used - piece, FALSE );
|
|
|
|
switch( ret )
|
|
{
|
|
case READ_NOW:
|
|
if( EVBUFFER_LENGTH( io->inbuf ) )
|
|
continue;
|
|
done = 1;
|
|
break;
|
|
|
|
case READ_LATER:
|
|
done = 1;
|
|
break;
|
|
|
|
case READ_ERR:
|
|
err = 1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
tr_globalUnlock( session );
|
|
}
|
|
}
|
|
|
|
#define _isBool(b) (((b)==0 || (b)==1))
|
|
|
|
tr_bool
|
|
tr_isPeerIo( const tr_peerIo * io )
|
|
{
|
|
return ( io != NULL )
|
|
&& ( io->magicNumber == MAGIC_NUMBER )
|
|
&& ( tr_isAddress( &io->addr ) )
|
|
&& ( _isBool( io->isEncrypted ) )
|
|
&& ( _isBool( io->isIncoming ) )
|
|
&& ( _isBool( io->peerIdIsSet ) )
|
|
&& ( _isBool( io->extendedProtocolSupported ) )
|
|
&& ( _isBool( io->fastExtensionSupported ) );
|
|
}
|
|
|
|
static void
|
|
event_read_cb( int fd, short event UNUSED, void * vio )
|
|
{
|
|
int res;
|
|
tr_peerIo * io = vio;
|
|
|
|
/* Limit the input buffer to 256K, so it doesn't grow too large */
|
|
size_t howmuch;
|
|
const tr_direction dir = TR_DOWN;
|
|
const size_t max = 256 * 1024;
|
|
const size_t curlen = EVBUFFER_LENGTH( io->inbuf );
|
|
|
|
howmuch = curlen >= max ? 0 : max - curlen;
|
|
howmuch = tr_bandwidthClamp( io->bandwidth, TR_DOWN, howmuch );
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
dbgmsg( io, "libevent says this peer is ready to read" );
|
|
|
|
/* if we don't have any bandwidth left, stop reading */
|
|
if( howmuch < 1 ) {
|
|
tr_peerIoSetEnabled( io, dir, FALSE );
|
|
return;
|
|
}
|
|
|
|
res = evbuffer_read( io->inbuf, fd, howmuch );
|
|
|
|
if( res > 0 )
|
|
{
|
|
tr_peerIoSetEnabled( io, dir, TRUE );
|
|
|
|
/* Invoke the user callback - must always be called last */
|
|
canReadWrapper( io );
|
|
}
|
|
else
|
|
{
|
|
short what = EVBUFFER_READ;
|
|
|
|
if( res == 0 ) /* EOF */
|
|
what |= EVBUFFER_EOF;
|
|
else if( res == -1 ) {
|
|
if( errno == EAGAIN || errno == EINTR ) {
|
|
tr_peerIoSetEnabled( io, dir, TRUE );
|
|
return;
|
|
}
|
|
what |= EVBUFFER_ERROR;
|
|
}
|
|
|
|
if( io->gotError != NULL )
|
|
io->gotError( io, what, io->userData );
|
|
}
|
|
}
|
|
|
|
static int
|
|
tr_evbuffer_write( tr_peerIo * io, int fd, size_t howmuch )
|
|
{
|
|
struct evbuffer * buffer = io->outbuf;
|
|
int n = MIN( EVBUFFER_LENGTH( buffer ), howmuch );
|
|
|
|
#ifdef WIN32
|
|
n = send(fd, buffer->buffer, n, 0 );
|
|
#else
|
|
n = write(fd, buffer->buffer, n );
|
|
#endif
|
|
dbgmsg( io, "wrote %d to peer (%s)", n, (n==-1?strerror(errno):"") );
|
|
|
|
if( n == -1 )
|
|
return -1;
|
|
if (n == 0)
|
|
return 0;
|
|
evbuffer_drain( buffer, n );
|
|
|
|
return n;
|
|
}
|
|
|
|
static void
|
|
event_write_cb( int fd, short event UNUSED, void * vio )
|
|
{
|
|
int res = 0;
|
|
short what = EVBUFFER_WRITE;
|
|
tr_peerIo * io = vio;
|
|
size_t howmuch;
|
|
const tr_direction dir = TR_UP;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
dbgmsg( io, "libevent says this peer is ready to write" );
|
|
|
|
/* Write as much as possible, since the socket is non-blocking, write() will
|
|
* return if it can't write any more data without blocking */
|
|
howmuch = tr_bandwidthClamp( io->bandwidth, dir, EVBUFFER_LENGTH( io->outbuf ) );
|
|
|
|
/* if we don't have any bandwidth left, stop writing */
|
|
if( howmuch < 1 ) {
|
|
tr_peerIoSetEnabled( io, dir, FALSE );
|
|
return;
|
|
}
|
|
|
|
res = tr_evbuffer_write( io, fd, howmuch );
|
|
if (res == -1) {
|
|
#ifndef WIN32
|
|
/*todo. evbuffer uses WriteFile when WIN32 is set. WIN32 system calls do not
|
|
* *set errno. thus this error checking is not portable*/
|
|
if (errno == EAGAIN || errno == EINTR || errno == EINPROGRESS)
|
|
goto reschedule;
|
|
/* error case */
|
|
what |= EVBUFFER_ERROR;
|
|
|
|
#else
|
|
goto reschedule;
|
|
#endif
|
|
|
|
} else if (res == 0) {
|
|
/* eof case */
|
|
what |= EVBUFFER_EOF;
|
|
}
|
|
if (res <= 0)
|
|
goto error;
|
|
|
|
if( EVBUFFER_LENGTH( io->outbuf ) )
|
|
tr_peerIoSetEnabled( io, dir, TRUE );
|
|
|
|
didWriteWrapper( io, res );
|
|
return;
|
|
|
|
reschedule:
|
|
if( EVBUFFER_LENGTH( io->outbuf ) )
|
|
tr_peerIoSetEnabled( io, dir, TRUE );
|
|
return;
|
|
|
|
error:
|
|
if( io->gotError != NULL )
|
|
io->gotError( io, what, io->userData );
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
static tr_peerIo*
|
|
tr_peerIoNew( tr_session * session,
|
|
const tr_address * addr,
|
|
tr_port port,
|
|
const uint8_t * torrentHash,
|
|
int isIncoming,
|
|
int socket )
|
|
{
|
|
tr_peerIo * io;
|
|
|
|
if( socket >= 0 )
|
|
tr_netSetTOS( socket, session->peerSocketTOS );
|
|
|
|
io = tr_new0( tr_peerIo, 1 );
|
|
io->magicNumber = MAGIC_NUMBER;
|
|
io->crypto = tr_cryptoNew( torrentHash, isIncoming );
|
|
io->session = session;
|
|
io->addr = *addr;
|
|
io->port = port;
|
|
io->socket = socket;
|
|
io->isIncoming = isIncoming != 0;
|
|
io->timeCreated = time( NULL );
|
|
io->inbuf = evbuffer_new( );
|
|
io->outbuf = evbuffer_new( );
|
|
|
|
event_set( &io->event_read, io->socket, EV_READ, event_read_cb, io );
|
|
event_set( &io->event_write, io->socket, EV_WRITE, event_write_cb, io );
|
|
|
|
__tr_list_init( &io->outbuf_datatypes );
|
|
|
|
tr_peerIoSetBandwidth( io, session->bandwidth );
|
|
|
|
return io;
|
|
}
|
|
|
|
tr_peerIo*
|
|
tr_peerIoNewIncoming( tr_session * session,
|
|
const tr_address * addr,
|
|
tr_port port,
|
|
int socket )
|
|
{
|
|
assert( session );
|
|
assert( tr_isAddress( addr ) );
|
|
assert( socket >= 0 );
|
|
|
|
return tr_peerIoNew( session, addr, port, NULL, 1, socket );
|
|
}
|
|
|
|
tr_peerIo*
|
|
tr_peerIoNewOutgoing( tr_session * session,
|
|
const tr_address * addr,
|
|
tr_port port,
|
|
const uint8_t * torrentHash )
|
|
{
|
|
int socket;
|
|
|
|
assert( session );
|
|
assert( tr_isAddress( addr ) );
|
|
assert( torrentHash );
|
|
|
|
socket = tr_netOpenTCP( session, addr, port );
|
|
|
|
return socket < 0
|
|
? NULL
|
|
: tr_peerIoNew( session, addr, port, torrentHash, 0, socket );
|
|
}
|
|
|
|
static void
|
|
trDatatypeFree( void * data )
|
|
{
|
|
struct tr_datatype * dt = __tr_list_entry( data, struct tr_datatype, head );
|
|
tr_free(dt);
|
|
}
|
|
|
|
static void
|
|
io_dtor( void * vio )
|
|
{
|
|
tr_peerIo * io = vio;
|
|
|
|
event_del( &io->event_read );
|
|
event_del( &io->event_write );
|
|
tr_peerIoSetBandwidth( io, NULL );
|
|
evbuffer_free( io->outbuf );
|
|
evbuffer_free( io->inbuf );
|
|
tr_netClose( io->socket );
|
|
tr_cryptoFree( io->crypto );
|
|
__tr_list_destroy( &io->outbuf_datatypes, trDatatypeFree );
|
|
|
|
io->magicNumber = 0xDEAD;
|
|
tr_free( io );
|
|
}
|
|
|
|
void
|
|
tr_peerIoFree( tr_peerIo * io )
|
|
{
|
|
if( io )
|
|
{
|
|
io->canRead = NULL;
|
|
io->didWrite = NULL;
|
|
io->gotError = NULL;
|
|
tr_runInEventThread( io->session, io_dtor, io );
|
|
}
|
|
}
|
|
|
|
tr_session*
|
|
tr_peerIoGetSession( tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( io->session );
|
|
|
|
return io->session;
|
|
}
|
|
|
|
const tr_address*
|
|
tr_peerIoGetAddress( const tr_peerIo * io,
|
|
tr_port * port )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
if( port )
|
|
*port = io->port;
|
|
|
|
return &io->addr;
|
|
}
|
|
|
|
const char*
|
|
tr_peerIoAddrStr( const tr_address * addr, tr_port port )
|
|
{
|
|
static char buf[512];
|
|
|
|
if( addr->type == TR_AF_INET )
|
|
tr_snprintf( buf, sizeof( buf ), "%s:%u", tr_ntop_non_ts( addr ), ntohs( port ) );
|
|
else
|
|
tr_snprintf( buf, sizeof( buf ), "[%s]:%u", tr_ntop_non_ts( addr ), ntohs( port ) );
|
|
return buf;
|
|
}
|
|
|
|
const char*
|
|
tr_peerIoGetAddrStr( const tr_peerIo * io )
|
|
{
|
|
return tr_peerIoAddrStr( &io->addr, io->port );
|
|
}
|
|
|
|
void
|
|
tr_peerIoSetIOFuncs( tr_peerIo * io,
|
|
tr_can_read_cb readcb,
|
|
tr_did_write_cb writecb,
|
|
tr_net_error_cb errcb,
|
|
void * userData )
|
|
{
|
|
io->canRead = readcb;
|
|
io->didWrite = writecb;
|
|
io->gotError = errcb;
|
|
io->userData = userData;
|
|
}
|
|
|
|
tr_bool
|
|
tr_peerIoIsIncoming( const tr_peerIo * c )
|
|
{
|
|
return c->isIncoming != 0;
|
|
}
|
|
|
|
int
|
|
tr_peerIoReconnect( tr_peerIo * io )
|
|
{
|
|
assert( !tr_peerIoIsIncoming( io ) );
|
|
|
|
if( io->socket >= 0 )
|
|
tr_netClose( io->socket );
|
|
|
|
io->socket = tr_netOpenTCP( io->session, &io->addr, io->port );
|
|
|
|
if( io->socket >= 0 )
|
|
{
|
|
tr_bandwidth * bandwidth = io->bandwidth;
|
|
tr_peerIoSetBandwidth( io, NULL );
|
|
tr_netSetTOS( io->socket, io->session->peerSocketTOS );
|
|
tr_peerIoSetBandwidth( io, bandwidth );
|
|
return 0;
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
void
|
|
tr_peerIoSetTorrentHash( tr_peerIo * io,
|
|
const uint8_t * hash )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
tr_cryptoSetTorrentHash( io->crypto, hash );
|
|
}
|
|
|
|
const uint8_t*
|
|
tr_peerIoGetTorrentHash( tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( io->crypto );
|
|
|
|
return tr_cryptoGetTorrentHash( io->crypto );
|
|
}
|
|
|
|
int
|
|
tr_peerIoHasTorrentHash( const tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( io->crypto );
|
|
|
|
return tr_cryptoHasTorrentHash( io->crypto );
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
void
|
|
tr_peerIoSetPeersId( tr_peerIo * io,
|
|
const uint8_t * peer_id )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
if( ( io->peerIdIsSet = peer_id != NULL ) )
|
|
memcpy( io->peerId, peer_id, 20 );
|
|
else
|
|
memset( io->peerId, 0, 20 );
|
|
}
|
|
|
|
const uint8_t*
|
|
tr_peerIoGetPeersId( const tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( io->peerIdIsSet );
|
|
|
|
return io->peerId;
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
void
|
|
tr_peerIoEnableFEXT( tr_peerIo * io,
|
|
tr_bool flag )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( _isBool( flag ) );
|
|
|
|
dbgmsg( io, "setting FEXT support flag to %d", (flag!=0) );
|
|
io->fastExtensionSupported = flag;
|
|
}
|
|
|
|
tr_bool
|
|
tr_peerIoSupportsFEXT( const tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
return io->fastExtensionSupported;
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
void
|
|
tr_peerIoEnableLTEP( tr_peerIo * io,
|
|
tr_bool flag )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( _isBool( flag ) );
|
|
|
|
dbgmsg( io, "setting LTEP support flag to %d", (flag!=0) );
|
|
io->extendedProtocolSupported = flag;
|
|
}
|
|
|
|
tr_bool
|
|
tr_peerIoSupportsLTEP( const tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
return io->extendedProtocolSupported;
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
static size_t
|
|
getDesiredOutputBufferSize( const tr_peerIo * io )
|
|
{
|
|
/* this is all kind of arbitrary, but what seems to work well is
|
|
* being large enough to hold the next 20 seconds' worth of input,
|
|
* or a few blocks, whichever is bigger.
|
|
* It's okay to tweak this as needed */
|
|
const double maxBlockSize = 16 * 1024; /* 16 KiB is from BT spec */
|
|
const double currentSpeed = tr_bandwidthGetPieceSpeed( io->bandwidth, TR_UP );
|
|
const double period = 20; /* arbitrary */
|
|
const double numBlocks = 5.5; /* the 5 is arbitrary; the .5 is to leave room for messages */
|
|
return MAX( maxBlockSize*numBlocks, currentSpeed*1024*period );
|
|
}
|
|
|
|
size_t
|
|
tr_peerIoGetWriteBufferSpace( const tr_peerIo * io )
|
|
{
|
|
const size_t desiredLen = getDesiredOutputBufferSize( io );
|
|
const size_t currentLen = EVBUFFER_LENGTH( io->outbuf );
|
|
size_t freeSpace = 0;
|
|
|
|
if( desiredLen > currentLen )
|
|
freeSpace = desiredLen - currentLen;
|
|
|
|
return freeSpace;
|
|
}
|
|
|
|
void
|
|
tr_peerIoSetBandwidth( tr_peerIo * io,
|
|
tr_bandwidth * bandwidth )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
if( io->bandwidth )
|
|
tr_bandwidthRemovePeer( io->bandwidth, io );
|
|
|
|
io->bandwidth = bandwidth;
|
|
|
|
if( io->bandwidth )
|
|
tr_bandwidthAddPeer( io->bandwidth, io );
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
tr_crypto*
|
|
tr_peerIoGetCrypto( tr_peerIo * c )
|
|
{
|
|
return c->crypto;
|
|
}
|
|
|
|
void
|
|
tr_peerIoSetEncryption( tr_peerIo * io,
|
|
int encryptionMode )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( encryptionMode == PEER_ENCRYPTION_NONE
|
|
|| encryptionMode == PEER_ENCRYPTION_RC4 );
|
|
|
|
io->encryptionMode = encryptionMode;
|
|
}
|
|
|
|
int
|
|
tr_peerIoIsEncrypted( const tr_peerIo * io )
|
|
{
|
|
return io != NULL && io->encryptionMode == PEER_ENCRYPTION_RC4;
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
void
|
|
tr_peerIoWrite( tr_peerIo * io,
|
|
const void * writeme,
|
|
size_t writemeLen,
|
|
int isPieceData )
|
|
{
|
|
struct tr_datatype * datatype;
|
|
|
|
assert( tr_amInEventThread( io->session ) );
|
|
dbgmsg( io, "adding %zu bytes into io->output", writemeLen );
|
|
|
|
datatype = tr_new( struct tr_datatype, 1 );
|
|
datatype->isPieceData = isPieceData != 0;
|
|
datatype->length = writemeLen;
|
|
|
|
__tr_list_init( &datatype->head );
|
|
__tr_list_append( &io->outbuf_datatypes, &datatype->head );
|
|
|
|
evbuffer_add( io->outbuf, writeme, writemeLen );
|
|
}
|
|
|
|
void
|
|
tr_peerIoWriteBuf( tr_peerIo * io,
|
|
struct evbuffer * buf,
|
|
int isPieceData )
|
|
{
|
|
const size_t n = EVBUFFER_LENGTH( buf );
|
|
|
|
tr_peerIoWrite( io, EVBUFFER_DATA( buf ), n, isPieceData );
|
|
evbuffer_drain( buf, n );
|
|
}
|
|
|
|
/**
|
|
***
|
|
**/
|
|
|
|
void
|
|
tr_peerIoWriteBytes( tr_peerIo * io,
|
|
struct evbuffer * outbuf,
|
|
const void * bytes,
|
|
size_t byteCount )
|
|
{
|
|
uint8_t * tmp;
|
|
|
|
switch( io->encryptionMode )
|
|
{
|
|
case PEER_ENCRYPTION_NONE:
|
|
evbuffer_add( outbuf, bytes, byteCount );
|
|
break;
|
|
|
|
case PEER_ENCRYPTION_RC4:
|
|
tmp = tr_new( uint8_t, byteCount );
|
|
tr_cryptoEncrypt( io->crypto, byteCount, bytes, tmp );
|
|
evbuffer_add( outbuf, tmp, byteCount );
|
|
tr_free( tmp );
|
|
break;
|
|
|
|
default:
|
|
assert( 0 );
|
|
}
|
|
}
|
|
|
|
void
|
|
tr_peerIoWriteUint8( tr_peerIo * io,
|
|
struct evbuffer * outbuf,
|
|
uint8_t writeme )
|
|
{
|
|
tr_peerIoWriteBytes( io, outbuf, &writeme, sizeof( uint8_t ) );
|
|
}
|
|
|
|
void
|
|
tr_peerIoWriteUint16( tr_peerIo * io,
|
|
struct evbuffer * outbuf,
|
|
uint16_t writeme )
|
|
{
|
|
uint16_t tmp = htons( writeme );
|
|
|
|
tr_peerIoWriteBytes( io, outbuf, &tmp, sizeof( uint16_t ) );
|
|
}
|
|
|
|
void
|
|
tr_peerIoWriteUint32( tr_peerIo * io,
|
|
struct evbuffer * outbuf,
|
|
uint32_t writeme )
|
|
{
|
|
uint32_t tmp = htonl( writeme );
|
|
|
|
tr_peerIoWriteBytes( io, outbuf, &tmp, sizeof( uint32_t ) );
|
|
}
|
|
|
|
/***
|
|
****
|
|
***/
|
|
|
|
void
|
|
tr_peerIoReadBytes( tr_peerIo * io,
|
|
struct evbuffer * inbuf,
|
|
void * bytes,
|
|
size_t byteCount )
|
|
{
|
|
assert( EVBUFFER_LENGTH( inbuf ) >= byteCount );
|
|
|
|
switch( io->encryptionMode )
|
|
{
|
|
case PEER_ENCRYPTION_NONE:
|
|
evbuffer_remove( inbuf, bytes, byteCount );
|
|
break;
|
|
|
|
case PEER_ENCRYPTION_RC4:
|
|
evbuffer_remove( inbuf, bytes, byteCount );
|
|
tr_cryptoDecrypt( io->crypto, byteCount, bytes, bytes );
|
|
break;
|
|
|
|
default:
|
|
assert( 0 );
|
|
}
|
|
}
|
|
|
|
void
|
|
tr_peerIoReadUint8( tr_peerIo * io,
|
|
struct evbuffer * inbuf,
|
|
uint8_t * setme )
|
|
{
|
|
tr_peerIoReadBytes( io, inbuf, setme, sizeof( uint8_t ) );
|
|
}
|
|
|
|
void
|
|
tr_peerIoReadUint16( tr_peerIo * io,
|
|
struct evbuffer * inbuf,
|
|
uint16_t * setme )
|
|
{
|
|
uint16_t tmp;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
tr_peerIoReadBytes( io, inbuf, &tmp, sizeof( uint16_t ) );
|
|
*setme = ntohs( tmp );
|
|
}
|
|
|
|
void
|
|
tr_peerIoReadUint32( tr_peerIo * io,
|
|
struct evbuffer * inbuf,
|
|
uint32_t * setme )
|
|
{
|
|
uint32_t tmp;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
tr_peerIoReadBytes( io, inbuf, &tmp, sizeof( uint32_t ) );
|
|
*setme = ntohl( tmp );
|
|
}
|
|
|
|
void
|
|
tr_peerIoDrain( tr_peerIo * io,
|
|
struct evbuffer * inbuf,
|
|
size_t byteCount )
|
|
{
|
|
uint8_t * tmp;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
tmp = tr_new( uint8_t, byteCount );
|
|
tr_peerIoReadBytes( io, inbuf, tmp, byteCount );
|
|
tr_free( tmp );
|
|
}
|
|
|
|
int
|
|
tr_peerIoGetAge( const tr_peerIo * io )
|
|
{
|
|
return time( NULL ) - io->timeCreated;
|
|
}
|
|
|
|
/***
|
|
****
|
|
***/
|
|
|
|
static int
|
|
tr_peerIoTryRead( tr_peerIo * io, size_t howmuch )
|
|
{
|
|
int res;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
howmuch = tr_bandwidthClamp( io->bandwidth, TR_DOWN, howmuch );
|
|
|
|
res = howmuch ? evbuffer_read( io->inbuf, io->socket, howmuch ) : 0;
|
|
|
|
dbgmsg( io, "read %d from peer (%s)", res, (res==-1?strerror(errno):"") );
|
|
|
|
if( EVBUFFER_LENGTH( io->inbuf ) )
|
|
canReadWrapper( io );
|
|
|
|
if( ( res <= 0 ) && ( io->gotError ) && ( errno != EAGAIN ) && ( errno != EINTR ) && ( errno != EINPROGRESS ) )
|
|
{
|
|
short what = EVBUFFER_READ | EVBUFFER_ERROR;
|
|
if( res == 0 )
|
|
what |= EVBUFFER_EOF;
|
|
io->gotError( io, what, io->userData );
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
static int
|
|
tr_peerIoTryWrite( tr_peerIo * io, size_t howmuch )
|
|
{
|
|
int n;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
howmuch = tr_bandwidthClamp( io->bandwidth, TR_UP, howmuch );
|
|
|
|
n = tr_evbuffer_write( io, io->socket, (int)howmuch );
|
|
|
|
if( n > 0 )
|
|
didWriteWrapper( io, n );
|
|
|
|
if( ( n < 0 ) && ( io->gotError ) && ( errno != EPIPE ) && ( errno != EAGAIN ) && ( errno != EINTR ) && ( errno != EINPROGRESS ) ) {
|
|
short what = EVBUFFER_WRITE | EVBUFFER_ERROR;
|
|
io->gotError( io, what, io->userData );
|
|
}
|
|
|
|
return n;
|
|
}
|
|
|
|
int
|
|
tr_peerIoFlush( tr_peerIo * io, tr_direction dir, size_t limit )
|
|
{
|
|
int ret;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( tr_isDirection( dir ) );
|
|
|
|
if( dir==TR_DOWN )
|
|
ret = tr_peerIoTryRead( io, limit );
|
|
else
|
|
ret = tr_peerIoTryWrite( io, limit );
|
|
|
|
return ret;
|
|
}
|
|
|
|
struct evbuffer *
|
|
tr_peerIoGetReadBuffer( tr_peerIo * io )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
return io->inbuf;
|
|
}
|
|
|
|
tr_bool
|
|
tr_peerIoHasBandwidthLeft( const tr_peerIo * io, tr_direction dir )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( tr_isDirection( dir ) );
|
|
|
|
return tr_bandwidthClamp( io->bandwidth, dir, 1024 ) > 0;
|
|
}
|
|
|
|
/***
|
|
****
|
|
****/
|
|
|
|
static void
|
|
event_enable( tr_peerIo * io, short event )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
if( event & EV_READ )
|
|
event_add( &io->event_read, NULL );
|
|
|
|
if( event & EV_WRITE )
|
|
event_add( &io->event_write, NULL );
|
|
}
|
|
|
|
static void
|
|
event_disable( struct tr_peerIo * io, short event )
|
|
{
|
|
assert( tr_isPeerIo( io ) );
|
|
|
|
if( event & EV_READ )
|
|
event_del( &io->event_read );
|
|
|
|
if( event & EV_WRITE )
|
|
event_del( &io->event_write );
|
|
}
|
|
|
|
|
|
void
|
|
tr_peerIoSetEnabled( tr_peerIo * io,
|
|
tr_direction dir,
|
|
tr_bool isEnabled )
|
|
{
|
|
short event;
|
|
|
|
assert( tr_isPeerIo( io ) );
|
|
assert( tr_isDirection( dir ) );
|
|
|
|
event = dir == TR_UP ? EV_WRITE : EV_READ;
|
|
|
|
if( isEnabled )
|
|
event_enable( io, event );
|
|
else
|
|
event_disable( io, event );
|
|
}
|