mirror of
https://github.com/transmission/transmission
synced 2025-03-11 14:43:42 +00:00
* chore: housekeeping * perf: short circuit peer has block check * refactor: track active requests in each respective peer * refactor: swap `ActiveRequests` with new request tracking method * refactor: use bitfield to store active requests per peer * perf: check active request numbers first * refactor: initialise candidate values in constructor * refactor: better naming * refactor: use `find_by_block()` more * refactor: store wishlist mediator in swarm object * test: make it compile * test: update endgame test * test: new test for choke event * test: remove redundant lines * test: new test for request event * test: new test for reject event * refactor: cache block have state in wishlist * test: fix `gotBlockResortsPiece` * fixup! refactor: track active requests in each respective peer * fixup! test: fix `gotBlockResortsPiece` * fix: count webseeds when calculating active requests * build: update xcode project * fix: add missing `candidates_dirty_` checks * chore: remove old `depends-on` comments * fixup! refactor: use bitfield to store active requests per peer * refactor: extract block peer event to separate function * perf: reorder conditions by overhead * perf: check for completed block instead of completed piece * chore: remove duplicated "unrequested piece" check * refactor: merge similar block size sanity check * refactor: use map to store number of requests in wishlist * refactor: add asserts * refactor: flush write buffer as soon as there is new data * refactor: more accurate function naming * fix: account for corrupt pieces in wishlist * fix: account for unaligned blocks in wishlist * Revert "fix: account for unaligned blocks in wishlist" This reverts commit c3fce93cbae49c11d62e26caccedf55c1987aa95. * fixup! refactor: use map to store number of requests in wishlist * fix: account for unaligned blocks in wishlist v2 * chore: add `[[nodiscard]]` * fixup! fix: account for unaligned blocks in wishlist v2 * fix: crash when handshake finishes in the middle of function
533 lines
15 KiB
C++
533 lines
15 KiB
C++
// This file Copyright © Mnemosyne LLC.
|
|
// It may be used under GPLv2 (SPDX: GPL-2.0-only), GPLv3 (SPDX: GPL-3.0-only),
|
|
// or any future license endorsed by Mnemosyne LLC.
|
|
// License text can be found in the licenses/ folder.
|
|
|
|
#include <algorithm> // std::adjacent_find, std::sort
|
|
#include <cstddef>
|
|
#include <functional>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include <small/map.hpp>
|
|
#include <small/vector.hpp>
|
|
|
|
#define LIBTRANSMISSION_PEER_MODULE
|
|
|
|
#include "libtransmission/transmission.h"
|
|
|
|
#include "libtransmission/bitfield.h"
|
|
#include "libtransmission/crypto-utils.h" // for tr_salt_shaker
|
|
#include "libtransmission/peer-mgr-wishlist.h"
|
|
|
|
// Asserts in this file are expensive, so hide them in #ifdef
|
|
#ifdef TR_WISHLIST_ASSERT
|
|
#include "libtransmission/tr-assert.h"
|
|
#endif
|
|
|
|
namespace
|
|
{
|
|
[[nodiscard]] std::vector<tr_block_span_t> make_spans(small::vector<tr_block_index_t> const& blocks)
|
|
{
|
|
if (std::empty(blocks))
|
|
{
|
|
return {};
|
|
}
|
|
|
|
auto spans = std::vector<tr_block_span_t>{};
|
|
spans.reserve(std::size(blocks));
|
|
for (auto span_begin = std::begin(blocks), end = std::end(blocks); span_begin != end;)
|
|
{
|
|
static auto constexpr NotAdjacent = [](tr_block_index_t const lhs, tr_block_index_t const rhs)
|
|
{
|
|
return lhs + 1U != rhs;
|
|
};
|
|
|
|
auto span_end = std::adjacent_find(span_begin, end, NotAdjacent);
|
|
if (span_end == end)
|
|
{
|
|
--span_end;
|
|
}
|
|
spans.push_back({ *span_begin, *span_end + 1 });
|
|
|
|
span_begin = std::next(span_end);
|
|
}
|
|
|
|
return spans;
|
|
}
|
|
} // namespace
|
|
|
|
class Wishlist::Impl
|
|
{
|
|
struct Candidate
|
|
{
|
|
Candidate(tr_piece_index_t piece_in, tr_piece_index_t salt_in, Mediator const* mediator)
|
|
: piece{ piece_in }
|
|
, block_span{ mediator->block_span(piece_in) }
|
|
, replication{ mediator->count_piece_replication(piece_in) }
|
|
, priority{ mediator->priority(piece_in) }
|
|
, salt{ salt_in }
|
|
, mediator_{ mediator }
|
|
{
|
|
n_reqs.reserve(block_span.end - block_span.begin);
|
|
for (auto [block, end] = block_span; block < end; ++block)
|
|
{
|
|
if (!mediator_->client_has_block(block))
|
|
{
|
|
n_reqs.try_emplace(block, mediator_->count_active_requests(block));
|
|
}
|
|
}
|
|
}
|
|
|
|
[[nodiscard]] int compare(Candidate const& that) const noexcept; // <=>
|
|
|
|
[[nodiscard]] auto operator<(Candidate const& that) const // less than
|
|
{
|
|
return compare(that) < 0;
|
|
}
|
|
|
|
tr_piece_index_t piece;
|
|
tr_block_span_t block_span;
|
|
|
|
small::map<tr_block_index_t, uint8_t> n_reqs;
|
|
|
|
// Caching the following 2 values are highly beneficial, because:
|
|
// - they are often used (mainly because resort_piece() is called
|
|
// every time we receive a block)
|
|
// - does not change as often compared to missing blocks
|
|
// - calculating their values involves sifting through bitfield(s),
|
|
// which is expensive.
|
|
size_t replication;
|
|
tr_priority_t priority;
|
|
|
|
tr_piece_index_t salt;
|
|
|
|
private:
|
|
Mediator const* mediator_;
|
|
};
|
|
|
|
using CandidateVec = std::vector<Candidate>;
|
|
|
|
public:
|
|
explicit Impl(Mediator& mediator_in);
|
|
|
|
[[nodiscard]] std::vector<tr_block_span_t> next(
|
|
size_t n_wanted_blocks,
|
|
std::function<bool(tr_piece_index_t)> const& peer_has_piece,
|
|
std::function<bool(tr_block_index_t)> const& has_active_request_to_peer);
|
|
|
|
private:
|
|
constexpr void set_candidates_dirty() noexcept
|
|
{
|
|
candidates_dirty_ = true;
|
|
}
|
|
|
|
// ---
|
|
|
|
TR_CONSTEXPR20 void dec_replication() noexcept
|
|
{
|
|
if (!candidates_dirty_)
|
|
{
|
|
std::for_each(
|
|
std::begin(candidates_),
|
|
std::end(candidates_),
|
|
[](Candidate& candidate) { --candidate.replication; });
|
|
}
|
|
}
|
|
|
|
TR_CONSTEXPR20 void dec_replication_bitfield(tr_bitfield const& bitfield)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (bitfield.has_none())
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (bitfield.has_all())
|
|
{
|
|
dec_replication();
|
|
return;
|
|
}
|
|
|
|
for (auto& candidate : candidates_)
|
|
{
|
|
if (bitfield.test(candidate.piece))
|
|
{
|
|
--candidate.replication;
|
|
}
|
|
}
|
|
|
|
std::sort(std::begin(candidates_), std::end(candidates_));
|
|
}
|
|
|
|
TR_CONSTEXPR20 void inc_replication() noexcept
|
|
{
|
|
if (!candidates_dirty_)
|
|
{
|
|
std::for_each(
|
|
std::begin(candidates_),
|
|
std::end(candidates_),
|
|
[](Candidate& candidate) { ++candidate.replication; });
|
|
}
|
|
}
|
|
|
|
void inc_replication_bitfield(tr_bitfield const& bitfield)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (bitfield.has_none())
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (bitfield.has_all())
|
|
{
|
|
inc_replication();
|
|
return;
|
|
}
|
|
|
|
for (auto& candidate : candidates_)
|
|
{
|
|
if (bitfield.test(candidate.piece))
|
|
{
|
|
++candidate.replication;
|
|
}
|
|
}
|
|
|
|
std::sort(std::begin(candidates_), std::end(candidates_));
|
|
}
|
|
|
|
TR_CONSTEXPR20 void inc_replication_piece(tr_piece_index_t piece)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (auto iter = find_by_piece(piece); iter != std::end(candidates_))
|
|
{
|
|
++iter->replication;
|
|
resort_piece(iter);
|
|
}
|
|
}
|
|
|
|
// ---
|
|
|
|
TR_CONSTEXPR20 void inc_active_request_span(tr_block_span_t block_span)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
for (auto block = block_span.begin; block < block_span.end;)
|
|
{
|
|
auto it_p = find_by_block(block);
|
|
if (it_p == std::end(candidates_))
|
|
{
|
|
set_candidates_dirty();
|
|
break;
|
|
}
|
|
|
|
auto& n_reqs = it_p->n_reqs;
|
|
|
|
auto it_b_begin = std::begin(n_reqs);
|
|
it_b_begin = it_b_begin->first >= block_span.begin ? it_b_begin : n_reqs.lower_bound(block_span.begin);
|
|
|
|
auto it_b_end = std::end(n_reqs);
|
|
it_b_end = std::prev(it_b_end)->first < block_span.end ? it_b_end : n_reqs.lower_bound(block_span.end);
|
|
|
|
for (auto it_b = it_b_begin; it_b != it_b_end; ++it_b)
|
|
{
|
|
++it_b->second;
|
|
}
|
|
|
|
block = it_p->block_span.end;
|
|
}
|
|
}
|
|
|
|
TR_CONSTEXPR20 void dec_active_request_block(tr_block_index_t block)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (auto it_p = find_by_block(block); it_p != std::end(candidates_))
|
|
{
|
|
auto& n_reqs = it_p->n_reqs;
|
|
if (auto it_b = n_reqs.find(block); it_b != std::end(n_reqs) && it_b->second > 0U)
|
|
{
|
|
--it_b->second;
|
|
}
|
|
}
|
|
}
|
|
|
|
TR_CONSTEXPR20 void dec_active_request_bitfield(tr_bitfield const& requests)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
for (auto& candidate : candidates_)
|
|
{
|
|
for (auto& [block, n_req] : candidate.n_reqs)
|
|
{
|
|
if (n_req > 0U && requests.test(block))
|
|
{
|
|
--n_req;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// ---
|
|
|
|
TR_CONSTEXPR20 void client_got_block(tr_block_index_t block)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (auto iter = find_by_block(block); iter != std::end(candidates_))
|
|
{
|
|
iter->n_reqs.erase(block);
|
|
resort_piece(iter);
|
|
}
|
|
}
|
|
|
|
// ---
|
|
|
|
TR_CONSTEXPR20 void peer_disconnect(tr_bitfield const& have, tr_bitfield const& requests)
|
|
{
|
|
dec_replication_bitfield(have);
|
|
dec_active_request_bitfield(requests);
|
|
}
|
|
|
|
// ---
|
|
|
|
[[nodiscard]] TR_CONSTEXPR20 CandidateVec::iterator find_by_piece(tr_piece_index_t const piece)
|
|
{
|
|
return std::find_if(
|
|
std::begin(candidates_),
|
|
std::end(candidates_),
|
|
[piece](auto const& c) { return c.piece == piece; });
|
|
}
|
|
|
|
[[nodiscard]] TR_CONSTEXPR20 CandidateVec::iterator find_by_block(tr_block_index_t const block)
|
|
{
|
|
return std::find_if(
|
|
std::begin(candidates_),
|
|
std::end(candidates_),
|
|
[block](auto const& c) { return c.block_span.begin <= block && block < c.block_span.end; });
|
|
}
|
|
|
|
void maybe_rebuild_candidate_list()
|
|
{
|
|
if (!candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
candidates_dirty_ = false;
|
|
candidates_.clear();
|
|
|
|
auto salter = tr_salt_shaker<tr_piece_index_t>{};
|
|
auto const is_sequential = mediator_.is_sequential_download();
|
|
auto const n_pieces = mediator_.piece_count();
|
|
candidates_.reserve(n_pieces);
|
|
for (tr_piece_index_t piece = 0U; piece < n_pieces; ++piece)
|
|
{
|
|
if (mediator_.client_has_piece(piece) || !mediator_.client_wants_piece(piece))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
auto const salt = is_sequential ? piece : salter();
|
|
candidates_.emplace_back(piece, salt, &mediator_);
|
|
}
|
|
std::sort(std::begin(candidates_), std::end(candidates_));
|
|
}
|
|
|
|
TR_CONSTEXPR20 void remove_piece(tr_piece_index_t const piece)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
if (auto iter = find_by_piece(piece); iter != std::end(candidates_))
|
|
{
|
|
candidates_.erase(iter);
|
|
}
|
|
}
|
|
|
|
TR_CONSTEXPR20 void resort_piece(CandidateVec::iterator const pos_old)
|
|
{
|
|
if (candidates_dirty_)
|
|
{
|
|
return;
|
|
}
|
|
|
|
auto const pos_begin = std::begin(candidates_);
|
|
|
|
// Candidate needs to be moved towards the front of the list
|
|
if (auto const pos_next = std::next(pos_old); pos_old > pos_begin && *pos_old < *std::prev(pos_old))
|
|
{
|
|
auto const pos_new = std::lower_bound(pos_begin, pos_old, *pos_old);
|
|
std::rotate(pos_new, pos_old, pos_next);
|
|
}
|
|
// Candidate needs to be moved towards the end of the list
|
|
else if (auto const pos_end = std::end(candidates_); pos_next < pos_end && *pos_next < *pos_old)
|
|
{
|
|
auto const pos_new = std::lower_bound(pos_next, pos_end, *pos_old);
|
|
std::rotate(pos_old, pos_next, pos_new);
|
|
}
|
|
}
|
|
|
|
CandidateVec candidates_;
|
|
bool candidates_dirty_ = true;
|
|
bool is_endgame_ = false;
|
|
|
|
std::array<libtransmission::ObserverTag, 13U> const tags_;
|
|
|
|
Mediator& mediator_;
|
|
};
|
|
|
|
Wishlist::Impl::Impl(Mediator& mediator_in)
|
|
: tags_{ {
|
|
mediator_in.observe_peer_disconnect([this](tr_torrent*, tr_bitfield const& b, tr_bitfield const& ar)
|
|
{ peer_disconnect(b, ar); }),
|
|
mediator_in.observe_got_bad_piece([this](tr_torrent*, tr_piece_index_t) { set_candidates_dirty(); }),
|
|
mediator_in.observe_got_bitfield([this](tr_torrent*, tr_bitfield const& b) { inc_replication_bitfield(b); }),
|
|
mediator_in.observe_got_block([this](tr_torrent*, tr_block_index_t b) { client_got_block(b); }),
|
|
mediator_in.observe_got_choke([this](tr_torrent*, tr_bitfield const& b) { dec_active_request_bitfield(b); }),
|
|
mediator_in.observe_got_have([this](tr_torrent*, tr_piece_index_t p) { inc_replication_piece(p); }),
|
|
mediator_in.observe_got_have_all([this](tr_torrent*) { inc_replication(); }),
|
|
mediator_in.observe_got_reject([this](tr_torrent*, tr_peer*, tr_block_index_t b) { dec_active_request_block(b); }),
|
|
mediator_in.observe_piece_completed([this](tr_torrent*, tr_piece_index_t p) { remove_piece(p); }),
|
|
mediator_in.observe_priority_changed([this](tr_torrent*, tr_file_index_t const*, tr_file_index_t, tr_priority_t)
|
|
{ set_candidates_dirty(); }),
|
|
mediator_in.observe_sent_cancel([this](tr_torrent*, tr_peer*, tr_block_index_t b) { dec_active_request_block(b); }),
|
|
mediator_in.observe_sent_request([this](tr_torrent*, tr_peer*, tr_block_span_t bs) { inc_active_request_span(bs); }),
|
|
mediator_in.observe_sequential_download_changed([this](tr_torrent*, bool) { set_candidates_dirty(); }),
|
|
} }
|
|
, mediator_{ mediator_in }
|
|
{
|
|
}
|
|
|
|
std::vector<tr_block_span_t> Wishlist::Impl::next(
|
|
size_t n_wanted_blocks,
|
|
std::function<bool(tr_piece_index_t)> const& peer_has_piece,
|
|
std::function<bool(tr_block_index_t)> const& has_active_request_to_peer)
|
|
{
|
|
if (n_wanted_blocks == 0U)
|
|
{
|
|
return {};
|
|
}
|
|
|
|
maybe_rebuild_candidate_list();
|
|
|
|
auto const max_peers = is_endgame_ ? EndgameMaxPeers : NormalMaxPeers;
|
|
auto blocks = small::vector<tr_block_index_t>{};
|
|
blocks.reserve(n_wanted_blocks);
|
|
for (auto const& candidate : candidates_)
|
|
{
|
|
// do we have enough?
|
|
if (std::size(blocks) >= n_wanted_blocks)
|
|
{
|
|
break;
|
|
}
|
|
|
|
// if the peer doesn't have this piece that we want...
|
|
if (candidate.replication == 0 || !peer_has_piece(candidate.piece))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
// walk the blocks in this piece that we don't have
|
|
for (auto const& [block, n_req] : candidate.n_reqs)
|
|
{
|
|
if (std::size(blocks) >= n_wanted_blocks)
|
|
{
|
|
break;
|
|
}
|
|
|
|
#ifdef TR_WISHLIST_ASSERT
|
|
auto const n_req_truth = mediator_.count_active_requests(block);
|
|
TR_ASSERT_MSG(
|
|
n_req == n_req_truth,
|
|
fmt::format("piece = {}, block = {}, n_req = {}, truth = {}", candidate.piece, block, n_req, n_req_truth));
|
|
#endif
|
|
|
|
// don't request from too many peers
|
|
if (n_req >= max_peers)
|
|
{
|
|
continue;
|
|
}
|
|
|
|
// don't request block from peers which we already requested from
|
|
if (has_active_request_to_peer(block))
|
|
{
|
|
continue;
|
|
}
|
|
|
|
blocks.emplace_back(block);
|
|
}
|
|
}
|
|
|
|
is_endgame_ = std::size(blocks) < n_wanted_blocks;
|
|
|
|
// Ensure the list of blocks are sorted
|
|
// The list needs to be unique as well, but that should come naturally
|
|
std::sort(std::begin(blocks), std::end(blocks));
|
|
return make_spans(blocks);
|
|
}
|
|
|
|
int Wishlist::Impl::Candidate::compare(Wishlist::Impl::Candidate const& that) const noexcept
|
|
{
|
|
// prefer pieces closer to completion
|
|
if (auto const val = tr_compare_3way(std::size(n_reqs), std::size(that.n_reqs)); val != 0)
|
|
{
|
|
return val;
|
|
}
|
|
|
|
// prefer higher priority
|
|
if (auto const val = tr_compare_3way(priority, that.priority); val != 0)
|
|
{
|
|
return -val;
|
|
}
|
|
|
|
// prefer rarer pieces
|
|
if (auto const val = tr_compare_3way(replication, that.replication); val != 0)
|
|
{
|
|
return val;
|
|
}
|
|
|
|
return tr_compare_3way(salt, that.salt);
|
|
}
|
|
|
|
// ---
|
|
|
|
Wishlist::Wishlist(Mediator& mediator_in)
|
|
: impl_{ std::make_unique<Impl>(mediator_in) }
|
|
{
|
|
}
|
|
|
|
Wishlist::~Wishlist() = default;
|
|
|
|
std::vector<tr_block_span_t> Wishlist::next(
|
|
size_t n_wanted_blocks,
|
|
std::function<bool(tr_piece_index_t)> const& peer_has_piece,
|
|
std::function<bool(tr_block_index_t)> const& has_active_pending_to_peer)
|
|
{
|
|
return impl_->next(n_wanted_blocks, peer_has_piece, has_active_pending_to_peer);
|
|
}
|