mirror of
https://github.com/oxen-io/lokinet.git
synced 2024-10-31 09:20:21 +00:00
Merge remote-tracking branch 'origin/master'
This commit is contained in:
commit
702273427e
@ -24,7 +24,7 @@ option(DEBIAN "build for debian" )
|
||||
option(TESTNET "testnet build" )
|
||||
option(WITH_SHARED "build shared library")
|
||||
option(WITH_COVERAGE "generate coverage data")
|
||||
option(WITH_SHELLHOOKS "enable shell hooks on compile time (dangerous)" OFF)
|
||||
option(USE_SHELLHOOKS "enable shell hooks on compile time (dangerous)" OFF)
|
||||
option(WARNINGS_AS_ERRORS "treat all warnings as errors. turn off for development, on for release" OFF)
|
||||
|
||||
include(cmake/target_link_libraries_system.cmake)
|
||||
|
@ -336,9 +336,10 @@ namespace llarp
|
||||
uint64_t txid = ++ids;
|
||||
TXOwner peer(askpeer, txid);
|
||||
TXOwner whoasked(OurKey(), txid);
|
||||
RouterID K;
|
||||
K.Randomize();
|
||||
pendingExploreLookups().NewTX(
|
||||
peer, whoasked, askpeer.as_array(),
|
||||
new ExploreNetworkJob(askpeer.as_array(), this));
|
||||
peer, whoasked, K, new ExploreNetworkJob(askpeer.as_array(), this));
|
||||
}
|
||||
|
||||
void
|
||||
|
@ -9,7 +9,7 @@ namespace llarp
|
||||
{
|
||||
struct XorMetric
|
||||
{
|
||||
const Key_t& us;
|
||||
const Key_t us;
|
||||
|
||||
XorMetric(const Key_t& ourKey) : us(ourKey)
|
||||
{
|
||||
|
@ -13,6 +13,7 @@
|
||||
#include <routing/transfer_traffic_message.hpp>
|
||||
#include <util/buffer.hpp>
|
||||
#include <util/endian.hpp>
|
||||
#include <util/logic.hpp>
|
||||
|
||||
#include <deque>
|
||||
|
||||
@ -145,7 +146,8 @@ namespace llarp
|
||||
if((currentStatus & LR_StatusRecord::SUCCESS) == 1)
|
||||
{
|
||||
llarp::LogDebug("LR_Status message processed, path build successful");
|
||||
HandlePathConfirmMessage(r);
|
||||
auto self = shared_from_this();
|
||||
r->logic()->queue_func([=]() { self->HandlePathConfirmMessage(r); });
|
||||
}
|
||||
else
|
||||
{
|
||||
|
@ -15,12 +15,14 @@
|
||||
namespace llarp
|
||||
{
|
||||
struct AsyncPathKeyExchangeContext
|
||||
: std::enable_shared_from_this< AsyncPathKeyExchangeContext >
|
||||
{
|
||||
using Path_t = path::Path_ptr;
|
||||
using PathSet_t = path::PathSet_ptr;
|
||||
PathSet_t pathset = nullptr;
|
||||
Path_t path = nullptr;
|
||||
using Handler = std::function< void(const AsyncPathKeyExchangeContext&) >;
|
||||
using Handler =
|
||||
std::function< void(std::shared_ptr< AsyncPathKeyExchangeContext >) >;
|
||||
|
||||
Handler result;
|
||||
size_t idx = 0;
|
||||
@ -96,13 +98,13 @@ namespace llarp
|
||||
{
|
||||
// farthest hop
|
||||
// TODO: encrypt junk frames because our public keys are not eligator
|
||||
logic->queue_func(std::bind(result, *this));
|
||||
logic->queue_func(std::bind(result, shared_from_this()));
|
||||
}
|
||||
else
|
||||
{
|
||||
// next hop
|
||||
worker->addJob(
|
||||
std::bind(&AsyncPathKeyExchangeContext::GenerateNextKey, *this));
|
||||
worker->addJob(std::bind(&AsyncPathKeyExchangeContext::GenerateNextKey,
|
||||
shared_from_this()));
|
||||
}
|
||||
}
|
||||
|
||||
@ -120,28 +122,28 @@ namespace llarp
|
||||
{
|
||||
LRCM.frames[i].Randomize();
|
||||
}
|
||||
pool->addJob(
|
||||
std::bind(&AsyncPathKeyExchangeContext::GenerateNextKey, *this));
|
||||
pool->addJob(std::bind(&AsyncPathKeyExchangeContext::GenerateNextKey,
|
||||
shared_from_this()));
|
||||
}
|
||||
};
|
||||
|
||||
static void
|
||||
PathBuilderKeysGenerated(const AsyncPathKeyExchangeContext& ctx)
|
||||
PathBuilderKeysGenerated(std::shared_ptr< AsyncPathKeyExchangeContext > ctx)
|
||||
{
|
||||
if(!ctx.pathset->IsStopped())
|
||||
if(!ctx->pathset->IsStopped())
|
||||
{
|
||||
const RouterID remote = ctx.path->Upstream();
|
||||
const ILinkMessage* msg = &ctx.LRCM;
|
||||
if(ctx.router->SendToOrQueue(remote, msg))
|
||||
const RouterID remote = ctx->path->Upstream();
|
||||
const ILinkMessage* msg = &ctx->LRCM;
|
||||
if(ctx->router->SendToOrQueue(remote, msg))
|
||||
{
|
||||
// persist session with router until this path is done
|
||||
ctx.router->PersistSessionUntil(remote, ctx.path->ExpireTime());
|
||||
ctx->router->PersistSessionUntil(remote, ctx->path->ExpireTime());
|
||||
// add own path
|
||||
ctx.router->pathContext().AddOwnPath(ctx.pathset, ctx.path);
|
||||
ctx.pathset->PathBuildStarted(ctx.path);
|
||||
ctx->router->pathContext().AddOwnPath(ctx->pathset, ctx->path);
|
||||
ctx->pathset->PathBuildStarted(ctx->path);
|
||||
}
|
||||
else
|
||||
LogError(ctx.pathset->Name(), " failed to send LRCM to ", remote);
|
||||
LogError(ctx->pathset->Name(), " failed to send LRCM to ", remote);
|
||||
}
|
||||
}
|
||||
|
||||
@ -419,15 +421,15 @@ namespace llarp
|
||||
return;
|
||||
lastBuild = Now();
|
||||
// async generate keys
|
||||
AsyncPathKeyExchangeContext ctx;
|
||||
ctx.router = router;
|
||||
ctx.pathset = GetSelf();
|
||||
auto path = std::make_shared< path::Path >(hops, this, roles);
|
||||
auto ctx = std::make_shared< AsyncPathKeyExchangeContext >();
|
||||
ctx->router = router;
|
||||
ctx->pathset = GetSelf();
|
||||
auto path = std::make_shared< path::Path >(hops, this, roles);
|
||||
LogInfo(Name(), " build ", path->HopsString());
|
||||
path->SetBuildResultHook(
|
||||
[this](Path_ptr p) { this->HandlePathBuilt(p); });
|
||||
ctx.AsyncGenerateKeys(path, router->logic(), router->threadpool(),
|
||||
&PathBuilderKeysGenerated);
|
||||
ctx->AsyncGenerateKeys(path, router->logic(), router->threadpool(),
|
||||
&PathBuilderKeysGenerated);
|
||||
}
|
||||
|
||||
void
|
||||
|
@ -893,20 +893,21 @@ namespace llarp
|
||||
Endpoint::OnLookup(const Address& addr, const IntroSet* introset,
|
||||
const RouterID& endpoint)
|
||||
{
|
||||
auto now = Now();
|
||||
auto& fails = m_state->m_ServiceLookupFails;
|
||||
auto& lookups = m_state->m_PendingServiceLookups;
|
||||
const auto now = router->Now();
|
||||
auto& fails = m_state->m_ServiceLookupFails;
|
||||
auto& lookups = m_state->m_PendingServiceLookups;
|
||||
if(introset == nullptr || introset->IsExpired(now))
|
||||
{
|
||||
LogError(Name(), " failed to lookup ", addr.ToString(), " from ",
|
||||
endpoint);
|
||||
fails[endpoint] = fails[endpoint] + 1;
|
||||
// inform one
|
||||
auto itr = lookups.find(addr);
|
||||
if(itr != lookups.end())
|
||||
// inform all
|
||||
auto range = lookups.equal_range(addr);
|
||||
auto itr = range.first;
|
||||
if(itr != range.second)
|
||||
{
|
||||
itr->second(addr, nullptr);
|
||||
lookups.erase(itr);
|
||||
itr = lookups.erase(itr);
|
||||
}
|
||||
return false;
|
||||
}
|
||||
@ -1169,12 +1170,12 @@ namespace llarp
|
||||
if(c)
|
||||
{
|
||||
c->UpdateIntroSet(true);
|
||||
for(auto& pending : traffic[r])
|
||||
for(auto& pending : m_state->m_PendingTraffic[r])
|
||||
{
|
||||
c->AsyncEncryptAndSendTo(pending.Buffer(), pending.protocol);
|
||||
}
|
||||
}
|
||||
traffic.erase(r);
|
||||
m_state->m_PendingTraffic.erase(r);
|
||||
},
|
||||
5000, true);
|
||||
}
|
||||
|
@ -52,24 +52,25 @@ namespace llarp
|
||||
do
|
||||
{
|
||||
auto& msg = sendq.front();
|
||||
while(msg.vecs.size() && sz >= msg.vecs.front().iov_len)
|
||||
while(msg.vecs.size() > 0 && sz >= msg.vecs.front().iov_len)
|
||||
{
|
||||
sz -= msg.vecs.front().iov_len;
|
||||
msg.vecs.pop_front();
|
||||
msg.fragments.pop_front();
|
||||
}
|
||||
if(msg.vecs.size() == 0)
|
||||
{
|
||||
msg.Delivered();
|
||||
sendq.pop_front();
|
||||
}
|
||||
else
|
||||
else if(sz)
|
||||
{
|
||||
auto& front = msg.vecs.front();
|
||||
front.iov_len -= sz;
|
||||
front.iov_base = ((byte_t*)front.iov_base) + sz;
|
||||
return;
|
||||
}
|
||||
else
|
||||
return;
|
||||
} while(sendq.size());
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user