2023-10-10 15:18:52 +00:00
|
|
|
package sweepbatcher
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
2024-05-24 08:52:26 +00:00
|
|
|
"errors"
|
2024-06-17 02:45:07 +00:00
|
|
|
"fmt"
|
2024-05-24 08:52:54 +00:00
|
|
|
"sync"
|
2023-10-10 15:18:52 +00:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2024-06-22 02:19:51 +00:00
|
|
|
"github.com/btcsuite/btcd/btcec/v2"
|
2023-10-10 15:18:52 +00:00
|
|
|
"github.com/btcsuite/btcd/btcutil"
|
|
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
|
|
"github.com/btcsuite/btcd/wire"
|
|
|
|
"github.com/lightninglabs/loop/loopdb"
|
|
|
|
"github.com/lightninglabs/loop/test"
|
2024-05-14 14:36:47 +00:00
|
|
|
"github.com/lightninglabs/loop/utils"
|
2023-10-10 15:18:52 +00:00
|
|
|
"github.com/lightningnetwork/lnd/chainntnfs"
|
2024-06-25 02:18:16 +00:00
|
|
|
"github.com/lightningnetwork/lnd/input"
|
2024-06-22 02:19:51 +00:00
|
|
|
"github.com/lightningnetwork/lnd/keychain"
|
2023-10-10 15:18:52 +00:00
|
|
|
"github.com/lightningnetwork/lnd/lntypes"
|
2024-06-22 02:20:52 +00:00
|
|
|
"github.com/lightningnetwork/lnd/lnwallet/chainfee"
|
2023-10-10 15:18:52 +00:00
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
|
|
|
swapInvoice = "lntb1230n1pjjszzgpp5j76f03wrkya4sm4gxv6az5nmz5aqsvmn4" +
|
|
|
|
"tpguu2sdvdyygedqjgqdq9xyerxcqzzsxqr23ssp5rwzmwtfjmsgranfk8sr" +
|
|
|
|
"4p4gcgmvyd42uug8pxteg2mkk23ndvkqs9qyyssq44ruk3ex59cmv4dm6k4v" +
|
|
|
|
"0kc6c0gcqjs0gkljfyd6c6uatqa2f67xlx3pcg5tnvcae5p3jju8ra77e87d" +
|
|
|
|
"vhhs0jrx53wnc0fq9rkrhmqqelyx7l"
|
|
|
|
|
|
|
|
eventuallyCheckFrequency = 100 * time.Millisecond
|
|
|
|
|
|
|
|
ntfnBufferSize = 1024
|
|
|
|
)
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// destAddr is a dummy p2wkh address to use as the destination address for
|
|
|
|
// the swaps.
|
|
|
|
var destAddr = func() btcutil.Address {
|
|
|
|
p2wkhAddr := "bcrt1qq68r6ff4k4pjx39efs44gcyccf7unqnu5qtjjz"
|
|
|
|
addr, err := btcutil.DecodeAddress(p2wkhAddr, nil)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
return addr
|
|
|
|
}()
|
|
|
|
|
2024-06-22 02:19:51 +00:00
|
|
|
var senderKey, receiverKey [33]byte
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
// Generate keys.
|
|
|
|
_, senderPubKey := test.CreateKey(1)
|
|
|
|
copy(senderKey[:], senderPubKey.SerializeCompressed())
|
|
|
|
_, receiverPubKey := test.CreateKey(2)
|
|
|
|
copy(receiverKey[:], receiverPubKey.SerializeCompressed())
|
|
|
|
}
|
|
|
|
|
|
|
|
func testVerifySchnorrSig(pubKey *btcec.PublicKey, hash, sig []byte) error {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
func testMuSig2SignSweep(ctx context.Context,
|
|
|
|
protocolVersion loopdb.ProtocolVersion, swapHash lntypes.Hash,
|
|
|
|
paymentAddr [32]byte, nonce []byte, sweepTxPsbt []byte,
|
|
|
|
prevoutMap map[wire.OutPoint]*wire.TxOut) (
|
|
|
|
[]byte, []byte, error) {
|
|
|
|
|
|
|
|
return nil, nil, nil
|
|
|
|
}
|
|
|
|
|
2024-06-25 02:18:16 +00:00
|
|
|
var customSignature = func() []byte {
|
|
|
|
sig := [64]byte{10, 20, 30}
|
|
|
|
return sig[:]
|
|
|
|
}()
|
|
|
|
|
|
|
|
func testSignMuSig2func(ctx context.Context, muSig2Version input.MuSig2Version,
|
|
|
|
swapHash lntypes.Hash, rootHash chainhash.Hash,
|
|
|
|
sigHash [32]byte) ([]byte, error) {
|
|
|
|
|
|
|
|
return customSignature, nil
|
|
|
|
}
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
var dummyNotifier = SpendNotifier{
|
2024-02-02 21:37:54 +00:00
|
|
|
SpendChan: make(chan *SpendDetail, ntfnBufferSize),
|
2023-10-10 15:18:52 +00:00
|
|
|
SpendErrChan: make(chan error, ntfnBufferSize),
|
|
|
|
QuitChan: make(chan bool, ntfnBufferSize),
|
|
|
|
}
|
|
|
|
|
2024-05-24 08:52:26 +00:00
|
|
|
func checkBatcherError(t *testing.T, err error) {
|
|
|
|
if !errors.Is(err, context.Canceled) &&
|
|
|
|
!errors.Is(err, ErrBatcherShuttingDown) &&
|
|
|
|
!errors.Is(err, ErrBatchShuttingDown) {
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// getOnlyBatch makes sure the batcher has exactly one batch and returns it.
|
|
|
|
func getOnlyBatch(batcher *Batcher) *batch {
|
|
|
|
if len(batcher.batches) != 1 {
|
|
|
|
panic(fmt.Sprintf("getOnlyBatch called on a batcher having "+
|
|
|
|
"%d batches", len(batcher.batches)))
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, batch := range batcher.batches {
|
|
|
|
return batch
|
|
|
|
}
|
|
|
|
|
|
|
|
panic("unreachable")
|
|
|
|
}
|
|
|
|
|
|
|
|
// testSweepBatcherBatchCreation tests that sweep requests enter the expected
|
2023-10-10 15:18:52 +00:00
|
|
|
// batch based on their timeout distance.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testSweepBatcherBatchCreation(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2023-10-10 15:18:52 +00:00
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
2024-05-24 08:52:26 +00:00
|
|
|
checkBatcherError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap1 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
err = store.CreateLoopOut(ctx, sweepReq1.SwapHash, swap1)
|
2023-10-10 15:18:52 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Insert the same swap twice, this should be a noop.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up a
|
|
|
|
// batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Create a second sweep request that has a timeout distance less than
|
|
|
|
// our configured threshold.
|
|
|
|
sweepReq2 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{2, 2, 2},
|
|
|
|
Value: 222,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{2, 2},
|
|
|
|
Index: 2,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap2 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance - 1,
|
|
|
|
AmountRequested: 222,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{2},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq2.SwapHash, swap2)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq2))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should not create a second batch as timeout distance is small
|
|
|
|
// enough.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Create a third sweep request that has more timeout distance than
|
|
|
|
// the default.
|
|
|
|
sweepReq3 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{3, 3, 3},
|
|
|
|
Value: 333,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{3, 3},
|
|
|
|
Index: 3,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap3 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance + 1,
|
|
|
|
AmountRequested: 333,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{3},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq3.SwapHash, swap3)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq3))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a second batch as timeout distance is greater
|
|
|
|
// than the threshold
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 2
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since the second batch got created we check that it registered its
|
|
|
|
// primary sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
require.Eventually(t, func() bool {
|
2024-05-29 01:35:08 +00:00
|
|
|
// Verify that each batch has the correct number of sweeps
|
|
|
|
// in it.
|
2023-10-10 15:18:52 +00:00
|
|
|
for _, batch := range batcher.batches {
|
|
|
|
switch batch.primarySweepID {
|
|
|
|
case sweepReq1.SwapHash:
|
|
|
|
if len(batch.sweeps) != 2 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
case sweepReq3.SwapHash:
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Check that all sweeps were stored.
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq1.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq2.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq3.SwapHash))
|
|
|
|
}
|
|
|
|
|
2024-06-22 02:19:51 +00:00
|
|
|
// testFeeBumping tests that sweep is RBFed with slightly higher fee rate after
|
|
|
|
// each block unless WithNoBumping is passed.
|
|
|
|
func testFeeBumping(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore, noFeeBumping bool) {
|
|
|
|
|
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Disable fee bumping, if requested.
|
|
|
|
var opts []BatcherOption
|
|
|
|
if noFeeBumping {
|
|
|
|
opts = append(opts, WithNoBumping())
|
|
|
|
}
|
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
|
|
|
testMuSig2SignSweep, testVerifySchnorrSig, lnd.ChainParams,
|
|
|
|
batcherStore, sweepStore, opts...)
|
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
|
|
|
checkBatcherError(t, err)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 1_000_000,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &SpendNotifier{
|
|
|
|
SpendChan: make(chan *SpendDetail, ntfnBufferSize),
|
|
|
|
SpendErrChan: make(chan error, ntfnBufferSize),
|
|
|
|
QuitChan: make(chan bool, ntfnBufferSize),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
swap1 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 1_000_000,
|
|
|
|
ProtocolVersion: loopdb.ProtocolVersionMuSig2,
|
|
|
|
HtlcKeys: loopdb.HtlcKeys{
|
|
|
|
SenderScriptKey: senderKey,
|
|
|
|
ReceiverScriptKey: receiverKey,
|
|
|
|
SenderInternalPubKey: senderKey,
|
|
|
|
ReceiverInternalPubKey: receiverKey,
|
|
|
|
ClientScriptKeyLocator: keychain.KeyLocator{
|
|
|
|
Family: 1,
|
|
|
|
Index: 2,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
DestAddr: destAddr,
|
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 111,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq1.SwapHash, swap1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Wait for tx to be published.
|
|
|
|
tx1 := <-lnd.TxPublishChannel
|
|
|
|
out1 := tx1.TxOut[0].Value
|
|
|
|
|
|
|
|
// Tick tock next block.
|
|
|
|
err = lnd.NotifyHeight(601)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Wait for another sweep tx to be published.
|
|
|
|
tx2 := <-lnd.TxPublishChannel
|
|
|
|
out2 := tx2.TxOut[0].Value
|
|
|
|
|
|
|
|
if noFeeBumping {
|
|
|
|
// Expect output to stay the same.
|
|
|
|
require.Equal(t, out1, out2, "expected out to stay the same")
|
|
|
|
} else {
|
|
|
|
// Expect output to drop.
|
|
|
|
require.Greater(t, out1, out2, "expected out to drop")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testSweepBatcherSimpleLifecycle tests the simple lifecycle of the batches
|
2023-10-10 15:18:52 +00:00
|
|
|
// that are created and run by the batcher.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testSweepBatcherSimpleLifecycle(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2023-10-10 15:18:52 +00:00
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
2024-05-24 08:52:26 +00:00
|
|
|
checkBatcherError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap1 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 111,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
err = store.CreateLoopOut(ctx, sweepReq1.SwapHash, swap1)
|
2023-10-10 15:18:52 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Eventually request will be consumed and a new batch will spin up.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// When batch is successfully created it will execute it's first step,
|
|
|
|
// which leads to a spend monitor of the primary sweep.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Find the batch and assign it to a local variable for easier access.
|
|
|
|
batch := &batch{}
|
|
|
|
for _, btch := range batcher.batches {
|
|
|
|
if btch.primarySweepID == sweepReq1.SwapHash {
|
|
|
|
batch = btch
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
// Batch should have the sweep stored.
|
|
|
|
return len(batch.sweeps) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// The primary sweep id should be that of the first inserted sweep.
|
|
|
|
require.Equal(t, batch.primarySweepID, sweepReq1.SwapHash)
|
|
|
|
|
|
|
|
err = lnd.NotifyHeight(601)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// After receiving a height notification the batch will step again,
|
|
|
|
// leading to a new spend monitoring.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return batch.currentHeight == 601
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Create the spending tx that will trigger the spend monitor of the
|
|
|
|
// batch.
|
|
|
|
spendingTx := &wire.MsgTx{
|
|
|
|
Version: 1,
|
|
|
|
// Since the spend monitor is registered on the primary sweep's
|
|
|
|
// outpoint we insert that outpoint here.
|
|
|
|
TxIn: []*wire.TxIn{
|
|
|
|
{
|
|
|
|
PreviousOutPoint: sweepReq1.Outpoint,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
TxOut: []*wire.TxOut{
|
|
|
|
{
|
|
|
|
PkScript: []byte{3, 2, 1},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
spendingTxHash := spendingTx.TxHash()
|
|
|
|
|
|
|
|
// Mock the spend notification that spends the swap.
|
|
|
|
spendDetail := &chainntnfs.SpendDetail{
|
|
|
|
SpentOutPoint: &sweepReq1.Outpoint,
|
|
|
|
SpendingTx: spendingTx,
|
|
|
|
SpenderTxHash: &spendingTxHash,
|
|
|
|
SpenderInputIndex: 0,
|
|
|
|
SpendingHeight: 601,
|
|
|
|
}
|
|
|
|
|
|
|
|
// We notify the spend.
|
|
|
|
lnd.SpendChannel <- spendDetail
|
|
|
|
|
|
|
|
// After receiving the spend, the batch is now monitoring for confs.
|
|
|
|
<-lnd.RegisterConfChannel
|
|
|
|
|
|
|
|
// The batch should eventually read the spend notification and progress
|
|
|
|
// its state to closed.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return batch.state == Closed
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
err = lnd.NotifyHeight(604)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// We mock the tx confirmation notification.
|
|
|
|
lnd.ConfChannel <- &chainntnfs.TxConfirmation{
|
|
|
|
Tx: spendingTx,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Eventually the batch receives the confirmation notification and
|
|
|
|
// confirms itself.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return batch.isComplete()
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testSweepBatcherSweepReentry tests that when an old version of the batch tx
|
2023-10-10 15:18:52 +00:00
|
|
|
// gets confirmed the sweep leftovers are sent back to the batcher.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testSweepBatcherSweepReentry(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2023-10-10 15:18:52 +00:00
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
2024-05-24 08:52:26 +00:00
|
|
|
checkBatcherError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
// Create some sweep requests with timeouts not too far away, in order
|
|
|
|
// to enter the same batch.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap1 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 111,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
err = store.CreateLoopOut(ctx, sweepReq1.SwapHash, swap1)
|
2023-10-10 15:18:52 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
sweepReq2 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{2, 2, 2},
|
|
|
|
Value: 222,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{2, 2},
|
|
|
|
Index: 2,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap2 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 222,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{2},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 111,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq2.SwapHash, swap2)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
sweepReq3 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{3, 3, 3},
|
|
|
|
Value: 333,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{3, 3},
|
|
|
|
Index: 3,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap3 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 333,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{3},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 111,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq3.SwapHash, swap3)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Feed the sweeps to the batcher.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// After inserting the primary (first) sweep, a spend monitor should be
|
|
|
|
// registered.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq2))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq3))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a batch for the sweeps.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Find the batch and store it in a local variable for easier access.
|
|
|
|
b := &batch{}
|
|
|
|
for _, btch := range batcher.batches {
|
|
|
|
if btch.primarySweepID == sweepReq1.SwapHash {
|
|
|
|
b = btch
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Batcher should contain all sweeps.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(b.sweeps) == 3
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Verify that the batch has a primary sweep id that matches the first
|
|
|
|
// inserted sweep, sweep1.
|
|
|
|
require.Equal(t, b.primarySweepID, sweepReq1.SwapHash)
|
|
|
|
|
|
|
|
// Create the spending tx. In order to simulate an older version of the
|
|
|
|
// batch transaction being confirmed, we only insert the primary sweep's
|
|
|
|
// outpoint as a TxIn. This means that the other two sweeps did not
|
|
|
|
// appear in the spending transaction. (This simulates a possible
|
|
|
|
// scenario caused by RBF replacements.)
|
|
|
|
spendingTx := &wire.MsgTx{
|
|
|
|
Version: 1,
|
|
|
|
TxIn: []*wire.TxIn{
|
|
|
|
{
|
|
|
|
PreviousOutPoint: sweepReq1.Outpoint,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
TxOut: []*wire.TxOut{
|
|
|
|
{
|
2024-05-29 01:35:08 +00:00
|
|
|
Value: int64(sweepReq1.Value.ToUnit(
|
|
|
|
btcutil.AmountSatoshi,
|
|
|
|
)),
|
2023-10-10 15:18:52 +00:00
|
|
|
PkScript: []byte{3, 2, 1},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
spendingTxHash := spendingTx.TxHash()
|
|
|
|
|
|
|
|
spendDetail := &chainntnfs.SpendDetail{
|
|
|
|
SpentOutPoint: &sweepReq1.Outpoint,
|
|
|
|
SpendingTx: spendingTx,
|
|
|
|
SpenderTxHash: &spendingTxHash,
|
|
|
|
SpenderInputIndex: 0,
|
|
|
|
SpendingHeight: 601,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send the spending notification to the mock channel.
|
|
|
|
lnd.SpendChannel <- spendDetail
|
|
|
|
|
|
|
|
// After receiving the spend notification the batch should progress to
|
|
|
|
// the next step, which is monitoring for confirmations.
|
|
|
|
<-lnd.RegisterConfChannel
|
|
|
|
|
|
|
|
// Eventually the batch reads the notification and proceeds to a closed
|
|
|
|
// state.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return b.state == Closed
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// While handling the spend notification the batch should detect that
|
|
|
|
// some sweeps did not appear in the spending tx, therefore it redirects
|
|
|
|
// them back to the batcher and the batcher inserts them in a new batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 2
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since second batch was created we check that it registered for its
|
|
|
|
// primary sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// We mock the confirmation notification.
|
|
|
|
lnd.ConfChannel <- &chainntnfs.TxConfirmation{
|
|
|
|
Tx: spendingTx,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Eventually the batch receives the confirmation notification,
|
|
|
|
// gracefully exits and the batcher deletes it.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Find the other batch, which includes the sweeps that did not appear
|
|
|
|
// in the spending tx.
|
|
|
|
b = &batch{}
|
|
|
|
for _, btch := range batcher.batches {
|
|
|
|
b = btch
|
|
|
|
}
|
|
|
|
|
|
|
|
// After all the sweeps enter, it should contain 2 sweeps.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(b.sweeps) == 2
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// The batch should be in an open state.
|
|
|
|
require.Equal(t, b.state, Open)
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testSweepBatcherNonWalletAddr tests that sweep requests that sweep to a non
|
2023-10-10 15:18:52 +00:00
|
|
|
// wallet address enter individual batches.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testSweepBatcherNonWalletAddr(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2023-10-10 15:18:52 +00:00
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
2024-05-24 08:52:26 +00:00
|
|
|
checkBatcherError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap1 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
|
|
|
IsExternalAddr: true,
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
err = store.CreateLoopOut(ctx, sweepReq1.SwapHash, swap1)
|
2023-10-10 15:18:52 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up a
|
|
|
|
// batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Insert the same swap twice, this should be a noop.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Create a second sweep request that has a timeout distance less than
|
|
|
|
// our configured threshold.
|
|
|
|
sweepReq2 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{2, 2, 2},
|
|
|
|
Value: 222,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{2, 2},
|
|
|
|
Index: 2,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap2 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance - 1,
|
|
|
|
AmountRequested: 222,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{2},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
IsExternalAddr: true,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq2.SwapHash, swap2)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq2))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a second batch as first batch is a non wallet
|
|
|
|
// addr batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 2
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Create a third sweep request that has more timeout distance than
|
|
|
|
// the default.
|
|
|
|
sweepReq3 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{3, 3, 3},
|
|
|
|
Value: 333,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{3, 3},
|
|
|
|
Index: 3,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap3 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance + 1,
|
|
|
|
AmountRequested: 333,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{3},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
IsExternalAddr: true,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq3.SwapHash, swap3)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq3))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a new batch as timeout distance is greater than
|
|
|
|
// the threshold
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 3
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
require.Eventually(t, func() bool {
|
2024-05-29 01:35:08 +00:00
|
|
|
// Verify that each batch has the correct number of sweeps
|
|
|
|
// in it.
|
2023-10-10 15:18:52 +00:00
|
|
|
for _, batch := range batcher.batches {
|
|
|
|
switch batch.primarySweepID {
|
|
|
|
case sweepReq1.SwapHash:
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
case sweepReq2.SwapHash:
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
case sweepReq3.SwapHash:
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Check that all sweeps were stored.
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq1.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq2.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq3.SwapHash))
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testSweepBatcherComposite tests that sweep requests that sweep to both wallet
|
2023-10-10 15:18:52 +00:00
|
|
|
// addresses and non-wallet addresses enter the correct batches.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testSweepBatcherComposite(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2023-10-10 15:18:52 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2023-10-10 15:18:52 +00:00
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
2024-05-24 08:52:26 +00:00
|
|
|
checkBatcherError(t, err)
|
2023-10-10 15:18:52 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap1 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
err = store.CreateLoopOut(ctx, sweepReq1.SwapHash, swap1)
|
2023-10-10 15:18:52 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Create a second sweep request that has a timeout distance less than
|
|
|
|
// our configured threshold.
|
|
|
|
sweepReq2 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{2, 2, 2},
|
|
|
|
Value: 222,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{2, 2},
|
|
|
|
Index: 2,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap2 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance - 1,
|
|
|
|
AmountRequested: 222,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{2},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq2.SwapHash, swap2)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Create a third sweep request that has less timeout distance than the
|
|
|
|
// default max, but is not spending to a wallet address.
|
|
|
|
sweepReq3 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{3, 3, 3},
|
|
|
|
Value: 333,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{3, 3},
|
|
|
|
Index: 3,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap3 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance - 3,
|
|
|
|
AmountRequested: 333,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{3},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
IsExternalAddr: true,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq3.SwapHash, swap3)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Create a fourth sweep request that has a timeout which is not valid
|
|
|
|
// for the first batch, so it will cause it to create a new batch.
|
|
|
|
sweepReq4 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{4, 4, 4},
|
|
|
|
Value: 444,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{4, 4},
|
|
|
|
Index: 4,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap4 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance + 1,
|
|
|
|
AmountRequested: 444,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{4},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq4.SwapHash, swap4)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Create a fifth sweep request that has a timeout which is not valid
|
|
|
|
// for the first batch, but a valid timeout for the new batch.
|
|
|
|
sweepReq5 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{5, 5, 5},
|
|
|
|
Value: 555,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{5, 5},
|
|
|
|
Index: 5,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap5 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance + 5,
|
|
|
|
AmountRequested: 555,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{5},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq5.SwapHash, swap5)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Create a sixth sweep request that has a valid timeout for the new
|
|
|
|
// batch, but is paying to a non-wallet address.
|
|
|
|
sweepReq6 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{6, 6, 6},
|
|
|
|
Value: 666,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{6, 6},
|
|
|
|
Index: 6,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap6 := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111 + defaultMaxTimeoutDistance + 6,
|
|
|
|
AmountRequested: 666,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{6},
|
2023-10-10 15:18:52 +00:00
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2023-10-10 15:18:52 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
IsExternalAddr: true,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq6.SwapHash, swap6)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up a
|
|
|
|
// batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Insert the same swap twice, this should be a noop.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq2))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should not create a second batch as timeout distance is small
|
|
|
|
// enough.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq3))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a second batch as this sweep pays to a non
|
|
|
|
// wallet address.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 2
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq4))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a third batch as timeout distance is greater
|
|
|
|
// than the threshold.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 3
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq5))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should not create a fourth batch as timeout distance is small
|
|
|
|
// enough for it to join the last batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 3
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq6))
|
2023-10-10 15:18:52 +00:00
|
|
|
|
|
|
|
// Batcher should create a fourth batch as this sweep pays to a non
|
|
|
|
// wallet address.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
return len(batcher.batches) == 4
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
// Verify that each batch has the correct number of sweeps in
|
|
|
|
// it.
|
|
|
|
for _, batch := range batcher.batches {
|
|
|
|
switch batch.primarySweepID {
|
|
|
|
case sweepReq1.SwapHash:
|
|
|
|
if len(batch.sweeps) != 2 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
case sweepReq3.SwapHash:
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
case sweepReq4.SwapHash:
|
|
|
|
if len(batch.sweeps) != 2 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
case sweepReq6.SwapHash:
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Check that all sweeps were stored.
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq1.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq2.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq3.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq4.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq5.SwapHash))
|
|
|
|
require.True(t, batcherStore.AssertSweepStored(sweepReq6.SwapHash))
|
|
|
|
}
|
2024-05-22 17:30:16 +00:00
|
|
|
|
|
|
|
// makeTestTx creates a test transaction with a single output of the given
|
|
|
|
// value.
|
|
|
|
func makeTestTx(value int64) *wire.MsgTx {
|
|
|
|
tx := wire.NewMsgTx(wire.TxVersion)
|
|
|
|
tx.AddTxOut(wire.NewTxOut(value, nil))
|
|
|
|
return tx
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testGetFeePortionForSweep tests that the fee portion for a sweep is correctly
|
2024-05-22 17:30:16 +00:00
|
|
|
// calculated.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testGetFeePortionForSweep(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2024-05-22 17:30:16 +00:00
|
|
|
tests := []struct {
|
|
|
|
name string
|
|
|
|
spendTxValue int64
|
|
|
|
numSweeps int
|
|
|
|
totalSweptAmt btcutil.Amount
|
|
|
|
expectedFeePortion btcutil.Amount
|
|
|
|
expectedRoundingDiff btcutil.Amount
|
|
|
|
}{
|
|
|
|
{
|
|
|
|
"Even Split",
|
|
|
|
100, 5, 200, 20, 0,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
"Single Sweep",
|
|
|
|
100, 1, 200, 100, 0,
|
|
|
|
},
|
|
|
|
{
|
|
|
|
"With Rounding Diff",
|
|
|
|
200, 4, 350, 37, 2,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, tt := range tests {
|
|
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
|
|
spendTx := makeTestTx(tt.spendTxValue)
|
|
|
|
feePortion, roundingDiff := getFeePortionForSweep(
|
|
|
|
spendTx, tt.numSweeps, tt.totalSweptAmt,
|
|
|
|
)
|
|
|
|
require.Equal(t, tt.expectedFeePortion, feePortion)
|
|
|
|
require.Equal(t, tt.expectedRoundingDiff, roundingDiff)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
}
|
2024-05-24 08:52:54 +00:00
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testRestoringEmptyBatch tests that the batcher can be restored with an empty
|
2024-05-24 08:52:54 +00:00
|
|
|
// batch.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testRestoringEmptyBatch(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2024-05-24 08:52:54 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2024-05-24 08:52:54 +00:00
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
_, err = batcherStore.InsertSweepBatch(ctx, &dbBatch{})
|
2024-05-24 08:52:54 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2024-05-24 08:52:54 +00:00
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
var runErr error
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
runErr = batcher.Run(ctx)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Wait for the batcher to be initialized.
|
|
|
|
<-batcher.initDone
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2024-05-24 08:52:54 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq.SwapHash, swap)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
2024-05-14 00:16:06 +00:00
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq))
|
2024-05-24 08:52:54 +00:00
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up a
|
|
|
|
// batch.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
// Make sure that the sweep was stored and we have exactly one
|
|
|
|
// active batch.
|
|
|
|
return batcherStore.AssertSweepStored(sweepReq.SwapHash) &&
|
|
|
|
len(batcher.batches) == 1
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Make sure we have only one batch stored (as we dropped the dormant
|
|
|
|
// one).
|
|
|
|
batches, err := batcherStore.FetchUnconfirmedSweepBatches(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, batches, 1)
|
|
|
|
|
2024-05-29 18:07:32 +00:00
|
|
|
// Now make the batcher quit by canceling the context.
|
2024-05-24 08:52:54 +00:00
|
|
|
cancel()
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
checkBatcherError(t, runErr)
|
|
|
|
}
|
2024-05-21 01:37:29 +00:00
|
|
|
|
|
|
|
type loopStoreMock struct {
|
|
|
|
loops map[lntypes.Hash]*loopdb.LoopOut
|
|
|
|
mu sync.Mutex
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// backend is the store passed to the test. An empty swap with the ID
|
|
|
|
// passed is stored to this place to satisfy SQL foreign key constraint.
|
|
|
|
backend testStore
|
|
|
|
|
|
|
|
// preimage is last preimage first byte used in fake swap in backend.
|
|
|
|
// It has to be unique to satisfy SQL constraint.
|
|
|
|
preimage byte
|
2024-05-21 01:37:29 +00:00
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
func newLoopStoreMock(backend testStore) *loopStoreMock {
|
2024-05-21 01:37:29 +00:00
|
|
|
return &loopStoreMock{
|
2024-06-17 02:45:07 +00:00
|
|
|
loops: make(map[lntypes.Hash]*loopdb.LoopOut),
|
|
|
|
backend: backend,
|
2024-05-21 01:37:29 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *loopStoreMock) FetchLoopOutSwap(ctx context.Context,
|
|
|
|
hash lntypes.Hash) (*loopdb.LoopOut, error) {
|
|
|
|
|
|
|
|
s.mu.Lock()
|
|
|
|
defer s.mu.Unlock()
|
|
|
|
|
|
|
|
out, has := s.loops[hash]
|
|
|
|
if !has {
|
|
|
|
return nil, errors.New("loop not found")
|
|
|
|
}
|
|
|
|
|
|
|
|
return out, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *loopStoreMock) putLoopOutSwap(hash lntypes.Hash, out *loopdb.LoopOut) {
|
|
|
|
s.mu.Lock()
|
|
|
|
defer s.mu.Unlock()
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
_, existed := s.loops[hash]
|
2024-05-21 01:37:29 +00:00
|
|
|
s.loops[hash] = out
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
if existed {
|
|
|
|
// The swap exists, no need to create one in backend, since it
|
|
|
|
// stores fake data anyway.
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
if _, ok := s.backend.(*loopdb.StoreMock); ok {
|
|
|
|
// Do not create a fake loop in loopdb.StoreMock, because it
|
|
|
|
// blocks on notification channels and this is not needed.
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Put a swap with the same ID to backend store to satisfy SQL foreign
|
|
|
|
// key constraint. Don't store the data to ensure it is not used.
|
|
|
|
err := s.backend.CreateLoopOut(context.Background(), hash,
|
|
|
|
&loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 999,
|
|
|
|
AmountRequested: 999,
|
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{s.preimage},
|
|
|
|
},
|
|
|
|
|
|
|
|
DestAddr: destAddr,
|
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
s.backend.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
s.preimage++
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
2024-05-21 01:37:29 +00:00
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// AssertLoopOutStored asserts that a swap is stored.
|
|
|
|
func (s *loopStoreMock) AssertLoopOutStored() {
|
|
|
|
s.backend.AssertLoopOutStored()
|
|
|
|
}
|
|
|
|
|
|
|
|
// testHandleSweepTwice tests that handing the same sweep twice must not
|
2024-05-21 01:37:29 +00:00
|
|
|
// add it to different batches.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testHandleSweepTwice(t *testing.T, backend testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2024-05-21 01:37:29 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
store := newLoopStoreMock(backend)
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2024-05-21 01:37:29 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2024-05-21 01:37:29 +00:00
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
var runErr error
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
runErr = batcher.Run(ctx)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Wait for the batcher to be initialized.
|
|
|
|
<-batcher.initDone
|
|
|
|
|
|
|
|
const shortCltv = 111
|
|
|
|
const longCltv = 111 + defaultMaxTimeoutDistance + 6
|
|
|
|
|
|
|
|
// Create two sweep requests with CltvExpiry distant from each other
|
|
|
|
// to go assigned to separate batches.
|
|
|
|
sweepReq1 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
loopOut1 := &loopdb.LoopOut{
|
|
|
|
Loop: loopdb.Loop{
|
|
|
|
Hash: lntypes.Hash{1, 1, 1},
|
|
|
|
},
|
|
|
|
Contract: &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: shortCltv,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2024-05-21 01:37:29 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
sweepReq2 := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{2, 2, 2},
|
|
|
|
Value: 222,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{2, 2},
|
|
|
|
Index: 2,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
loopOut2 := &loopdb.LoopOut{
|
|
|
|
Loop: loopdb.Loop{
|
|
|
|
Hash: lntypes.Hash{2, 2, 2},
|
|
|
|
},
|
|
|
|
Contract: &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: longCltv,
|
|
|
|
AmountRequested: 222,
|
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2024-05-21 01:37:29 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
store.putLoopOutSwap(sweepReq1.SwapHash, loopOut1)
|
|
|
|
store.putLoopOutSwap(sweepReq2.SwapHash, loopOut2)
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq1))
|
|
|
|
|
|
|
|
// Since two batches were created we check that it registered for its
|
|
|
|
// primary sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Deliver the second sweep. It will go to a separate batch,
|
|
|
|
// since CltvExpiry values are distant enough.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq2))
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up
|
|
|
|
// batches.
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
// Make sure that the sweep was stored and we have exactly one
|
|
|
|
// active batch.
|
|
|
|
return batcherStore.AssertSweepStored(sweepReq1.SwapHash) &&
|
|
|
|
batcherStore.AssertSweepStored(sweepReq2.SwapHash) &&
|
|
|
|
len(batcher.batches) == 2
|
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Change the second sweep so that it can be added to the first batch.
|
|
|
|
// Change CltvExpiry.
|
|
|
|
loopOut2 = &loopdb.LoopOut{
|
|
|
|
Loop: loopdb.Loop{
|
|
|
|
Hash: lntypes.Hash{2, 2, 2},
|
|
|
|
},
|
|
|
|
Contract: &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: shortCltv,
|
|
|
|
AmountRequested: 222,
|
|
|
|
},
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2024-05-21 01:37:29 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
store.putLoopOutSwap(sweepReq2.SwapHash, loopOut2)
|
|
|
|
|
|
|
|
// Re-add the second sweep. It is expected to stay in second batch,
|
|
|
|
// not added to both batches.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq2))
|
|
|
|
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
// Make sure there are two batches.
|
|
|
|
batches := batcher.batches
|
|
|
|
if len(batches) != 2 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// Find the batch with largest ID. It must be the second batch.
|
|
|
|
// Variable batches is a map, not a slice, so we have to visit
|
|
|
|
// all the items and find the one with maximum id.
|
|
|
|
var secondBatch *batch
|
|
|
|
for _, batch := range batches {
|
|
|
|
if secondBatch == nil || batch.id > secondBatch.id {
|
|
|
|
secondBatch = batch
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-05-21 01:37:29 +00:00
|
|
|
// Make sure the second batch has the second sweep.
|
2024-06-17 02:45:07 +00:00
|
|
|
sweep2, has := secondBatch.sweeps[sweepReq2.SwapHash]
|
2024-05-21 01:37:29 +00:00
|
|
|
if !has {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure the second sweep's timeout has been updated.
|
2024-06-17 02:45:07 +00:00
|
|
|
return sweep2.timeout == shortCltv
|
2024-05-21 01:37:29 +00:00
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Make sure each batch has one sweep. If the second sweep was added to
|
|
|
|
// both batches, the following check won't pass.
|
2024-06-17 02:45:07 +00:00
|
|
|
for _, batch := range batcher.batches {
|
|
|
|
require.Equal(t, 1, len(batch.sweeps))
|
|
|
|
}
|
2024-05-21 01:37:29 +00:00
|
|
|
|
2024-05-29 18:07:32 +00:00
|
|
|
// Now make the batcher quit by canceling the context.
|
2024-05-21 01:37:29 +00:00
|
|
|
cancel()
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
checkBatcherError(t, runErr)
|
|
|
|
}
|
2024-05-29 18:07:32 +00:00
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testRestoringPreservesConfTarget tests that after the batch is written to DB
|
2024-05-29 18:07:32 +00:00
|
|
|
// and loaded back, its batchConfTarget value is preserved.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testRestoringPreservesConfTarget(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2024-05-29 18:07:32 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
2024-05-29 18:07:32 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2024-05-29 18:07:32 +00:00
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
var runErr error
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
runErr = batcher.Run(ctx)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Wait for the batcher to be initialized.
|
|
|
|
<-batcher.initDone
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
},
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2024-05-29 18:07:32 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 123,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
err = store.CreateLoopOut(ctx, sweepReq.SwapHash, swap)
|
2024-05-29 18:07:32 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq))
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up a
|
|
|
|
// batch.
|
|
|
|
require.Eventually(t, func() bool {
|
2024-06-17 02:45:07 +00:00
|
|
|
// Make sure that the sweep was stored
|
|
|
|
if !batcherStore.AssertSweepStored(sweepReq.SwapHash) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure there is exactly one active batch.
|
|
|
|
if len(batcher.batches) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get the batch.
|
|
|
|
batch := getOnlyBatch(batcher)
|
|
|
|
|
|
|
|
// Make sure the batch has one sweep.
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure the batch has proper batchConfTarget.
|
|
|
|
return batch.cfg.batchConfTarget == 123
|
2024-05-29 18:07:32 +00:00
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Make sure we have stored the batch.
|
|
|
|
batches, err := batcherStore.FetchUnconfirmedSweepBatches(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, batches, 1)
|
|
|
|
|
|
|
|
// Now make the batcher quit by canceling the context.
|
|
|
|
cancel()
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
// Make sure the batcher exited without an error.
|
|
|
|
checkBatcherError(t, runErr)
|
|
|
|
|
|
|
|
// Now launch it again.
|
|
|
|
batcher = NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-05-14 14:36:47 +00:00
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
2024-05-29 18:07:32 +00:00
|
|
|
ctx, cancel = context.WithCancel(context.Background())
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
runErr = batcher.Run(ctx)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Wait for the batcher to be initialized.
|
|
|
|
<-batcher.initDone
|
|
|
|
|
|
|
|
// Wait for batch to load.
|
|
|
|
require.Eventually(t, func() bool {
|
2024-06-17 02:45:07 +00:00
|
|
|
// Make sure that the sweep was stored
|
|
|
|
if !batcherStore.AssertSweepStored(sweepReq.SwapHash) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure there is exactly one active batch.
|
|
|
|
if len(batcher.batches) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get the batch.
|
|
|
|
batch := getOnlyBatch(batcher)
|
|
|
|
|
|
|
|
// Make sure the batch has one sweep.
|
|
|
|
return len(batch.sweeps) == 1
|
2024-05-29 18:07:32 +00:00
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
|
|
|
// Make sure batchConfTarget was preserved.
|
2024-06-17 02:45:07 +00:00
|
|
|
require.Equal(t, 123, int(getOnlyBatch(batcher).cfg.batchConfTarget))
|
2024-05-29 18:07:32 +00:00
|
|
|
|
|
|
|
// Expect registration for spend notification.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Now make the batcher quit by canceling the context.
|
|
|
|
cancel()
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
// Make sure the batcher exited without an error.
|
|
|
|
checkBatcherError(t, runErr)
|
|
|
|
}
|
2024-05-14 14:36:47 +00:00
|
|
|
|
|
|
|
type sweepFetcherMock struct {
|
|
|
|
store map[lntypes.Hash]*SweepInfo
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *sweepFetcherMock) FetchSweep(ctx context.Context, hash lntypes.Hash) (
|
|
|
|
*SweepInfo, error) {
|
|
|
|
|
|
|
|
return f.store[hash], nil
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testSweepFetcher tests providing custom sweep fetcher to Batcher.
|
|
|
|
func testSweepFetcher(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
|
|
|
|
// Extract payment address from the invoice.
|
|
|
|
swapPaymentAddr, err := utils.ObtainSwapPaymentAddr(
|
|
|
|
swapInvoice, lnd.ChainParams,
|
|
|
|
)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2024-06-22 02:20:52 +00:00
|
|
|
swapHash := lntypes.Hash{1, 1, 1}
|
|
|
|
|
|
|
|
// Provide min fee rate for the sweep.
|
|
|
|
feeRate := chainfee.SatPerKWeight(30000)
|
|
|
|
amt := btcutil.Amount(1_000_000)
|
|
|
|
weight := lntypes.WeightUnit(445) // Weight for 1-to-1 tx.
|
|
|
|
bumpedFee := feeRate + 100
|
|
|
|
expectedFee := bumpedFee.FeeForWeight(weight)
|
|
|
|
|
|
|
|
swap := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 222,
|
|
|
|
AmountRequested: amt,
|
|
|
|
ProtocolVersion: loopdb.ProtocolVersionMuSig2,
|
|
|
|
HtlcKeys: loopdb.HtlcKeys{
|
|
|
|
SenderScriptKey: senderKey,
|
|
|
|
ReceiverScriptKey: receiverKey,
|
|
|
|
SenderInternalPubKey: senderKey,
|
|
|
|
ReceiverInternalPubKey: receiverKey,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
DestAddr: destAddr,
|
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 321,
|
|
|
|
}
|
|
|
|
|
|
|
|
htlc, err := utils.GetHtlc(
|
|
|
|
swapHash, &swap.SwapContract, lnd.ChainParams,
|
|
|
|
)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
sweepInfo := &SweepInfo{
|
|
|
|
ConfTarget: 123,
|
|
|
|
Timeout: 111,
|
|
|
|
SwapInvoicePaymentAddr: *swapPaymentAddr,
|
|
|
|
MinFeeRate: feeRate,
|
|
|
|
ProtocolVersion: loopdb.ProtocolVersionMuSig2,
|
|
|
|
HTLCKeys: loopdb.HtlcKeys{
|
|
|
|
SenderScriptKey: senderKey,
|
|
|
|
ReceiverScriptKey: receiverKey,
|
|
|
|
SenderInternalPubKey: senderKey,
|
|
|
|
ReceiverInternalPubKey: receiverKey,
|
|
|
|
},
|
|
|
|
HTLC: *htlc,
|
|
|
|
HTLCSuccessEstimator: htlc.AddSuccessToEstimator,
|
|
|
|
DestAddr: destAddr,
|
|
|
|
}
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
sweepFetcher := &sweepFetcherMock{
|
|
|
|
store: map[lntypes.Hash]*SweepInfo{
|
2024-06-22 02:20:52 +00:00
|
|
|
swapHash: sweepInfo,
|
2024-05-14 14:36:47 +00:00
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq := SweepRequest{
|
2024-06-22 02:20:52 +00:00
|
|
|
SwapHash: swapHash,
|
|
|
|
Value: amt,
|
2024-06-17 02:45:07 +00:00
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a swap in the DB. It is needed to satisfy SQL constraints in
|
|
|
|
// case of SQL test. The data is not actually used, since we pass sweep
|
|
|
|
// fetcher, so put different conf target to make sure it is not used.
|
2024-06-22 02:20:52 +00:00
|
|
|
err = store.CreateLoopOut(ctx, swapHash, swap)
|
2024-06-17 02:45:07 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
2024-05-14 14:36:47 +00:00
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
2024-06-22 02:20:52 +00:00
|
|
|
testMuSig2SignSweep, testVerifySchnorrSig, lnd.ChainParams,
|
|
|
|
batcherStore, sweepFetcher)
|
2024-05-14 14:36:47 +00:00
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
var runErr error
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
runErr = batcher.Run(ctx)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Wait for the batcher to be initialized.
|
|
|
|
<-batcher.initDone
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq))
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Once batcher receives sweep request it will eventually spin up a
|
|
|
|
// batch.
|
|
|
|
require.Eventually(t, func() bool {
|
2024-06-17 02:45:07 +00:00
|
|
|
// Make sure that the sweep was stored
|
2024-06-22 02:20:52 +00:00
|
|
|
if !batcherStore.AssertSweepStored(swapHash) {
|
2024-06-17 02:45:07 +00:00
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure there is exactly one active batch.
|
|
|
|
if len(batcher.batches) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Get the batch.
|
|
|
|
batch := getOnlyBatch(batcher)
|
|
|
|
|
|
|
|
// Make sure the batch has one sweep.
|
|
|
|
if len(batch.sweeps) != 1 {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure the batch has proper batchConfTarget.
|
|
|
|
return batch.cfg.batchConfTarget == 123
|
2024-05-14 14:36:47 +00:00
|
|
|
}, test.Timeout, eventuallyCheckFrequency)
|
|
|
|
|
2024-06-22 02:20:52 +00:00
|
|
|
// Get the published transaction and check the fee rate.
|
|
|
|
tx := <-lnd.TxPublishChannel
|
|
|
|
out := btcutil.Amount(tx.TxOut[0].Value)
|
|
|
|
gotFee := amt - out
|
|
|
|
require.Equal(t, expectedFee, gotFee, "fees don't match")
|
|
|
|
|
2024-05-14 14:36:47 +00:00
|
|
|
// Make sure we have stored the batch.
|
|
|
|
batches, err := batcherStore.FetchUnconfirmedSweepBatches(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, batches, 1)
|
|
|
|
|
|
|
|
// Now make the batcher quit by canceling the context.
|
|
|
|
cancel()
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
// Make sure the batcher exited without an error.
|
|
|
|
checkBatcherError(t, runErr)
|
|
|
|
}
|
2024-06-15 02:31:23 +00:00
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testSweepBatcherCloseDuringAdding tests that sweep batcher works correctly
|
2024-06-15 02:31:23 +00:00
|
|
|
// if it is closed (stops running) during AddSweep call.
|
2024-06-17 02:45:07 +00:00
|
|
|
func testSweepBatcherCloseDuringAdding(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
2024-06-15 02:31:23 +00:00
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
|
|
|
testMuSig2SignSweep, nil, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore)
|
|
|
|
go func() {
|
|
|
|
err := batcher.Run(ctx)
|
|
|
|
checkBatcherError(t, err)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Add many swaps.
|
|
|
|
for i := byte(1); i < 255; i++ {
|
|
|
|
swapHash := lntypes.Hash{i, i, i}
|
|
|
|
|
|
|
|
// Create a swap contract.
|
|
|
|
swap := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
2024-06-17 02:45:07 +00:00
|
|
|
|
|
|
|
// Make preimage unique to pass SQL constraints.
|
|
|
|
Preimage: lntypes.Preimage{i},
|
2024-06-15 02:31:23 +00:00
|
|
|
},
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
DestAddr: destAddr,
|
2024-06-15 02:31:23 +00:00
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, swapHash, swap)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
// Add many sweeps.
|
|
|
|
for i := byte(1); i < 255; i++ {
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{i, i, i},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{i, i},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
|
|
|
err := batcher.AddSweep(&sweepReq)
|
|
|
|
if err == ErrBatcherShuttingDown {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
require.NoError(t, err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
// Close sweepbatcher during addings.
|
|
|
|
time.Sleep(1 * time.Millisecond)
|
|
|
|
cancel()
|
|
|
|
}()
|
|
|
|
|
|
|
|
// We don't know how many spend notification registrations will be
|
|
|
|
// issued, so accept them while waiting for two goroutines to stop.
|
|
|
|
quit := make(chan struct{})
|
|
|
|
registrationChan := make(chan struct{})
|
|
|
|
go func() {
|
|
|
|
defer close(registrationChan)
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-lnd.RegisterSpendChannel:
|
|
|
|
case <-quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
wg.Wait()
|
|
|
|
close(quit)
|
|
|
|
<-registrationChan
|
|
|
|
}
|
2024-06-17 02:45:07 +00:00
|
|
|
|
2024-06-25 02:18:16 +00:00
|
|
|
// testCustomSignMuSig2 tests the operation with custom musig2 signer.
|
|
|
|
func testCustomSignMuSig2(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
|
|
|
defer test.Guard(t)()
|
|
|
|
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
|
|
|
|
sweepStore, err := NewSweepFetcherFromSwapStore(store, lnd.ChainParams)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Use custom MuSig2 signer function.
|
|
|
|
batcher := NewBatcher(lnd.WalletKit, lnd.ChainNotifier, lnd.Signer,
|
|
|
|
nil, testVerifySchnorrSig, lnd.ChainParams, batcherStore,
|
|
|
|
sweepStore, WithCustomSignMuSig2(testSignMuSig2func))
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
|
|
|
|
|
|
|
var runErr error
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
runErr = batcher.Run(ctx)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Wait for the batcher to be initialized.
|
|
|
|
<-batcher.initDone
|
|
|
|
|
|
|
|
// Create a sweep request.
|
|
|
|
sweepReq := SweepRequest{
|
|
|
|
SwapHash: lntypes.Hash{1, 1, 1},
|
|
|
|
Value: 111,
|
|
|
|
Outpoint: wire.OutPoint{
|
|
|
|
Hash: chainhash.Hash{1, 1},
|
|
|
|
Index: 1,
|
|
|
|
},
|
|
|
|
Notifier: &dummyNotifier,
|
|
|
|
}
|
|
|
|
|
|
|
|
swap := &loopdb.LoopOutContract{
|
|
|
|
SwapContract: loopdb.SwapContract{
|
|
|
|
CltvExpiry: 111,
|
|
|
|
AmountRequested: 111,
|
|
|
|
ProtocolVersion: loopdb.ProtocolVersionMuSig2,
|
|
|
|
HtlcKeys: loopdb.HtlcKeys{
|
|
|
|
SenderScriptKey: senderKey,
|
|
|
|
ReceiverScriptKey: receiverKey,
|
|
|
|
SenderInternalPubKey: senderKey,
|
|
|
|
ReceiverInternalPubKey: receiverKey,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
|
|
|
|
DestAddr: destAddr,
|
|
|
|
SwapInvoice: swapInvoice,
|
|
|
|
SweepConfTarget: 111,
|
|
|
|
}
|
|
|
|
|
|
|
|
err = store.CreateLoopOut(ctx, sweepReq.SwapHash, swap)
|
|
|
|
require.NoError(t, err)
|
|
|
|
store.AssertLoopOutStored()
|
|
|
|
|
|
|
|
// Deliver sweep request to batcher.
|
|
|
|
require.NoError(t, batcher.AddSweep(&sweepReq))
|
|
|
|
|
|
|
|
// Since a batch was created we check that it registered for its primary
|
|
|
|
// sweep's spend.
|
|
|
|
<-lnd.RegisterSpendChannel
|
|
|
|
|
|
|
|
// Wait for tx to be published.
|
|
|
|
tx := <-lnd.TxPublishChannel
|
|
|
|
|
|
|
|
// Check the signature.
|
|
|
|
gotSig := tx.TxIn[0].Witness[0]
|
|
|
|
require.Equal(t, customSignature, gotSig, "signatures don't match")
|
|
|
|
|
|
|
|
// Now make the batcher quit by canceling the context.
|
|
|
|
cancel()
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
checkBatcherError(t, runErr)
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// TestSweepBatcherBatchCreation tests that sweep requests enter the expected
|
|
|
|
// batch based on their timeout distance.
|
|
|
|
func TestSweepBatcherBatchCreation(t *testing.T) {
|
|
|
|
runTests(t, testSweepBatcherBatchCreation)
|
|
|
|
}
|
|
|
|
|
2024-06-22 02:19:51 +00:00
|
|
|
// TestFeeBumping tests that sweep is RBFed with slightly higher fee rate after
|
|
|
|
// each block unless WithNoBumping is passed.
|
|
|
|
func TestFeeBumping(t *testing.T) {
|
|
|
|
t.Run("regular", func(t *testing.T) {
|
|
|
|
runTests(t, func(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
|
|
|
testFeeBumping(t, store, batcherStore, false)
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("WithNoBumping", func(t *testing.T) {
|
|
|
|
runTests(t, func(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore) {
|
|
|
|
|
|
|
|
testFeeBumping(t, store, batcherStore, true)
|
|
|
|
})
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// TestSweepBatcherSimpleLifecycle tests the simple lifecycle of the batches
|
|
|
|
// that are created and run by the batcher.
|
|
|
|
func TestSweepBatcherSimpleLifecycle(t *testing.T) {
|
|
|
|
runTests(t, testSweepBatcherSimpleLifecycle)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestSweepBatcherSweepReentry tests that when an old version of the batch tx
|
|
|
|
// gets confirmed the sweep leftovers are sent back to the batcher.
|
|
|
|
func TestSweepBatcherSweepReentry(t *testing.T) {
|
|
|
|
runTests(t, testSweepBatcherSweepReentry)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestSweepBatcherNonWalletAddr tests that sweep requests that sweep to a non
|
|
|
|
// wallet address enter individual batches.
|
|
|
|
func TestSweepBatcherNonWalletAddr(t *testing.T) {
|
|
|
|
runTests(t, testSweepBatcherNonWalletAddr)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestSweepBatcherComposite tests that sweep requests that sweep to both wallet
|
|
|
|
// addresses and non-wallet addresses enter the correct batches.
|
|
|
|
func TestSweepBatcherComposite(t *testing.T) {
|
|
|
|
runTests(t, testSweepBatcherComposite)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestGetFeePortionForSweep tests that the fee portion for a sweep is correctly
|
|
|
|
// calculated.
|
|
|
|
func TestGetFeePortionForSweep(t *testing.T) {
|
|
|
|
runTests(t, testGetFeePortionForSweep)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestRestoringEmptyBatch tests that the batcher can be restored with an empty
|
|
|
|
// batch.
|
|
|
|
func TestRestoringEmptyBatch(t *testing.T) {
|
|
|
|
runTests(t, testRestoringEmptyBatch)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestHandleSweepTwice tests that handing the same sweep twice must not
|
|
|
|
// add it to different batches.
|
|
|
|
func TestHandleSweepTwice(t *testing.T) {
|
|
|
|
runTests(t, testHandleSweepTwice)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestRestoringPreservesConfTarget tests that after the batch is written to DB
|
|
|
|
// and loaded back, its batchConfTarget value is preserved.
|
|
|
|
func TestRestoringPreservesConfTarget(t *testing.T) {
|
|
|
|
runTests(t, testRestoringPreservesConfTarget)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestSweepFetcher tests providing custom sweep fetcher to Batcher.
|
|
|
|
func TestSweepFetcher(t *testing.T) {
|
|
|
|
runTests(t, testSweepFetcher)
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestSweepBatcherCloseDuringAdding tests that sweep batcher works correctly
|
|
|
|
// if it is closed (stops running) during AddSweep call.
|
|
|
|
func TestSweepBatcherCloseDuringAdding(t *testing.T) {
|
|
|
|
runTests(t, testSweepBatcherCloseDuringAdding)
|
|
|
|
}
|
|
|
|
|
2024-06-25 02:18:16 +00:00
|
|
|
// TestCustomSignMuSig2 tests the operation with custom musig2 signer.
|
|
|
|
func TestCustomSignMuSig2(t *testing.T) {
|
|
|
|
runTests(t, testCustomSignMuSig2)
|
|
|
|
}
|
|
|
|
|
2024-06-17 02:45:07 +00:00
|
|
|
// testBatcherStore is BatcherStore used in tests.
|
|
|
|
type testBatcherStore interface {
|
|
|
|
BatcherStore
|
|
|
|
|
|
|
|
// AssertSweepStored asserts that a sweep is stored.
|
|
|
|
AssertSweepStored(id lntypes.Hash) bool
|
|
|
|
}
|
|
|
|
|
|
|
|
type loopdbBatcherStore struct {
|
|
|
|
BatcherStore
|
|
|
|
|
|
|
|
sweepsSet map[lntypes.Hash]struct{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpsertSweep inserts a sweep into the database, or updates an existing sweep
|
|
|
|
// if it already exists. This wrapper was added to update sweepsSet.
|
|
|
|
func (s *loopdbBatcherStore) UpsertSweep(ctx context.Context,
|
|
|
|
sweep *dbSweep) error {
|
|
|
|
|
|
|
|
err := s.BatcherStore.UpsertSweep(ctx, sweep)
|
|
|
|
if err == nil {
|
|
|
|
s.sweepsSet[sweep.SwapHash] = struct{}{}
|
|
|
|
}
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// AssertSweepStored asserts that a sweep is stored.
|
|
|
|
func (s *loopdbBatcherStore) AssertSweepStored(id lntypes.Hash) bool {
|
|
|
|
_, has := s.sweepsSet[id]
|
|
|
|
return has
|
|
|
|
}
|
|
|
|
|
|
|
|
// testStore is loopdb used in tests.
|
|
|
|
type testStore interface {
|
|
|
|
loopdb.SwapStore
|
|
|
|
|
|
|
|
// AssertLoopOutStored asserts that a swap is stored.
|
|
|
|
AssertLoopOutStored()
|
|
|
|
}
|
|
|
|
|
|
|
|
// loopdbStore wraps loopdb.SwapStore and implements testStore interface.
|
|
|
|
type loopdbStore struct {
|
|
|
|
loopdb.SwapStore
|
|
|
|
|
|
|
|
t *testing.T
|
|
|
|
|
|
|
|
loopOutStoreChan chan struct{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// newLoopdbStore creates new loopdbStore instance.
|
|
|
|
func newLoopdbStore(t *testing.T, swapStore loopdb.SwapStore) *loopdbStore {
|
|
|
|
return &loopdbStore{
|
|
|
|
SwapStore: swapStore,
|
|
|
|
t: t,
|
|
|
|
loopOutStoreChan: make(chan struct{}, 1),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// CreateLoopOut adds an initiated swap to the store.
|
|
|
|
func (s *loopdbStore) CreateLoopOut(ctx context.Context, hash lntypes.Hash,
|
|
|
|
swap *loopdb.LoopOutContract) error {
|
|
|
|
|
|
|
|
err := s.SwapStore.CreateLoopOut(ctx, hash, swap)
|
|
|
|
if err == nil {
|
|
|
|
s.loopOutStoreChan <- struct{}{}
|
|
|
|
}
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// AssertLoopOutStored asserts that a swap is stored.
|
|
|
|
func (s *loopdbStore) AssertLoopOutStored() {
|
|
|
|
s.t.Helper()
|
|
|
|
|
|
|
|
select {
|
|
|
|
case <-s.loopOutStoreChan:
|
|
|
|
case <-time.After(test.Timeout):
|
|
|
|
s.t.Fatalf("expected swap to be stored")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// runTests runs a test with both mock and loopdb.
|
|
|
|
func runTests(t *testing.T, testFn func(t *testing.T, store testStore,
|
|
|
|
batcherStore testBatcherStore)) {
|
|
|
|
|
|
|
|
t.Run("mocks", func(t *testing.T) {
|
|
|
|
store := loopdb.NewStoreMock(t)
|
|
|
|
batcherStore := NewStoreMock()
|
|
|
|
testFn(t, store, batcherStore)
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("loopdb", func(t *testing.T) {
|
|
|
|
sqlDB := loopdb.NewTestDB(t)
|
|
|
|
typedSqlDB := loopdb.NewTypedStore[Querier](sqlDB)
|
|
|
|
lnd := test.NewMockLnd()
|
|
|
|
batcherStore := NewSQLStore(typedSqlDB, lnd.ChainParams)
|
|
|
|
testStore := newLoopdbStore(t, sqlDB)
|
|
|
|
testBatcherStore := &loopdbBatcherStore{
|
|
|
|
BatcherStore: batcherStore,
|
|
|
|
sweepsSet: make(map[lntypes.Hash]struct{}),
|
|
|
|
}
|
|
|
|
testFn(t, testStore, testBatcherStore)
|
|
|
|
})
|
|
|
|
}
|