24 KiB
Numeric Outcome DLCs
Introduction
When dealing with enumerated outcomes, DLCs require a single nonce and Contract Execution Transactions (CETs) are claimed using a single oracle signature. This scheme results in DLCs which contain a unique CET for every possible outcome, which is only feasible if the number of possible outcomes is of manageable size.
When an outcome can be any of a large range of numbers, then using a simple enumeration of all possible numbers in this range is unwieldy. We optimize by using numeric decomposition in which the oracle signs each digit of the outcome individually so that many possible outcomes can be compressed into a single CET by ignoring certain signatures. There will be as many nonces as there are possible digits required and CETs are claimed using some number of these signatures, not necessarily all of them. If not all signatures are used, then this CET corresponds to all events which agree on the digits for which signatures are used and may have any value at all other digits for which signatures are ignored.
Table of Contents
TODO
Adaptor Points with Multiple Signatures
Given public key P and nonces R1, ..., Rn we can compute n individual signature points for
a given event (d1, ..., dn) in the usual way: si * G = Ri + H(P, Ri, di)*P.
To compute a composite adaptor point for all events which agree on the first m digits, where
m is any positive number less than or equal to n, the sum of the corresponding signature
points is used: s(1..m) * G = (s1 + s2 + ... + sm) * G = s1 * G + s2 * G + ... + sm * G.
When the oracle broadcasts its n signatures s1, ..., sn, the corresponding adaptor secret can be
computed as s(1..m) = s1 + s2 + ... + sm which can be used to broadcast the CET.
Rationale
This design allows implementations to re-use all transaction construction code without modification because every CET needs as input exactly one adaptor point just like in the single-nonce setting.
Another design that was considered was adding keys to the funding output so that parties could collaboratively
construct m adaptor signatures and where n signatures are put on-chain in every CET which would reveal
all oracle signatures to both parties when a CET is published.
This design's major drawbacks is that it creates a very distinct fingerprint and makes CET fees significantly worse.
This design's only benefit is that it results in simpler (although larger) fraud proofs.
The large multi-signature design was abandoned because the above proposal is sufficient to generate fraud proofs.
If an oracle incorrectly signs for an event, then only the sum of the digit signatures s(1..m)
is recoverable on-chain using the adaptor signature which was given to one's counter-party.
This sum is sufficient information to determine what was signed however as one can iterate through
all possible composite adaptor points until they find one whose pre-image is the signature sum found on-chain.
This will determine what digits (d1, ..., dm) were signed and these values along with the oracle
announcement and s(1..m) is sufficient information to generate a fraud proof in the multi-nonce setting.
Contract Execution Transaction Compression
Anytime there is a range of numeric outcomes [start, end] which result in the same payouts for all parties,
then a compression function can be run to reduce the number of CETs to be O(log(L)) instead of L where
L = end - start + 1 is the length of the interval being compressed.
Most contracts are expected to be concerned with some subset of the total possible domain and every outcome before or after that range will result in some constant maximal or minimal payout. This means that compression will drastically reduce the number of CETs to be of the order of the size of the probable domain, with further optimizations available when parties are willing to do some rounding.
The compression algorithm takes as input a range [start, end], a base B, and the number of digits
n (being signed by the oracle) and returns an array of arrays of integers (which will all be in the range [0, B-1]).
An array of integers corresponds to a single event equal to the concatenation of these integers (interpreted in base B).
Example
Before specifying the algorithm, let us run through a general example.
Consider the range [(prefix)wxyz, (prefix)WXYZ] where prefix is some string of digits in base B which
start and end share and w, x, y, and z are the unique digits of start in base B while W, X, Y, and Z
are the unique digits of end in base B.
To cover all cases while looking at as few digits as possible in this (general) range we need only consider
(prefix)wxyz, (prefix)WXYZ and the following cases:
wxy(z+1), wxy(z+2), ..., wxy(B-1),
wx(y+1)_, wx(y+2)_, ..., wx(B-1)_,
w(x+1)__, w(x+2)__, ..., w(B-1)__,
(w+1)___, (w+2)___, ..., (W-1)___,
W0__, W1__, ..., W(X-1)__,
WX0_, WX1_, ..., WX(Y-1)_,
WXY0, WXY1, ..., WXY(Z-1)
where _ refers to an ignored digit (an omission from the array of integers) and all of these cases have the prefix.
I refer to the first three rows as the front groupings the fourth row as the middle grouping and the last three rows
as the back groupings.
Notice that the patterns for the front and back groupings are nearly identical, and that in total the number of CETs that
will be required to cover the range will be equal to the sum of the unique digits of end plus the sum of B-1 minus the
unique digits of start.
This means that the number of CETs required to cover a range of length L will be O(B*log_B(L)) because log_B(L)
corresponds to the number of unique digits between the start and end of the range and for each unique digit a row is
generated in both the front and back groupings of length at most B-1 which corresponds to the coefficient in the order bound.
This counting shows us that base 2 is the optimal base to be using in general cases as it will outperform all larger bases
in both large and small ranges in general.
Note that there are two more possible optimizations to be made, which I call the row optimization, using the outliers wxyz and WXYZ.
If z=0 then the entire first row can be replaced with wxy_ and if Z=B-1 then the entire last row can be replaced with WXY_.
There are another two possible optimizations in the case where the front or back groupings are not needed, which
I call grouping optimization, that again use the outliers to the above pattern wxyz and WXYZ.
For the front this is the case when x=y=z=0 so that the front groupings can be replaced with w___.
Likewise if X = Y = Z = B-1 then the back groupings can be replace with W___.
Lastly, if both grouping optimizations can be made, w = 0 and W = B-1 then a total optimization can be made and the
whole range can be represented using only a single CET corresponding to (prefix)____.
Algorithms
We will first need a function to decompose any number into its digits in a given base.
def decompose(num: Long, base: Int, numDigits: Int): Vector[Int] = {
var currentNum: Long = num
// Note that (0 until numDigits) includes 0 and excludes numDigits
val backwardsDigits = (0 until numDigits).toVector.map { _ =>
val digit = currentNum % base
currentNum = currentNum / base // Note that this is integer division
digit.toInt
}
backwardsDigits.reverse
}
I will use this function for the purposes of this specification but note that when iterating through a range of sequential numbers, there are faster algorithms for computing sequential decompositions.
We will then need a function to compute the shared prefix and the unique digits of start and end.
def separatePrefix(start: Long, end: Long, base: Int, numDigits: Int): (Vector[Int], Vector[Int], Vector[Int]) = {
val startDigits = decompose(startIndex, base, numDigits)
val endDigits = decompose(endIndex, base, numDigits)
val prefixDigits = startDigits
.zip(endDigits)
.takeWhile { case (startDigit, endDigit) => startDigit == endDigit }
.map(_._1)
(prefixDigits, startDigits.drop(prefixDigits.length), endDigits.drop(prefixDigits.length))
}
Now we need the algorithms for computing the groupings.
def frontGroupings(
digits: Vector[Int], // The unique digits of the range's start
base: Int): Vector[Vector[Int]] = {
if (digits.tail.forall(_ == 0)) { // Grouping Optimization
Vector(Vector(digits.head))
} else {
val fromFront = digits.reverse.zipWithIndex.init.flatMap { // Note the flatMap collapses the rows of the grouping
case (lastImportantDigit, unimportantDigits) =>
val fixedDigits = digits.dropRight(unimportantDigits + 1).reverse
(lastImportantDigit + 1).until(base).map { lastDigit => // Note that this range excludes lastImportantDigit and base
fixedDigits :+ lastDigit
}
}
if (digits.last == 0) { // Row Optimization
digits.init +: fromFront.drop(base - 1) // Note init drops the last digit and drop(base-1) drops the fist row
} else {
digits +: fromFront // Add outlier
}
}
}
def backGroupings(
digits: Vector[Int], // The unique digits of the range's end
base: Int): Vector[Vector[Int]] = {
if (digits.tail.forall(_ == base - 1)) { // Grouping Optimization
Vector(Vector(digits.head))
} else {
// Here we compute the back groupings in reverse so as to use the same iteration as in front groupings
val fromBack = digits.reverse.zipWithIndex.init.flatMap { // Note the flatMap collapses the rows of the grouping
case (lastImportantDigit, unimportantDigits) =>
val fixedDigits = digits.dropRight(unimportantDigits + 1)
0.until(lastImportantDigit).reverse.toVector.map { // Note that this range excludes lastImportantDigit
lastDigit =>
fixedDigits :+ lastDigit
}
}
if (uniqueEndDigits.head == base - 1) { // Row Optimization
fromBack.drop(base - 1).reverse :+ digits.init // Note init drops the last digit and drop(base-1) drops the last row
} else {
fromBack.reverse :+ digits // Add outlier
}
}
}
def middleGrouping(
firstDigitStart: Int, // The first unique digit of the range's start
firstDigitEnd: Int): Vector[Vector[Int]] = { // The first unique digit of the range's end
(firstDigitStart + 1).until(firstDigitEnd).toVector.map { firstDigit => // Note that this range excludes firstDigitEnd
Vector(firstDigit)
}
}
Finally we are able to use all of these pieces to compress a range to an approximately minimal number of outcomes (by ignoring digits).
def groupByIgnoringDigits(start: Long, end: Long, base: Int, numDigits: Int): Vector[Vector[Int]] = {
val (prefixDigits, startDigits, endDigits) = separatePrefix(start, end, base, numDigits)
if (start == end) { // Special Case: Range Length 1
Vector(startDigits)
} else if (startDigits.forall(_ == 0) && endDigits.forall(_ == base - 1)) { // Total Optimization
Vector(prefixDigits)
} else if (prefix.length == numDigits - 1) { // Special Case: Front Grouping = Back Grouping
startDigits.last.to(endDigits.last).toVector.map { lastDigit =>
prefixDigits :+ lastDigit
}
} else {
val front = frontGroupings(startDigits, base)
val middle = middleGrouping(startDigits.head, endDigits.head)
val back = backGroupings(endDigits, base)
val groupings = front ++ middle ++ back
groupings.map { digits =>
prefixDigits ++ digits
}
}
}
General Payout Curves
Design
The goal of this specification is to enable general payout curve shapes efficiently and compactly while also ensuring that simpler and more common payout curves (such as a straight line for a forward contract) do not become complex while conforming to the generalized structure.
To this end, we let the set of all supported payout curves be the set of piecewise polynomial functions (with no continuity requirements between pieces). Since lines are polynomials, simple curves remain simple when represented in the language of piecewise polynomial functions. Any interesting (e.g. non-random) payout curve can be closely approximated using a cleverly constructed polynomial interpolation. And lastly, serializing these functions can be done compactly by providing only a few points of each polynomial piece so as to enable the receiving party in the communication to interpolate the polynomials from this minimal amount of information.
It is important to note however, that due to Runge's phenomenon, it will usually be preferable for clients to construct their payout curves using some choice of spline interpolation instead of directly using polynomial interpolation (unless linear approximation is sufficient) where a spline is made up of polynomial pieces so that the resulting interpolation can be written as a piecewise polynomial one.
Curve Serialization
In this section we detail the TLV serialization for a general payout_function.
Version 0 payout_function
- type: ??? (
payout_function_v0) - data:
- [
u16:num_pts] - [
boolean:is_endpoint_1] - [
bigsize:event_outcome_1] - [
bigsize:outcome_payout_1] - ...
- [
boolean:is_endpoint_num_pts] - [
bigsize:event_outcome_num_pts] - [
bigsize:outcome_payout_num_pts]
- [
num_pts is the number of points on the payout curve that will be provided for interpolation.
Each point consists of a boolean and two bigsize integers.
The boolean is called is_endpoint and if this is true, then this point marks the end of a
polynomial piece and the beginning of a new one. If this is false then this is a midpoint
between the previous and next endpoints.
The first integer is called event_outcome and contains the actual number that could be signed by the oracle
(note: not in the serialization used by the oracle) which corresponds to an x-coordinate on the payout curve.
The second integer is called outcome_payout and is set equal to the local party's payout should
event_outcome be signed which corresponds to a y-coordinate on the payout curve.
Note that if you wish to construct the counter-party's payout_function, then this can be accomplished by replacing
all outcome_payouts in your payout_function with total_collateral - outcome_payout and interpolating the
resulting function will yield the same result as defining your counter-party's function to be total_collateral - computed_payout
at every possible event_outcome.
Requirements
num_ptsMUST be at least2.event_outcomeMUST strictly increase. If a discontinuity is desired, a sequentialevent_outcomeshould be used in the second point. This is done to avoid ambiguity about the value at a discontinuity.
General Function Evaluation
There are many ways to compute the unique polynomial determined by some set of interpolation points. I choose to detail Lagrange Interpolation here due to its relative simplicity, but any algorithm should work so long is it does not result in approximations with an error too large so as to fail validation. To name only a few other algorithms, if you are interested in alternatives you may wish to use a Vandermonde matrix or another alternative, the method of Divided Differences.
Given a potential event_outcome compute the outcome_payout as follows:
- Binary search the endpoints by
event_outcome- If found, return
outcome_payout - Else let
pointsbe all of the interpolation points between (inclusive) the previous and next endpoint
- If found, return
- Let
lagrange_line(i, j) = (event_outcome - points(j).event_outcome)/(points(i).event_outcome - points(j).event_outcome) - Let
lagrange(i) := PROD(j = 0, j < points.length && j != i, lagrange_line(i, j)) - Return
SUM(i = 0, i < points.length, points(i).outcome_payout * lagrange(i))
Optimized Evaluation During CET Calculation
There are many optimizations to this piecewise interpolation function that can be made when repeatedly and sequentially evaluating an interpolation as is done during CET calculation.
-
The binary search in step 1 can be avoided when computing for sequential inputs.
-
The value
points(i).outcome_payout / PROD(j = 0, j < points.length && j != i, points(i).event_outcome - points(j).event_outcome)can be cached for eachiin a polynomial piece, call thiscoef_i. For a givenevent_outcomeletall_prod = PROD(i = 0, i < points.length, event_outcome - points(i).event_outcome).The sum can then be computed as
SUM(i = 0, i < points.length, coef_i * all_prod / (event_outcome - points(i).event_outcome)). -
When precision ranges are introduced, derivatives of the polynomial can be used to reduce the number of calculations needed. For example, when dealing with a cubic piece, if you are going left to right and enter a new value modulo precision while the first derivative (slope) is positive and the second derivative (concavity) is negative then you can take the tangent line to the curve at this point and find it's intersection with the next value boundary modulo precision. If the derivatives' signs are the same, then the interval from the current x-coordinate to the x-coordinate of the intersection is guaranteed to all be the same value modulo precision.
Precision Ranges
As detailed in the section on CET compression, any time some continuous interval of the domain results in the same payout value, we can compress the CETs required by that interval to be logarithmic in size compared to using one CET per point on that interval. As such, it can be beneficial to reduce the precision of the payout function to allow for bounded approximation of pieces of the payout curve by constant-valued intervals. For example, if me and my counter-party are both willing to round payout values to the nearest 100 satoshis, we can have significant savings on the number of CETs required to enforce our contract. To this end, we allow parties to negotiate precision ranges which may vary along the curve, allowing more precision near more probably outcomes and allowing more rounding to occur near extremes.
Each party has their own minimum precision_ranges and the precision to be used at a given event_outcome is the minimum of both party's precisions.
If P is the precision to be used for a given event_outcome and the result of function evaluation for that event_outcome is value, then the amount
to be used in the CET output for this party will be the closer of value - (value % P) or value - (value % P) + P, rounding up in the case of a tie.
Precision Range Serialization
- type: ??? (
precision_ranges_v0) - data:
- [
u16:num_precision_ranges] - [
bigsize:begin_range_1] - [
bigsize:precision_1] - ...
- [
bigsize:begin_range_num_precision_ranges] - [
bigsize:precision_num_precision_ranges]
- [
num_precision_ranges is the number of precision ranges specified in this function and can be
zero in which case a precision of 1 is used everywhere.
Each serialized precision range consists of two bigsize integers.
The first integer is called begin_range and refers to the x-coordinate (event_outcome) at which this range begins.
The second integer is called precision and contains the precision modulus to be used in this range.
If begin_range_1 is strictly greater than 0, then the interval between 0 and begin_range_1 has a precision of 1.
Requirements
begin_range_1, if it exists, MUST be non-negative.begin_rangeMUST strictly increase.
Contract Execution Transaction Calculation
Given a payout function, a total_collateral amount and precision ranges, we wish to compute a list of pairs of arrays of
integers (corresponding to digits) and Satoshi values.
Each of these pairs will then be turned into a CET whose adaptor point is computed from the list of integers and whose values
will be equal to the Satoshi value and total_collateral minus that value.
We must first modify the pure function given to us by interpolating points by applying precision ranges, as well as setting all
negative payouts to 0 and all computed payouts above total_collateral to equal total_collateral.
Next, we split the function's domain into two kinds of intervals:
- Intervals in which the modified function's value is constant.
- Intervals in which the modified function's values are changing at every point.
This can be done by evaluating the modified function at every point in the domain and keeping track of whether or not the value has changed to construct the intervals, but this is not a particularly efficient solution. There are countless ways to go about making this process more efficient such as binary searching for function value changes or looking at the unmodified function's derivatives.
Regardless of how these intervals are computed, it is required that the constant-valued intervals be as large as possible. For example if you have two constant-valued intervals in a row with the same value, these must be merged.
Finally, once these intervals have been computed, the CET compression algorithm is run on each constant-valued interval which generates
a list of integers to be paired with the (constant) value for that interval.
For variable-value intervals, a unique CET is constructed for every event_outcome where all digits of that event_outcome are included
in the list of integers and the Satoshi value is equal to the output of the modified function for that event_outcome.
Contract Execution Transaction Signature Validation
To validate the adaptor signatures for CETs given in a dlc_accept or dlc_sign message, do the process above of computing the list of pairs of
lists of digits and payout values to construct the CETs and then run the adaptor_verify funciton.
However, if adaptor_verify results in a failed validation, do not terminate the CET signature process.
Instead, you must look at whether you rounded up (to value - (value % precision) + precision) or down (to value - (value % precision)).
If you rounded up, compute the CET resulting from rounding down or if you rounded down, compute the CET resulting from rounding up.
Call the adaptor_verify function against this new CET and if it passes verification, consider that adaptor signature valid and continue.
This extra step is necessary because there is no way to introduce deterministic floating point computations into this specification without also introducing complexity of magnitude much larger than that of this entire specification. This is because there are no guarantees provided by hardware, operating systems, or programming language compilers that doing the same floating point computation twice will yield the same results. Thus, this specification instead takes the stance that clients must be resilient to off-by-precision (off-by-one * precision) differences between machines.
Authors
Nadav Kohen nadavk25@gmail.com
This work is licensed under a Creative Commons Attribution 4.0 International License.