#data-quality #liquidations #open-interest

Why Liquidation and OI Signals Don't Work: The Data Is Broken

Academic research shows exchanges systematically misreport open interest and delay liquidation data. We tested 40 hypotheses on this data and found no edge. Here's why that makes sense.

The Thesis We Wanted to Prove

The idea was compelling: liquidation cascades create forced selling pressure. When leveraged positions get liquidated, it triggers more liquidations. This creates predictable price dislocations that should be tradeable.

Open interest tells you how much leverage is in the system. When OI spikes relative to price, the market is overleveraged. When OI drops sharply, positions are being flushed. These should be signals.

We built a full pipeline to test this: real-time feeds from Binance, HyperLiquid, OKX, and Bybit. 653,000 liquidation events. Months of open interest snapshots. 40 hypotheses tested across 275+ parameter configurations.

The result: zero reliable edge. Not one configuration was profitable out-of-sample.

The Literature Warned Us

After our testing failed, we found academic research that explained why.

Giagkiozis & Said (2024)

"Open interest in Bitcoin perpetual swaps is systematically misquoted by some of the largest derivatives exchanges; however, the degree varies, with some exchanges reporting open interest that is wholly implausible to others that seem to be delaying messages of forced trades, i.e., liquidations."

Source: arXiv:2310.14973, published in Ledger, Volume 9 (2024)

The researchers connected directly to seven major derivatives exchanges and analyzed tick-by-tick data. Their findings:

  • Some exchanges report "wholly implausible" OI numbers. The data doesn't reconcile with traded volume.
  • Others delay liquidation messages. This makes OI appear higher than reality during volatile periods.
  • The degree of misreporting varies widely. Some exchanges are more reliable than others, but none are perfect.

This isn't a fringe finding. The paper was peer-reviewed and published in Ledger, a respected blockchain research journal. The methodology is sound: they cross-referenced OI changes against liquidation events and found systematic discrepancies.

What We Tested

Before finding this research, we ran an exhaustive test of liquidation and OI-based signals:

Liquidation Cascade Signals (H001 series)

Hypothesis Approach Result
H001a-v2 225-config grid search on cascade mean reversion 0/225 profitable
H001d Counter-trend cascade continuation 25% WR (worse than random)
H001e Rare extreme events ($30M+ in 15min) 50% WR, n=6 (inconclusive)

Open Interest Signals (H029-H031)

Hypothesis Approach Result
H029 OI extremes (z-score > 2.0) Marginal edge on ETH only (56% WR)
H030 OI + liquidation combined No improvement over baseline
H031 Low OI extremes Weak, inconsistent

Liquidation Imbalance (H002 series)

Hypothesis Approach Result
H002 Accumulated long/short imbalance 62% WR on 24 trades → collapsed to 42% on 57 trades
H002-v2 Full 90-day validation 0/20 configs profitable

We tested different thresholds, cooldowns, holding periods, and directions. The pattern was the same every time: promising results on small samples that completely collapsed when tested on larger datasets.

Why It Makes Sense

Build signals on broken data and you'll fit to noise instead of finding alpha. Here's why exchange data quality issues kill liquidation/OI signals:

1. Delayed Liquidation Messages Create False Timing

If an exchange delays liquidation reports by even a few seconds during volatile periods, your "real-time" cascade detection is already stale. The price has moved. The opportunity (if it ever existed) is gone.

We found cross-exchange timing varied by 100-500ms. That's an eternity in crypto markets during a cascade.

2. Implausible OI Numbers Mean Positioning Data Is Wrong

OI is supposed to tell you how leveraged the market is. If exchanges report numbers that don't reconcile with actual trades, your read on market positioning is garbage.

When we tested OI extreme signals, the "extreme" levels we detected may not have been extreme at all. They could just be artifacts of reporting inconsistencies.

3. Cross-Exchange Inconsistency Breaks Confirmation

One strategy was to use cross-exchange confirmation: only trade when multiple exchanges show the same signal. But if each exchange has different reporting delays and accuracy, you're not confirming a real signal. You're confirming that two broken datasets momentarily agreed.

4. Small Sample Artifacts

Our most "promising" results came from small samples. H002 showed 62% win rate on 24 trades. When we expanded to 57 trades, it collapsed to 42%. That's worse than a coin flip after fees.

This is what happens when you fit to noise: small samples look great because you're pattern-matching to randomness. Expand the sample and the "pattern" disappears.

The Only Signal That Survived

One finding held up: extreme SHORT liquidation spikes ($30M+ in a 15-minute window) showed 86% favorable price movement at the 4-hour horizon.

But there were only 7 such events in 90 days. That's not a tradeable strategy. It's a curiosity. And even this may be explained by the extreme magnitude cutting through the noise of bad data reporting.

Implications

If you're building trading signals on exchange-reported liquidation or OI data, you should be aware:

  1. The data is not ground truth. Academic research confirms what our testing suggested. Exchange data has systematic issues.
  2. Small sample results are unreliable. We saw multiple 60%+ win rates collapse to below 40% when tested on larger samples.
  3. Cross-exchange "confirmation" may be meaningless. If both sources are unreliable, agreement proves nothing.
  4. Timing-sensitive strategies are especially vulnerable. Delayed liquidation messages make real-time cascade trading nearly impossible.

What Actually Works

After 40 failed hypotheses on liquidation/OI data, we pivoted to signals that don't rely on exchange-reported derivative metrics:

  • Funding rate extremes. Still exchange-reported, but mathematically constrained. The funding rate affects P&L directly, so it can't be arbitrarily fake.
  • Cross-sectional momentum. Uses only price data, which is harder to systematically misreport.
  • On-chain data. Blockchain data is immutable and verifiable.

The lesson: when building quantitative signals, start by asking whether your data source can be trusted. In crypto derivatives, the answer is often "less than you think."

Source: Our testing covered 653K BTC liquidations across Binance, HyperLiquid, OKX, and Bybit (Nov 2025 - Jan 2026), plus OI snapshots at 5-minute intervals. Academic reference: Giagkiozis & Said (2024), "Reconciling Open Interest with Traded Volume in Perpetual Swaps", Ledger, Volume 9.

Get research like this in your inbox