Should we let ourselves see the future?

This is a cross-post from my new substack, Calibrations. Cryptography professor and solid Twitter follow Matthew Green asks why we should really care what prediction markets say:

It would seem the perfect first post for a blog titled Calibrations. So are prediction markets any good at predicting things, and should we care what they say?

Prediction markets allow participants to buy shares of an event occurring or not, analogous to a bet or wager. For example, the website Polymarket has political event markets. A “share” pays out if the election occurs as you predict and pays nothing if you predict incorrectly. A market is created by people buying and selling shares on the outcomes of the events, and the price of the share represents the market’s current probability estimate for the outcome. For example, you can buy a share of Kamala Harris winning the presidential election which will pay out if she wins, but will be worth nothing if she loses. A price of 60 cents would indicate a 60% chance of her winning.

Data on Accuracy

So are prediction markets any good? Well, we can actually look at the history of prediction markets and see how often a market that predicts any given outcome actually ends up resolving in that way. Maxim Lott at electionbettingodds.com has a track record page for the prediction markets he tracks.

This is the calibration of the markets. If something that is predicted to happen 60% of the time actually happens that often, we say it’s well calibrated (and thus the name of this blog). We can also see the calibration for some other prediction sites here from calibration.city which has some excellent visualizations:

I suspect some of the systematic bias you see in the chart above and below the 50% mark is from the fact that many of these markets default to 50% and require betters to provide liquidity to move the market away from the start. To make an enticing market to bet in, the market makers have to incorrectly price a market so that bettors have an incentive to wager and earn their winnings, and we see this in the data as a the market makers systematically “betting” the wrong way to start the market. In fact, if you go to the calibration.city page and weight the y-axis for resolutions by market volume, this bias is reduced (since larger volume markets would have less weight on the initial price) although it does not disappear.

Alright, prediction markets are well-calibrated, but does that mean they are accurate? Are they actually predicting anything? Not necessarily!

Suppose we know that Republicans win the presidency about half the time and Democrats win the other half. We check the prediction markets and they always say democrats have a 50% chance to win. This is a well calibrated prediction, but it doesn’t actually tell us much information at any given point in time. We want to know about this particular upcoming election. What we can do then is use a Brier score to mathematically measure accuracy, and not just calibration. The Brier score is the means squared error for a set of predictions and their outcomes, and thus we can use it to measure the accuracy of probabilistic predictions. Here are the Brier scores from calibration.city:

I arranged by market volume and did a time weighted average, but you can also try it yourself. Overall it seems market volume didn’t have much effect, perhaps because the existing Brier scores are already pretty low around 0.15, and also perhaps because higher volume markets could reflect exciting changes in the underlying event leading to both more uncertainty and more market trading.

We can actually decompose the Brier score into calibration, resolution, and uncertainty components, although exactly why this works is beyond my current statistical understanding. This stack exchange post was helpful, but unfortunately I don’t have the time to put together a full script to grab the data from Manifold and run the decomposition.

There’s also real limitations to comparing Brier scores. You’re really only supposed to compare them across the same events; if the underlying event is more uncertain, you would expect higher (worse) Brier scores. But Kalshi, Manifold, and Polymarket are all predicting slightly different things. The legal difficulties prediction market platforms face exacerbates the accuracy measurement problem; PredictIt has to limit the number and size of bets, Polymarket and Betfair are banned in the United States, and Manifold only uses play money. Still, you’d expect much of this to contribute to worse Brier scores. The fact that these are pretty solid is encouraging. If anyone wants to run an actual Murphy decomposition on prediction market data on a narrow topic, I would be interested to see the results.

I think we can still draw some conclusions though:

  • Prediction markets are well calibrated: if they say something will happen with a given percentage chance, that’s actually a good estimate of how likely it will happen.
  • Prediction markets have pretty good (low) Brier scores, and that puts an lower limit on how practically bad their resolution decomposition component could be even if they were perfectly calibrated.
  • Prediction markets do all this despite major regulatory barriers.

The Bigger Picture

Taking a step back though, I think Matthew Green’s critique gets to something deeper: sure prediction markets are well calibrated in the face of uncertainty, but they don’t have any magical insight besides pushing the best guesses to the top. In other words, they’re not actually reducing the fundamental uncertainty in the world. Why is everyone so excited about this?

I think the answer is price discovery.

First let’s take the demand side. Perhaps you’ve looked up a prediction market price for an event and found the price to be…about what you expected.

The Fed has been talking about rate cuts this year, they indicated in their last meeting that it’s likely coming and the market took a sharp tumble last week before recovering. It seems pretty obvious to anyone paying attention that the likelihood of a Fed rate cut by the end of the year is high. But not everyone was paying attention. A prediction market’s existence creates a publicly known “price” of the best estimate of whether an event will occur according to the market participants. Being able to know a well calibrated estimate of an uncertain event is a huge positive externality that prediction markets provide the whole world.

This is a big deal! If you don’t have a price to check, you’re left doing all the research yourself and you only have so much time. Worse, you might have to rely on very non-quantified pundit opinions with no track record. A price cuts through everything to give you a single best guess.

Next let’s take the supply side. Think about the amount of time, effort, resources, money, labor, algorithm design, etcetera that trading firms undertake to get an edge in financial trading. There are potentially billions of dollars on the line if you can correctly predict whether an equity or future will rise or fall in price. Prediction markets can harness these incentives to uncover truth not just about company financial data, but about anything we are curious about. They tie reality and truth about the world to financial incentives. They push us to discover information about what is happening and what will happen.

Unfortunately, they are also legally limited. Event based betting is for all practical purposes banned in the United States even though political decision making is very high stakes. Our lives are worse because of this! We should want our political decision-making processes scrutinized and better understood; prediction markets provide financial incentives to achieve those ends. Prediction markets could be providing us with real time information on whether Congress will pass legislation or how an administration will respond to crises.

Of course, the actual application of prediction markets could be much broader than politics. Nominal GDP futures markets could do a better job informing the Fed of forward looking expectations than stock markets or lagging economic data. Or imagine if we had an ongoing prediction market about possible pandemics on the horizon. That would have been very handy to refer to in January or February 2020 and may have alerted us sooner than simply relying on broad media reporting, which didn’t pick up the potential danger of the pandemic until early March.

Perhaps today prediction markets don’t tell us much besides aggregating some polling data, but if we legalized them, maybe they could tell us a lot more!

Uncharted Territory

The finance industry may be driven purely by profit, but the development of new financial instruments has actually brought all sorts of benefits, some of them very widely distributed. Index funds and ETFs allow regular people with savings access to market returns without needing to expose themselves to specific stock risk and at a fraction of the cost of mutual funds. Foreign exchange and commodity derivatives can help companies reduce their risk by locking in prices now; airlines are trying to run an airline business, not predict oil prices in six months.

Prediction markets could do even more if we let them.

A conditional prediction market is a single trading market with two events allowing investors to trade on the relationship between the two outcomes. For example, you could make a prediction market for a political parties’ election prospects conditional on whether they replace their candidate or not. Democrats took months to come around to the idea that their current presidential candidate might do poorly. If they had been able to compare conditional bets on who would win, they might have been able to select a different candidate much earlier!

Conditional markets could have all sorts of interesting applications. We could answer questions like “what will different policies’ impact be on unemployment next year?” or “which scientific research approach is most likely to improve 5 year survival rates of a major cancer?”. But I think the most important point to underline is that we don’t know because the markets aren’t allowed. It took a while to develop ETFs that regular investors could use for their retirement accounts. If we allowed people to experiment and try things, we could probably create some pretty cool things.

Even today, with all the legal challenges prediction markets face, they offer great value! People on Twitter will continue to claim that prediction markets are systemically biased or simply reflecting quirky beliefs of people overrepresented in the betting pool. I think this is too quick of a dismissal, but to some extent, they’re right! And I think it’s because you have to jump through a bunch of hoops to correct the markets and collect your winnings; as long as the CFTC wages a crusade against PMs, it will be challenging to make much money on them and thus they may often have incorrect prices. But we don’t have to live this way! We could look into the future if we allowed ourselves to.

Links 2018-07-09

My new series focusing on policy summaries made me realize that while the political world and Twittersphere may not discuss policy much, there are groups of people who research policy professionally and have probably covered some of what I want to do with my “Policies in 500 Words or Less” series.  So after looking around, I found that the Cato Institute has an excellent page called the Cato Handbook for Policymakers. It contains a ridiculous 80 entries of policy discussions including a top agenda of important items, a focus on legal and government reforms, fiscal, health, entitlement, regulation, and foreign policies. I will definitely be pulling some ideas from that page for future policy summaries.

I recently found the YouTube channel of Isaac Arthur, who makes high quality, well researched, and lengthy videos on futurism topics, including space exploration. I’d like to take a moment to highlight the benefits of a free and decentralized market in the internet age. Adam Smith’s division of labor is incredibly specialized with the extent of our market. Arthur has a successful Patreon with weekly videos on bizarre and niche topics that regularly get hundreds of thousands of views (24 million total for his channel), and they are available completely free, no studio backing necessary. Such an informative career could not have existed even 10 years ago.

The 80000 Hours Podcast, which was recently mentioned in our top podcasts post, had Dr. Anders Sandberg on (broken into two episodes) to discuss a variety of related topics: existential risk, solutions to the Fermi Paradox, and how to colonize the galaxy. Sandberg is a very interesting person and I found the discussion enlightening, even if it didn’t focus much on how to change your career to have large impacts, like 80000 Hours usually does.

Reason magazine’s July issue is titled “Burn After Reading”. It contains various discussions and instructional articles on how to do things that are on the border between legal and illegal, such as how to build a handgun or how to make good pot brownies or how to hack your own DNA with CRISPR kits. It’s an impressive demonstration of the power of free speech, but also important to the cyberpunk ideal that information is powerful and can’t be contained.

George Will writes in support of Bill Weld’s apparent aim to become the 2020 Libertarian Party nominee. I admit I wasn’t hugely impressed with Weld’s libertarian bona fide’s when he was running in 2016, but I thought his campaigning and demeanor was easily better than Gary Johnson’s, who was already the LP’s best candidate in years, maybe ever. I think a better libertarian basis paired with Weld’s political skills would be an excellent presidential candidate for the LP.

Related: last week was the 2018 Libertarian Party National Convention. I don’t know if it’s worth discussing or whether it’s actually going to matter, but I have seen some good coverage from Matt Welch at Reason and Shawn Levasseur.

I read this very long piece by Democratic Senator (and likely Presidential hopeful) Cory Booker at Brookings. It was a pretty sad look at current issues of employment, worker treatment, and stagnant wages. There was a compelling case that firms are getting better at figuring out ways to force labor to compete through sub-contracting out labor to avoid paying employee benefits. This leads to monopsony labor purchasing by large firms, squeezing workers who don’t have the same amount of market bargaining power. He also mentions non-compete clauses and growing differences between CEO pay and average pay for workers. I don’t have good answers to these points, although his suggestion of a federal jobs guarantee seems very expensive and likely wasteful. His proposed rules about stock buybacks also seem to miss the point. Maybe stricter reviews of mergers would work, but perhaps larger firms are more efficient in today’s high tech economy, it’s hard to know. Definitely a solid piece from a source I disagree with, which is always valuable.

Somewhat related: Scott Alexander’s post from a couple months ago on why a jobs guarantee isn’t that great, especially compared to a basic income guarantee. Also worth reading, Scott’s fictional post on the Gattaca sequels.

Uber might have suspended testing of self driving automobiles, but Waymo is going full steam ahead. They recently ordered over 80,000 new cars to outfit with their autonomous driving equipment, in preparation for rolling out a taxi service in Phoenix. Timothy B. Lee at Ars Technica has a very interesting piece, arguing the setbacks for autonomous vehicles only exist if you ignore the strides Waymo has made.

Augur, a decentralized prediction market platform similar to Paul Sztorc’s Hivemind (which I’ve discussed before), is launching on the Ethereum mainnet today. Ethereum has its own scaling problems, although I’d hope at some point sharding will actually be a real thing. But for now, transactions on Augur may be pretty expensive, and complex prediction markets may remain illiquid. That may mean the only competitive advantage Augur will offer is the ability to create markets of questionable legality.  Exactly what that will be remains to be seen, but this is an exciting development in the continuing development of prediction markets.

 

A Few Thoughts on Bitcoin

I have been aware of Bitcoin’s existence for a while, and while I was excited about it a few years ago, it had somewhat dropped off my radar. Perhaps because over the past few months, Bitcoin has seen a big increase in value, I started to revisit it and analyze it as a technology. My experience has been nothing short of breathtaking.

A few years ago, Bitcoin was pretty cool. I even wrote a paper about it, discussing the huge potential of the technology and decentralized, autonomous transactions could totally upend the banking industry. But back when I first got into Bitcoin, I was also interested in Austrian Economics, which I’m largely over now. Their focus on control of the money supply and dire warnings about the Federal Reserve weren’t really borne out by the rather mundane economic growth of the last few years.

Nonetheless, the Bitcoin community has been working on without me, and it has paid off: you can now use Bitcoin to purchase from all sorts of retailers, including Dell, Overstock.com, Newegg, and more. You can also buy all sorts of internet specific services, which to me seems like the clearest use case. These include Steam credit, VPNs, cloud hosting, and even Reddit gold.

The price has jumped up to over $1000 at the end of April 2017 (that’s over $18 billion in total market value of all Bitcoins), and it was briefly even higher a month ago on speculation the SEC would allow for a Bitcoin ETF. The ETF was rejected, but the potential of the currency remains. And technologically, Bitcoin is far more impressive than it was, most notably with a concept called the Lightning Network.

This technology would allow for instantaneous Bitcoin transactions (without having to accept risky zero confirmation transactions). These transactions would have the full security of the Bitcoin network, and would also likely allow massive scaling of the Bitcoin payment network. Drivechain is another project with great potential to scale Bitcoin and allow for applications to be built on top of the Bitcoin blockchain. It would create a two-way peg, enforced by miners, that allowed tokens to be converted from Bitcoin to other sidechains and back again. This would allow experimentation of tons of new applications without risk to the original Bitcoin blockchain.

Hivemind is particularly exciting as a decentralized prediction market that is not subject to a central group creating markets; anyone can create and market and rely on a consensus algorithm to declare outcomes. If attached to the Bitcoin blockchain, it also wouldn’t suffer from cannibalization that Ethereum blockchains like Augur can suffer from.

Mimblewimble is another interesting sidechain idea. It combines concepts of confidential transactions with (I think) homomorphic encryption to allow for completely unknowable transaction amounts and untraceable transaction histories. It would also do this while keeping the required data to run the blockchain fairly low (the Bitcoin blockchain grows over time). It would have to be implemented as a sidechain, but any transactions that occur there would be completely untraceable.

And there are even more cool projects: Namecoin, JoinMarket, the Elements Project, and of course other cryptocurrencies like Ethereum, Monero, and Zcash. This really makes the future of Bitcoin and cryptocurrencies seem pretty bright.

However, we’ve skipped a big point, which is that most of these cool innovations for Bitcoin can’t be done with Bitcoin’s present architecture. Moreover, the current number of Bitcoin transactions per block has just about maxed out at ~1800. This has resulted in something called the Scaling Debate, which centers about the best way to scale the Bitcoin blockchain. Upgrades to the blockchain must be done through consensus where miners mine new types of blocks, institutions running nodes approve of those new blocks, and users continue to create transactions that are included in new blocks (or else find another cryptocurrency). Before any of that can happen, developers have to write the code that miners, validation nodes, and users will run.

Right now, there is a big political fight that could very briefly be described as between users who support the most common implementation of the Bitcoin wallet and node (known as Bitcoin Core) and those who generally oppose that implementation and the loose group of developers behind it. I certainly am not here to take sides, and in fact it would probably have some long term benefits if both groups could go their separate ways and have the market decide which blockchain consensus rules are better. However, there is not much incentive to do that, as there are network effects in Bitcoin and any chain split would reduce the value of the entire ecosystem. The network effects would likely mean any two-chain system would quickly collapse to one chain or the other. No one wants to be on the losing side, yet no side can convince the other, and so there has been political infighting and digging in, resulting in the current stalemate.

There will eventually be a conclusion to this stalemate; there is too much money on the line to avoid it. Either the sides will figure out a compromise, the users or the miners will trigger a fork of the chain in some way and force the issue, or eventually a couple years down the road another cryptocurrency will overtake Bitcoin as the most prominent store of value and widely used blockchain. A compromise would obviously be the least costly, a chain split would be more expensive, but could possibly solve the disagreement more completely than a compromise, while another cryptocurrency winning would be by far the most expensive and damaging outcome. All development and code security that went into Bitcoin would have to be redone on any new crytocurrency. Nonetheless, Litecoin just this week seems to have approved of Segregated Witness, the code piece that is currently causing the Bitcoin stalemate. If Bitcoin’s stalemate continues for years, Litecoin is going to start looking pretty great.

Obviously it’s disappointing that even a system built on trustless transactions can’t avoid the pettiness of human politics, but it’s a good case study demonstrating just how pervasive and pernicious human political fights are. Ultimately, because cryptocurrencies are built in a competitive market, politics cannot derail this technology forever. And when the technology does win out, the impact on the word will be revolutionary. I just hope it’s sooner rather than later.

 


Bitcoin featured picture is a public domain image.

Leave a comment on the official reddit thread.