The idea is that if the coin flip goes in the player’s favor, they win double their bet. After winning, they can either collect their winnings, or risk them all on another coin flip to have a chance at doubling them. The initial bet is fixed at, let’s say $1.

Mathematically, this seems like a fair game. The expected value of each individual round is zero for both house and player.

Intuitively, though, I can’t shake the notion that the player will tend to keep flipping until they lose. In theory, it isn’t the wrong decision to keep flipping since the expected value of the flip doesn’t change, but it feels like it is.

Any insight?

  • Ada@lemmy.blahaj.zone
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    Once the player loses, the chain ends, and the house wins. So as long as the house can afford to keep pushing the player in to trying again, they’re going to create more opportunities for the player to return their winnings to the house.

    • HandwovenConsensus@lemm.eeOP
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      Right, and as the chain continues, the probability of the player maintaining their streak becomes infinitesimal. But the potential payout scales at the same rate.

      If the player goes for 3 rounds, they only have a 1/8 chance of winning… but they’ll get 8 times their initial bet. So it’s technically a fair game, right?

      • rowinxavier@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        2 months ago

        If everyone has the same amount of starting capital it is a fair game assuming both can opt out at any time.

        That said, the house appears to not be able to opt out (they definitely can, you just don’t think about that part), and the house has more capital. For them each time someone plays a round there are only 3 possible outcomes. Half are the player loses, then a quarter are the player wins and plays another round, and lastly a quarter are the player wins and ends the game. The only case where the player wins is option 3, in all other cases, so 75%, the house wins because the next round has another chance to make the player lose directly at a 50/50 chance or play another round.

  • solrize@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    If both players have infinite bankrolls, but only one of them is allowed to stop the game once they are ahead, the one with the option of stopping has an advantage. They can play until they are in the lead, then stop. The reason this doesn’t work in real life is that real bankrolls aren’t infinite.

    See also: https://en.wikipedia.org/wiki/Gambler's_ruin

  • General_Effort@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    That looks like the St. Petersburg Paradox. Much ink has been spilled over it.

    The expected payout is infinite. At any point, the “rational” (profit-maximizing) decision is to keep flipping, since you wager a finite sum of money to win an infinite sum. It’s very counter-intuitive, hence called a paradox.

    In reality, a casino has finite money. You can work out how many coin flips it takes to bankrupt it. So you can work out how likely it is to reach that point with a given, finite sum of money. Martingale strategies have already been mentioned.

    • HandwovenConsensus@lemm.eeOP
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      Not quite the same, since in my scenario the player loses everything after a loss while in the St. Petersburg Paradox it seems they keep their winnings. But it does seem relevant in explaining that expected value isn’t everything.

  • commie@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    0
    ·
    2 months ago

    the house can only make $1 per play, and the bettor can make a functionally unlimited amount.

    see the martingale strategy. you are basically sticking the house with a martingale strategy in which you get to decide when they get.

  • OsrsNeedsF2P@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    2 months ago

    If you have 100$, and you bet 1$ at a time, infinitely, you will lose.

    More generally (simplified to assume you’re always betting the same amount):

    P(ruin after X bets) = (edit: I removed my formula because it was wrong…but I’m sure you could mathematically prove a formula)

    • HandwovenConsensus@lemm.eeOP
      link
      fedilink
      arrow-up
      0
      ·
      2 months ago

      You’re saying that the player pays a dollar each time they decide to “double-or-nothing”? I was thinking they’d only be risking the dollar they bet to start the game.

      That change in the ruleset would definitely tilt the odds in the house’s favor.