Certainly any algorithm that's been peer-reviewed multiple times must not have any obvious errors, right?
What about algorithms by leaders in the field, which come with proofs of correctness, and form the basis for tons of later research? Nothing of that stature could be flawed, could it?
But those are purely academic concerns, with no practical impact. It surely couldn't be the case that something as basic as a sorting algorithm, which was implemented multiple times and tested, fails to operate correctly?
In fact, I'll show examples of all three of these. Let's have a conversation about the ways in which algorithms fail--- and the ways to increase confidence that your algorithms and designs are correct.
Mark Gritter is Vault Advisor at HashiCorp, his third startup experience. Previously, Mark co-founded Tintri, an enterprise storage company, in 2008 and helped grow it to a 2017 IPO (and, unfortunately, a 2018 bankruptcy.) Before that he worked at Kealia, a video streaming startup acquired by Sun Microsystems in 2004.
Mark's previous Minnebar presentations have covered topics such as correctness of algorithms, combinatorial auctions, scaling a startup, building a file system, and procedural content generation.