[HN Gopher] The McNamara fallacy: Measurement is not understanding ___________________________________________________________________ The McNamara fallacy: Measurement is not understanding Author : wenc Score : 174 points Date : 2022-02-01 18:02 UTC (4 hours ago) (HTM) web link (mcnamarafallacy.com) (TXT) w3m dump (mcnamarafallacy.com) | [deleted] | serverlessmom wrote: | "In war, whichever side may call itself the victor, there are no | winners, but all are losers,"- Neville Chamberlain. | | This is a really interesting look at this type of mindset. I have | often wondered what it is to "win" a war, and the quote that came | to mind was the one I posted above. Measurement is certainly not | understanding in all situations. | sandworm101 wrote: | What is the name for the fallacy of thinking that wars are always | meant to be won? Throughout history wars have been fought without | any intention of "winning". Sometimes a war can serve a religious | or ceremonial purpose, one that doesn't require a clear winner. | Other wars have been fought for completely economic reasons. | Others are proxy contests whereby greater powers can demonstrate | their abilities without directly engaging each other. The false | thinking is the assumption that participants always want or even | care about winning. | missedthecue wrote: | Wars might have unclear objectives and mission-creep, but I | don't think anyone is fighting to lose. | serial_dev wrote: | The question is about what "winning" means for the people who | start wars and keep fighting them. It might not be the same | as for you. | | To quote Julian Assange, "The goal is to use Afghanistan to | wash money out of the tax bases of the US and Europe through | Afghanistan and back into the hands of a transnational | security elite. The goal is an endless war, not a successful | war". | missedthecue wrote: | I just don't buy that line of thinking at all. If that is | truly their goal, there are way easier and subtle ways to | do it without causing a collapse in political capital for | the parties and politicians involved, which the Vietnam, | Iraq, and Afghanistan wars did. | [deleted] | eumoria wrote: | McNamara basically admits this himself in the documentary The Fog | of War: | | https://en.wikipedia.org/wiki/The_Fog_of_War | | It's worth a watch but it's very soft on him and his role. Still | a very good documentary. | ProAm wrote: | Great documentary. I felt it was a person trying to come clean | and ease his conscious on his death bed. | 2OEH8eoCRo0 wrote: | Agreed. Great documentary. The part where Castro said he | urged Russia to launch their nukes from Cuba to the US | knowing it would destroy Cuba was chilling. Humans are not | always logical. Don't assume somebody wont drag an entire | country or the world to total destruction for some deranged | cause. | | I didn't get the death bed vibe from McNamara but I | definitely felt that he was genuinely reflecting on the past. | | The documentary on Rumsfeld was the polar opposite. I could | also see Rumsfeld not wanting to give the enemy of an ongoing | conflict any shred of material. It makes for a less | interesting documentary. | nickdothutton wrote: | Morris himself said something like he didnt really feel he | got to know Rumsfeld and had no real idea what was in his | mind compared to his previous interview with McNamara. | wayoutthere wrote: | Guess Donald himself was one of the "unknown unknowns" | smaug7 wrote: | Was looking for this comment. In McNamara's reflection, a | tenant he called out was to understand the enemy. US didn't | understand the Viet Congs motivation for fighting the war | (freedom from colonizers) whereas the US viewed the war as a | larger Cold War. This is the same as what happened in | Afghanistan that we didn't learn. | calyth2018 wrote: | I'd argue the US has not tried to understand the enemy since | the Cuban missile crisis. There are more failures outside of | Afghanistan, and I think the US is going to walk right into | another one. | calyth2018 wrote: | I haven't live in that era, so maybe it's not my place for me | to say whether Morris was particularly soft on him. I think the | fact that he said that it was the president's responsibility, | that revealed a lot about him. | | On the other hand, given the Fog of War did play back a | recording where Johnson had a much different view on Vietnam | than JFK, I wouldn't put the burden solely on McNamara's | shoulders either. | | Regardless of what one might think of his role, it was still | quite enlightening, and I think more people should watch it. I | think the lessons outlined in it are useful, but too few have | taken heed of it. | jkingsbery wrote: | I haven't seen that one, but I've been watching the Ken Burns | documentary recently. It seems suitably fair. Where maybe some | of the proposed "McNamara Fallacy" breaks down from OP, | according to archives that they go through in the documentary | he knew his approach wasn't working for a long time, he just | did not (or would not, or could not, depending on your | perspective) say so publicly and didn't seem to have any other | way to measure progress. | johnp271 wrote: | Does the McNamara Fallacy have any application to our response to | COVID? I often hear pundits of all sorts, medical doctors, | epidemiologists, politicians, CDC scientists, etc, make | statements such as "the data shows this" and "the data says that" | and then follow up with "therefore the science says we must all | do such-and-such". I hold a rather narrow, rigorous - maybe | closed minded - opinion of what is 'science' (so to me 'social | science' is an oxymoron) thus I have a degree of skepticism when | data analysis is relied on to heavily for making conclusions that | are then called 'scientific'. | p_l wrote: | More like the data about COVID is used as fig leaf for metrics | in other areas actually driving the decisions, or | ideology/dogma. | | If you go with hard data and experience, you'd do hard moves | like China (and many other asian countries did). In fact, | similar moves have been done in the past in Europe (on the | communist side of Iron Curtain) to stop epidemics, including | even manual contact tracing. | | But because a non-trivial force in decision making has strong | other incentives, and because of dogma like disbelief in | aerosol transmission, we end up with really bad decisions with | fig leaf of data analysis. | calyth2018 wrote: | > because of dogma like disbelief in aerosol transmission | | It's not just that COVID can spread via aerosol. It's more so | that the western world extrapolated the definition via a | study on TB, neglecting that TB needed to infect deep in the | lungs. | | https://www.wired.com/story/the-teeny-tiny-scientific- | screwu... | | There is a scientific paper version of this, btw. | jscode wrote: | Quick story: I was the CFO for a company that sold to a private | equity group (PEG). I took over as the CEO as the founders | retired, leaving me to deal with the PEG. It quickly became | apparent that the PEG managers looked at everything through the | lens of an Excel spreadsheet. These guys were brilliant attorneys | and analysts but lacked experience building businesses and | managing teams. Ultimately, they couldn't add much value in terms | of operations or strategy, but they were great at financial | modeling/quantitative analysis and forcing us to justify | expenses. That may sound good at first--eliminating wasteful | spending--but it ultimately led to the gradual erosion of the | company culture and employee loyalty. It's easy to cut benefits | and pay given that many workers lack the leverage to do anything | about it, while it's much harder to reduce hard costs like | materials and equipment. That meant employees just kept getting | squeezed, and it was surprisingly difficult to quantify the | impact that terminating an employee or cutting benefits would | have on morale/culture/performance. | | The moral of the story is that people with analyst mindsets play | an essential role in our economy, but sometimes giving those | people power over large organizations can have disastrous | consequences. There truly is a disconnect between measurement and | understanding. | boringg wrote: | While your experience sounds painful it also sounds like | something that would have happened in the 90s. I don't think | most tier 1 organizations still think in that way. | apohn wrote: | Less than a decade ago I worked at a company that was | acquired by a private equity group. What jscode said matches | my experience. For a while I also tracked (on Glassdoor and | some other sites) companies that were purchased by that firm, | and seems like employees at different companies had the same | experience. EBITA was king, nothing else matters. | bb88 wrote: | The company exists for the stockholders, not for the | employees. | | Stack ranking is still used widely throughout Fortune 500 | companies, which is one of the most culture destroying | management practices known to man. | milesvp wrote: | I worked at a company that started to go through the "you can't | improve what you don't measure" phase. In general it was good | for the org I was in, but I used to have to remind management | that there's a corollary to that saying, which is: you | necessarily improve things you measure at the expense of the | things which are difficult or impossible to measure. | | This seems to be a hard one for some types to truly grok. A | common response is that we need to figure out how to measure | it, thinking there was some single magic number that things | could be distilled down to. But often even if you figured out | how to measure some of them, there's always other intangibles | you're not tracking. So you need to always be conscious of it. | hn_version_0023 wrote: | I'm not the type to pass up making a Star Wars reference, so | here goes: | | "One would think you Jedi would understand the difference | between _knowledge_ and... heh heh... _wisdom_ " | apohn wrote: | I worked at a company that went through a private equity | acquisition and I have a question you might be able to answer. | | If you exclude layoffs and incentivized retirement, it seemed | to be that a greater percentage of individual contributors left | as compared to managers. Lots of managers stuck around for 12+ | months, and it seemed like the percentage of managers of | managers (e.g. directors, VPs) who stuck around was even | greater. Almost the entire C-suite stayed for years after the | acquision. | | Was there any financial or other incentives given to managers | to stay? As an individual contributor, the morale was just | terrible. I just couldn't understand why the managers and other | people in leadership positions stuck around. | jscode wrote: | > I just couldn't understand why the managers and other | people in leadership positions stuck around. | | Money. Investors typically carve out equity to retain key | personnel (i.e., management units/stock). The units are | worthless unless the company appreciates in value, so | management becomes laser-focused on doing whatever it takes | to increase the company's valuation. Everything else becomes | a secondary concern. | mrxd wrote: | Just to play devil's advocate, surely their approach is more | rational than that. They're probably looking at it from the | perspective that the business needs to have a profit margin of | X in order to justify investing in it. | | They probably do understand that cutting costs impacts company | culture and morale. But shutting the company down probably | impacts that much more. | jscode wrote: | They do understand that cutting costs will have an impact on | culture and morale, they just think the marginal benefit | exceeds the marginal cost. Keep in mind, PEG managers are | chasing a carried interest bonus which they only achieve | after covering the minimum return promised to their | investors. Plus, leveraged buyouts--which PEGs frequently use | --increase a company's risk of failure. Everyone's under | intense pressure to perform. | | Massive Financial Incentives + Highly Leveraged Balance Sheet | + Intense Pressure = Risky Decision Making | mcguire wrote: | " _But shutting the company down probably impacts that much | more._ " | | Is that the _only_ other option? | jkingsbery wrote: | "Frupidity" is a term I've heard used for this. | Buttons840 wrote: | Interesting observation. The PEG managers would probably admit | that turnover had some financial cost, but I doubt they ever | actually put a number on it and added it to their calculations. | Am I right? | musicale wrote: | > private equity group | | The goal of a private equity group can sometimes be be to | extract as much money from the company as possible in a given | time frame, rather than to ensure the long-term success or | survival of the company. | giva wrote: | And the goal of PEG's empolyees is to get bonuses by hitting | the metrics set for them. | | Metrics are the only goal that means there. It's all about | making the numbers look pretty. | peteradio wrote: | I worked at a place with a lean 6 sigma certified specialist | who towards the end of the companies doom effectively had the | lead engineer cleaning out molding machines to track down every | last tiny molded part that over the course of several years of | continuous running had flung outside of its target. Same guy | told me if the coke machine ever stole my change that he'd help | me get it back from the vendor. | 7thaccount wrote: | All the lean sigma stuff seems like another useless | management fad to me that only benefits consultants. Is that | what you're saying here? | Spooky23 wrote: | It's an expression of distrust. | | If you ever work in government, procurement people think | like that because their goal is objective competitive | process that that meets the minimum standard to fulfill the | purpose. | | There's a certain logic to it. You don't want to see random | government employees driving around in Teslas, so generally | speaking they will be in nondescript 4-door sedans. Having | a human say "no" makes them accountable, so a complex | process will determine what kind of car you need. | | Taken to extreme, it becomes a problem. Procurement | officers get lazy and focus on their process instead of the | needs of the customer. So they treat humans like Ford | Tauruses and allow vendors who understand how to game the | process walk out the door with millions. | peteradio wrote: | I'm only speaking towards this one particularly useless | buffoon, but the fact that he was allowed to wield any sort | of power over anyone says something. | thereddaikon wrote: | Like many management system fads, it started as a useful | kernel of wisdom or obvious maxim that idiots ran with and | turned into a monster. In the case of six sigma, the idea | is about constantly optimizing your workflows and not | accepting "we've always done it this way" as an excuse. | | But most people lack the critical thinking skills to | correctly apply wisdom when necessary and instead need a | solid framework to operate within. That's how these things | inevitably develop. Just like how Agile is supposed to be | about getting working code over being bogged down in | process but inevitably ends up with with half baked | products that have massive issues. | | Six Sigma also gets applied to industries it has no | business being in. The mentality works best when you have a | fixed workflow. In manufacturing it would be if you are | making a lot of one thing. You can do a lot of optimizing. | But I have seen it employed in organizations where every | project was vastly different. Instead of a lean approach it | should have, and previously was, following a house of | quality philosophy. This happened to be in an industry | where cost was rarely a consideration but performance and | reliability were. | 7thaccount wrote: | Great comment and thank you. It has only popped up | sparingly in my industry thankfully. | | What you're saying makes sense to me in that you should | never fall into the "this is how I've always done it" | trap, but putting a complex beauracratic process around | that is just going to create a whole new problem. | jancsika wrote: | > The moral of the story is that people with analyst mindsets | play an essential role in our economy, but sometimes giving | those people power over large organizations can have disastrous | consequences. | | I'm gonna cosplay an "analyst mindset": | | 1. Need to measure costs and benefits of slashing benefits/pay. | | 2. A benefit-- slashing benefits/pay allows us to hit some | obvious financial goal | | 3. A cost-- Uh oh, I don't yet know how to reliably measure | _any_ of the costs. | | 4. Good analysts don't take action without measuring. | | 5. I'm a good analyst. | | Conclusion: I cannot take the action of slashing benefits/pay | | The only way to make it work is to add a step "3b: cherry pick | metrics for the costs of slashing benefits/pay such that the | phony metrics justify the decision management already wants to | make of slashing benefits/pay." But now we've shifted from | "analyst mentality" to "the mindset of the little Beetle-like | bureaucrats described by George Orwell in 1984." | jacobr1 wrote: | Having dealt with 2 PE exists, rarely is the proposal | something as upfront and silly as slash everyones pay. That | probably does happen for a company being restructured in the | red, but the more subtle actions tend to be things like: | | * Comp bands for are now targeting p50 averages rather than | p75 or top of market. So you can't close new hires that are | going competitors. And you can't give raises to your top | performers | | * The health benefits are less generous when renegotiated for | the following year | | * T&E that would have been approved - granted some maybe that | shouldn't - but importantly some that should have for top- | sales people, are no longer approvable. So your top sales | people leave. Or similarly the accelerators or other measures | are changed, that might look good on paper but rub top sales | people the wrong way. | | * Head-count isn't replaced, so teams have to take on more | work | | * Perks like conference attendance or hardware upgrades, | which arguable aren't perks but investments in your team's | productivity, are cut/limited | bb88 wrote: | Cash is still King. The more cash on hand, the easier it is | for the business to survive in an economic downturn, and | the more dividends and stock buybacks can happen for the | investors. | | Software engineers from a CFO perspective aren't any really | different than plumbers or carpenters. It's just labor. | Getting a cheaper rate on labor is far more beneficial to | the company than say making sure it's employees are happy. | | If the company could cut wages 50% across the board and | then give the executives 20% raises for saving 50% in | labor, they would do it in a heartbeat. | Aeolun wrote: | > Getting a cheaper rate on labor is far more beneficial | to the company than say making sure it's employees are | happy. | | Whoa, citation needed. I'm fairly certain yhis isn't true | unless you are optimizing for just the next month. | jscode wrote: | >> Having dealt with 2 PE exists, rarely is the proposal | something as upfront and silly as slash everyones pay. | | Agreed, but a gradual erosion can occur over the course of | several years. PEGs and the operating company's management | team have massive incentives to hit their growth metrics. | If decreasing 401K contributions helps management hit their | EBITDA target, many would argue that they should do exactly | that. However, the impact of these decisions accumulates | over time and can eventually derail a company's | performance. | andrei_says_ wrote: | What were the longer term changes you saw and their | outcomes for the health of the teams and companies? | serverlessmom wrote: | I definitely agree with you. I would further argue that the | "analyst mentality" has become so engrossed in the foundation | of capitalism(and always has been, really) that we see | companies across many industries closing down due to the | "worker shortage". When in reality the CEO's of the companies | complaining the loudest that "no one wants to work" have | decided that paying living wages in a deeply competitive job | market is less doable than letting the company totally | stagnate and dissolve due to lack of people willing to work | hard for pennies. | abhishekjha wrote: | Does this have any QM implications? | bell-cot wrote: | "The moral is to the physical as three to one." - Napoleon I | seanwilson wrote: | > What McNamara didn't keep track of was the narrative of the | war, the meaning that it had both within the military forces of | each side, but also in the civilian populations of the nations | involved. | | I'm probably not following and not saying it's easy but aren't | there some metrics you could track that with? How was it | determined/measured later that the meaning was important? | Couldn't you at least do polls in your own country? | asplake wrote: | But that's the point - it's not just how many, but what they | are willing and able to do, their positional and operational | strengths and weaknesses, and so on. Condensing all that into a | number is impossible. | rossdavidh wrote: | I believe the McNamara Fallacy is a real thing, although I would | say Goodhart's Law expresses the more fundamental issue. But, the | examples they give (Vietnam and Afghanistan wars) are bad | examples, because in both cases the issue was primarily not that | the U.S. military was relying too much on quantitative metrics, | but rather that it was being asked to do a job which militaries | are not good at. | | In neither case, despite what many deniers and apologists | proclaim, was the war "winnable" by the military, because in both | cases the primary problem was that the government of the nation | in question did not have the respect (and thus support) of its | people. There is no strategy that an outside military can use, | which will fix this, because it's not a military problem, and a | military has the wrong personnel, the wrong training, and thus | the wrong institutional mindset. | | It is as if you sent the Red Cross in to defeat the Viet Cong or | the Taliban, and then concluded that the reason they failed was | that they were measuring numbers of hospital beds and counting | how much medical equipment they needed. It was the political | leadership, who decided this was a job for a military to perform, | that was the root problem. McNamara's metrics were neither here | nor there. | shellback3 wrote: | McNamara cut his teeth on real world problems in the AAF's Office | of Statistics and worked with General Curtis LaMay to plan how | best to use the B29 bombers. He became an expert with statistics | and other math tools and thought they could be widely applied to | business and, of course, war. Once he had these marvelous tools | he found nails everywhere. | | For instance the rate of German tank production was accurately | estimated by collecting the serial numbers of all the tanks (and | some tank components) that were knocked out. Best methods of | attacking submarines was worked out this way as well as well as | where to place armor in bombers, etc. | topspin wrote: | > Once he had these marvelous tools he found nails everywhere. | | He also found two presidents and a sycophantic media that hung | on his every word. | | His figures were excellent. In the earliest days of Vietnam, | long before anyone outside of Asia could find it on a map, he | produced eerily precise predictions of the costs -- in lives, | dollars and time -- of the future conflict. He told them what a | civil war in the jungle would look like and they pulled the | trigger. | | As far as the value of measurement goes; I think most of the | low hanging fruit has been picked (a consequence of the | "Information Age") and what we struggle to 'measure' today is | far less tractable. As a result our measurements are frequently | corrupted in the service of prevailing agendas or not permitted | at all for fear of undesirable results. | dempedempe wrote: | This article is good, but not great - the author only gives one | example of how quantitative-only reasoning can be bad (the | example of the poppies). The other "example" is just the US | military lying. | | There are also no specific examples of non-quantitative reasoning | that, if ignored, would be damaging. | | I feel like the Wikipedia article does a better job explaining | this: https://en.wikipedia.org/wiki/McNamara_fallacy | | Also, why is there an entire webpage dedicated to this? | hummusandsushi wrote: | The webpage appears to be made to serve as a warning to data- | driven businesses to not fall into the same perverse set of | incentives that McNamara created, to encourage business | managers to diversify their accounts of the success of their | business beyond just the quantitative narrative. | pakitan wrote: | I'd be more cynical and suspect this is some kind of | elaborate SEO strategy. Submit the site to social networks -> | wait till gets some love from the Google algorithm -> put ads | on it -> profit. | hummusandsushi wrote: | Possible, but this seems of rather a niche interest for a | successful SEO strategy. I could grant that it would then | be targeted specifically at the HN sort of social networks. | mulmen wrote: | Why not just start with ads? | enkid wrote: | The example with the US military lying certainly demonstrates | another issue with over valuing metrics - making sure you have | good data. People who get wrapped around metrics also tend to | not look at the data they are being presented. One of the | issues McNamara had was he was receiving inflated body count | figures. [0] Not only was he measuring the wrong thing, he was | measuring it badly. | | [0] | https://en.wikipedia.org/wiki/Vietnam_War_body_count_controv... | the_af wrote: | It's also a good example of perverse incentives or "gaming | the metrics" (in a horrifying way). If the main measure of | success was body count, and every kill was considered an | enemy by default, this encouraged just killing people and | counting them as successes. | shellback3 wrote: | I was in Vietnam in a Naval Support Group. I had | conversations with army officers that had combat experience | and learned that since no statistics were collected about | dead civilians all the dead were counted as enemy soldiers | - and everyone from the top down knew it. | paulpauper wrote: | I don't understands what the fallacy is. It is that an | overreliance on data can overlook factors not in the data? Why is | that a surprise. You would also have to show that not relying on | data would generate better results. | warning26 wrote: | _> It is that an overreliance on data can overlook factors not | in the data? Why is that a surprise?_ | | The vast majority of PMs I've worked with don't seem to | understand this at all. If it's not in the metrics, it doesn't | exist! | 0xdeadbeefbabe wrote: | If understanding is what you seek then | https://www.youtube.com/watch?v=eNDsd798HR4 is a nice long | discussion on McNamara overlooking factors not in the data. The | fallacy didn't earn its name for simple reasons. | _Nat_ wrote: | It's probably one of those things that might seem more obvious | in simple cases, but might surprise folks in more complex | cases. | | For example, a simple case: Say you go on vacation for a few | weeks with a certain amount of cash to spend. Upon arriving at | your destination, you immediately purchase some indulgence, | and, hey, you're feeling better! Why not immediately keep | spending as much as possible to maximize? | | For example, a complex case: Say you're leading a country. You | set up policies that, after a few years, seem to have led to a | higher GDP than was previously expected. Does that suggest that | the policies are leading to a better future? | | In the simple cases, like the vacation-example above, it's easy | enough to understand the scenario and what's going on. And we | can imagine that, hey, immediately spending all of the cash | might lead to poor consequences for the rest of the vacation, | even if they display some quick-satisfaction at first. | | But in more complex cases, like with a country's policies | leading to a higher GDP, stuff can get trickier. We might say | that it's the lack of a top-level model: unlike in the vacation | scenario, where we were easily able to predict that there'd be | a lack of money later in the vacation, it might be harder to | say what else might be going on besides the GDP going up. And | all other things held equal, presumably a higher GDP would be | better than a lower GDP, and therefore all evidence points | toward the policies being a good idea, right? | enkid wrote: | GDP is actually a great example. It's actually very easy to | increase a country's GDP if that's the only thing you care | about. You just borrow more money and then spend it. GDP is | literally a measure of money changing hands inside a country. | If you borrow as much money as you can and then spend it on | things like infrastructure projects, you can instantly | increase the GDP. Banks see the growing GDP figure and assume | that its a good thing and let you borrow more money. That is, | until they don't. | | This happened with Brazil in the 1970's, when the oil crisis | made Brazil believe that they were going to face a downturn. | To overcome this, they borrowed money and went on an | infrastructure spending spree. This made Brazil look like an | economic miracle, growing when everyone else was struggling. | This encouraged more banks to lend to Brazil. The problem is | the money was spent on short term growth instead of things | that would more systematically grow the economy over the long | term. The government (and lending banks) was substituting | year-to-year GDP numbers for economic health. Once credit | tightened in the early 1980's, Brazil's growth plummeted. | | This is of course an over simplification, but I think we are | seeing similar problems in modern economies. People often | cite China's GDP growth, but they don't balance that out with | the debt they are taking on in order to finance that growth. | potatolicious wrote: | There are a bunch of factors and implications from the piece, | which I had hoped they'd go into. But some of the more obvious | implications which are often more organizational than they have | to do with the data per se: | | - the data you are collecting may not include factors that are | consequential to outcomes. Treating these unmeasured factors as | inconsequential is hazardous. The piece specifically mentions | this effect. This effect is not a surprise but yet many | organizations fall into this trap so it seems to be worth | mentioning. | | - the difficulty of measuring some factors will _result in | their exclusion_ from the dataset. Some things are | intrinsically hard to measure, and many organizations will as a | result refuse to measure them, and make decisions without them. | There needs to be a conscious and active process | organizationally to resist this and find effective ways of | measuring them. | | - the difficulty of measuring some factors will _result in | easier but less useful proxies being used in their place_. Same | as above just with a somewhat different outcome. Organizations | are often blind to this happening as they come to believe the | proxy is as good as the real measure (or sometimes even that | the proxy _is_ the measure). A good industry example of this is | clickthrough rates being treated synonymously with audience | interest or content quality. | | And a point I wish the piece made but did not: | | - measuring something does not grant automatic understanding of | the phenomenon and gives you no predictive power. It's one | thing to quantifiably know the blue button gets more clicks | than the green button, but that grants you no insight into why. | Many tech companies fall into this trap - where despite | investing heavily in experimentation their modeling of the | product space doesn't improve over time, since they fail to | take the step to convert observation to generalizable | hypotheses that improve their model for the product and market. | | This last effect IMO is _huge_ in our industry, and is why the | same low-level experiments are being re-done over and over | again. This creates more product churn and reduces your product | velocity - since the lack of proven product models means you | 're mostly flying blind and using ex-post-facto experimentation | on live users to figure out what to do. | | Experimenting on button colors is all well and good, but the | end result of that data should be a color theory that explains | what colors to use when, not forever A/B testing every single | button color for the rest of time. | torginus wrote: | It probably tries to make the point that when a good metric | becomes a target, it ceases to be a good metric - either | through people gaming the metric, or because of diminishing | returns. | | But I admit the article gets to its point in a muddy and | circuitous way, typical of business school anecdotes. | asveikau wrote: | > Why is that a surprise | | It's surprising to many. | | Many people will decide they are "data driven" and act like | this gives them infallibility and total correctness. Lots of | people need to be told this. | facorreia wrote: | My understanding is that the fallacy is cherry-picking a few | data points that are easy to measure and then claiming that | these are the most important data points, and then optimizing | execution for moving them in the "right" direction. This is | described in the metaphor of "searching under the streetlight". | The key to this con is to act confident and to challenge any | objections as lacking supporting data. | monkeybutton wrote: | I think the fallacy is when factors not captured or tracked by | data are treated as non-existing or consequential. In Vietnam, | they tried to quantify how pacified each village was as | measurement of success. It did not work. This isn't saying no | data is better, just that data on its own does not win a war, | or the hearts and minds of villagers. | | See also: | https://www.theatlantic.com/technology/archive/2017/10/the-c... | martincmartin wrote: | Two examples come to mind: | | - Investing your money with the fund manager who has the highest | returns over the last year, or even 5 years, is called | "performance chasing" and generally has worse returns than | investing in an index. Think investing in dot com stocks in 1999, | and real estate in 2007. | | - In Moneyball, about using stats to improve a baseball team, | they didn't care how many home runs someone hit. In fact, getting | runs that way was considered bad. They wanted to get players on | base and them get them home, even if through walks. There's a | part of the book where a player is getting a lot of home runs, | and therefore increasing the team's score, but the manager | doesn't like it. "It's a process", "we have a process" he keeps | saying. | beaconstudios wrote: | > In another case of military metrics gone wrong, the US military | reported success in undermining Taliban financing after it paid | Afghan farmers to destroy their crops of opium poppies. What went | unreported, however, was that the farmers planted larger fields | of opium poppies in response, in the hopes that they might be | paid by the US military to destroy the crops again. When US | payments didn't come through, the opium was harvested and entered | the international drug trade. Much of the profit went to support | the Taliban's anti-American military operations. | | That's pretty ironic because its a near perfect replication of | the Cobra effect, the canonical example for bad incentives | leading to unintended outcomes: | https://en.m.wikipedia.org/wiki/Perverse_incentive | aeternum wrote: | From a game-theory POV, all wars are caused by misunderstandings | / partial information. | | If each side really knew the true strength of the other side, it | would be clear which would win and therefore actual war is | illogical and unnecessary. | chasd00 wrote: | i'll add that there's a lot of people that see death as the | door to paradise and fighting for their cause as the key to | that door. No amount of logic or information is going to stop | them. | | when you're dealing with the mentally ill, all bets are off. | aeturnum wrote: | This seems untrue. | | There are also genuine uncertainties about the future: things | that _neither_ side knows but will impact the outcome. Because | of that, there is always a range of outcomes for a conflict, | and so both sides would still have an incentive to carry out | the war. | | Consider, for instance, if there are two countries (A and B) | looking at conflict with perfect information. Given their | relative strengths, the war will cost both $1 billion, but A | will win and reap $1 billion + $1 in plunder. In your model, B | will always surrender and give A $1, but the reality is there | is a big spread about both the costs and the gains. Lots | depends on how well each side fights and responds, even with | perfect knowledge of the other. | | So...you know, if it's the US versus...the Philippines[1], then | sure, the uncertain range of outcomes is small enough that | rational thing for the Philippines to do is surrender. But for | countries that are even reasonably well-matched, they would | want to fight with the goal of making the conflict unattractive | for the other side. Probably the best historical example of | this is the USSR in Afghanistan. | | [1] I am sure the US would never invade and occupy them just | because the US found it convenient ;) | thaumasiotes wrote: | > Consider, for instance, if there are two countries (A and | B) looking at conflict with perfect information. Given their | relative strengths, the war will cost both $1 billion, but A | will win and reap $1 billion + $1 in plunder. In your model, | B will always surrender and give A $1 | | In the perfect information scenario, B will surrender more | than $1. B is better off at any tribute level up to | $1,000,000,000. | crazy1van wrote: | > If each side really knew the true strength of the other side, | it would be clear which would win | | This is too simplistic. Often wars are heavily influenced by | things other than just the strengths of both sides. For example | - What if the ground hadn't been soaked from days of rain at | Agincourt and Waterloo? What if the Germans hadn't held back | their armor reinforcements for so long on D-Day? | asdff wrote: | They'd still lose. Wars are fought not from individual | battles but from logistics. Modern military's don't even | engage if there isn't a lopsided power imbalance favoring | their success. | chasd00 wrote: | if you find yourself in a fair fight it's time to re-think | your strategy | thaumasiotes wrote: | It only takes one side to engage. | asdff wrote: | Look at the orders of magnitudes differences in | casualties in desert storm, iraq, and afghanistan. When | Americans battle they ensure they have a massive power | advantage for every fight and this is abundantly clear | just from the casualty data. Even if the other side | engages first like in desert storm, Americans ensure that | they don't land boots unless they have a tactical | advantage which they had in Desert Storm thanks to | technology that Saddam and his army, which was one of the | largest and experienced land forces in the world at the | time, could not begin to compete with. | 0xdeadbeefbabe wrote: | Yeah like in Black Hawk Down. | asdff wrote: | In real life the battle of mogadishu saw Americans with | 10x fewer casualties. Most modern battles have numbers | like this with an order of magnitude fewer casualties on | the American side, because American logistic planners try | to ensure a massive power advantage. | hackeraccount wrote: | Is that true? What if the situation is that I am currently | stronger then you but can expect to become weaker in the | future? Why not force a war now if I can expect to win? | | I still haven't finished it (too depressing) but my reading of | the book The Sleepwalkers about the events leading up to WWI is | that all the parties thought they were in that place; even if | they all had a full picture it's certainly possible those | beliefs would in fact be true. | aeternum wrote: | If I know that you are currently stronger and would | ultimately win a war, I should logically concede now since I | will lose anyway. | | If I know that you are currently stronger but we both also | know that I can draw out a war such that I eventually become | stronger, then you shouldn't go to war since you will | ultimately lose. | vkou wrote: | > If I know that you are currently stronger and would | ultimately win a war, I should logically concede now since | I will lose anyway. | | You'll often lose less if you fight to a loss. | | In fact, you often won't have to fight at all if you make | it clear to your adversary that you will fight to a loss. | | If I'm a street thug, I'm not going to try to mug someone | who clearly communicates that their response to a mugging | is detonating their suicide vest. I will, however, | repeatedly victimize someone who clearly communicate that | they are a pushover. | | This is an iterative game, against the same players. Loss | aversion is not a winning strategy to pursue in the long- | term. Something resembling tit-for-tat is. | vlovich123 wrote: | That's a fallacious line of reasoning on several fronts. | | A) The beneficiaries of a war need not be the ones who pay the | cost. If you personally stand to benefit from the war itself, | regardless of win or lose, you want the war. | | B) It's applying a rational economics approach when it's pretty | clear that humans aren't rational. Japan kept fighting even | when it was clear they would lose. Pride, image, your future | after the war. All things that impact the decision that gets | made. | | C) War isn't typically total like it was in WWII. Most wars are | just skirmishes or prolonged skirmishes that don't meaningfully | change borders. In such a scenario, a single "war" loss may | still be an important fight to have in the middle of the | overall conflict (i.e. to signal to your opponent that you're | not a pushover and that continuing on their current path has | consequences). | | D) Win/lose isn't the only outcome of war. Lose/lose is also | another scenario. Additionally, you might lose the war but | winning may not be the goal. For example, maybe it only takes | me 1/10th the resources for me to wage the war that it takes | for you to win. A single war might cost me 100 units, but it's | important for me to be able to cause you to spend 1000 units | because then I have a 900 unit advantage in some other area. | | There's many more, but I don't think imperfect information is | the only reason war exists. | asdff wrote: | I really don't think the GP is wrong here. Yes there are | ideologues, but there were plenty in the Japanese high | command who understood logistics and understood that war with | the U.S. would end the Japanese empire. It was the same way | with Germany. The people who understood logistics knew it | would be absolutely impossible to win a two front war and | nearly impossible to win a one front war with the U.S. having | a thumb on the scales just with lend lease alone. Germany | sent their best generals to the east and they knew they had | no hope but to delay the inevitable, and no alternative but | to listen to Hitler. They continued to fight because to deny | Hitler's madness often meant less than favorable outcomes for | these commanders than making peace with Allies in 1945. The | war in Japan ended when it became impossible for the emperor | to continue with the cognitive dissonance and ignore the fact | that the fight had ended years before it even began. | Logistics have won every war there ever was. | tomrod wrote: | (1) Externalities are handled in the game theory literature | | (2) Preference definitions are not in scope for most game | theory lit; what you're describing are preferences and | expected utility, which ARE in the literature and metamodel | | (3) This is not really true. It depends on how big you are | and how big your coalition is. Consider that Saddam Hussain | is no longer in charge of Iraq. | | (4) Yes, potential outcomes are considered under game theory. | | I get your point that OP's statement feels reductionist, but | I wanted to explain that the OP is actually correct. | [deleted] | s28l wrote: | There is a lot to disagree with here. First, the article itself | provides an example of an army with a superior fighting force | that had far fewer casualties than the other side, yet still | lost the war. That was due to the civilians back at home | lacking the stomach for a long, protracted war abroad. So even | if one side has superior strength, and even if the side with | the superior strength wins every battle, they might still lose | the war. | | Another issue is your implicit assumption that the side with | the superior fighting force will always win the war, but I | don't think you can make that assumption. Just like the better | football team can lose to the underdog, there is a stochastic | element to warfare. A sudden bit of bad weather can turn the | tide of a battle. There are countless other elements that are | unobservable and unpredictable that can decide which side wins. | | There are also asymmetric payoffs to going to war. A nation | might have a slim chance of victory, but the cost of defeat or | surrender might be genocide or subjection. How do you assign a | payoff to the choice "surrender" when the outcome is the | destruction of your nation as it once existed? | | [0] | https://en.wikipedia.org/wiki/Strategy_(game_theory)#Pure_an... | 13415 wrote: | Wait a minute. Shouldn't the goal of the defender be to make it | clear to the attacker that they would incur losses that | outweigh possible wins for the attacker, thereby making the | attack a lose-lose scenario? This is only possible if the | defender makes it credible to the potential attacker that they | will defend their country even when the costs would be | extremely high and they cannot possibly win the conflict. In | other words, at least from deterrence point of view it's not | about winning, it's about making sure the attacker would | overall lose more than gain (which is not the same as winning | against the attacker). | sf_rob wrote: | This isn't a great article IMO (I say as someone who is a big fan | of qualitative user research, mixed methods, etc). | | Per the first example, many types of attitudinal data can be | quantified and conflating attitudinal data with qualitative data | is itself a fallacy. It's possible that there were quantitative | attitudinal signals that could have been captured or created as | inputs to a more accurate model. | | Per the second example, this is more a question of data validity | than the metric itself. If the metric could be validated through | better design and gamification prevented then it would likely | still be a helpful indicator. Granted this is a very hard | problem. | serial_dev wrote: | My thoughts, too. It's not that quantitative data is bad, he | just picked the wrong one. | scubbo wrote: | Surprised not to see Goodhart's Law[0] referenced here - "When a | measure becomes a target, it ceases to be a good measure". Not | the same concept, but a related one (as is the Cobra Effect[1], | of which the poppy-field burning is an example) | | [0] https://en.wikipedia.org/wiki/Goodhart's_law [1] | https://en.wikipedia.org/wiki/Perverse_incentive#The_origina... | musicale wrote: | From [1]: "It was discovered that, by providing company | executives with bonuses for reporting higher earnings, | executives at Fannie Mae and other large corporations were | encouraged to artificially inflate earnings statements and make | decisions targeting short-term gains at the expense of long- | term profitability." | | What a shocking revelation! Could it possibly apply to other | companies that report quarterly results? ;-) | scubbo wrote: | I vividly remember a "small-hands" (smaller all-hands) very | early in my career, where the VP basically said "we recognize | that oncall is a burden, and we want to compensate you for | it. At first we thought we should give you a bonus related to | how many times you get paged - but then we realized that | that's incentivizing you to build flaky systems. Then we | considered giving a bonus in inverse relation to how often | you're paged - but then we though, no, they're smart | engineers, they'll just turn off the monitoring and collect | the bonus. So we're giving you a flat rate". | | In the absence of some untamperable objective way to measure | service health (the concept of SLAs was a distant dream at | that point), can't fault that reasoning, tbh! | csours wrote: | A different view of the problem: | | https://www.nngroup.com/articles/campbells-law/ | snidane wrote: | I wonder if it gets widespread to oppose the dominant | hyperquantitative, every fart tracked in jira, can't measure - | can't manage trend. | | Which was never pushed by anybody from the industry, but got | spread by project management frauds. | | https://deming.org/myth-if-you-cant-measure-it-you-cant-mana... | skybrian wrote: | Maybe a better way of thinking about it is that a process that | gathers numbers can be useful, but understanding the context in | which the numbers were gathered (methodology and whatever else is | happening at the time) is more important than the numbers | themselves. Without context, you don't know what you have. | | For example, surveys gather numbers, but if you don't understand | how the people who answered the questions interpreted them, you | don't know what the numbers mean. Asking people what they really | meant by their answer is only possible for the people doing the | survey. | hwers wrote: | At this point "fallacy" just strikes the same note to me as | "misinformation". It's mostly used as a political device to get | ideas you want to be true into social consensus but the reality | is usually more ambiguous. | [deleted] | NovemberWhiskey wrote: | Or, put slightly more succinctly: "you get what you measure" | shellback3 wrote: | It should be remembered that McNamara cut his teeth working for | the office of statistics with Curtis LeMay in the use of B29 | aircraft. He was proud that he was part of a larger effort using | statistics and other math that helped to win the war and, with | that hammer, most everything could be a nail. ___________________________________________________________________ (page generated 2022-02-01 23:00 UTC)