The Innovators Studio with Phil McKinney cover art

The Innovators Studio with Phil McKinney

The Innovators Studio with Phil McKinney

By: Phil McKinney
Listen for free

Summary

Forty years of billion-dollar innovation decisions. The real stories, the hard calls, and the patterns that repeat across every organization that's ever tried to build something new. Phil McKinney shares what those decisions actually look like. Phil was HP's CTO when Fast Company named it one of the most innovative companies in the world three years running. He co-founded a company and took it public. Now he runs CableLabs, the R&D engine behind the global broadband industry. This isn't theory. It's what happened. And what you can see coming if you know what to look for. Running since 2005, originally as The Killer Innovations Show, now The Innovators Studio. Tens of millions of downloads. Full archive at killerinnovations.com. New episodes at philmckinney.com.Copyright 2005-2026 Techtrend Group LLC. See philmckinney.com Economics Leadership Management & Leadership Personal Development Personal Success
Episodes
  • Why Most Organizations Aren't Funding Innovation
    Apr 29 2026
    Twelve official definitions for R&D. Zero agreement. The US government publishes at least a dozen distinct official definitions across agencies, accounting standards, tax authorities, and international bodies. Not one agrees with the others on where research ends and development begins. Trillions of dollars flow through R&D budgets every year. Boards approve them. Investors evaluate them. Governments subsidize them. Analysts benchmark them. And the term at the center of all of it has no settled definition. A company can gut its research investment without triggering a single alarm on its income statement. Researchers who gained rare access to confidential federal R&D data found exactly this: when companies face financial pressure, they cut research while leaving development essentially untouched, and the combined number barely moves. Every benchmark, every board conversation, every investment thesis built around the R&D line may be built on sand. Innovation, ideas made real, requires both. Research is how you find the idea. Development is how you make it real. Strip out the research and you're not innovating, you're iterating on what already exists. Strip out the development and you're just experimenting. The problem is that nobody in the room knows which one they're actually funding, because the definition that would tell them doesn't exist. Someone needs to draw the line. This episode is about why nobody has, and the definition I think should replace the chaos. By the end, I'm going to put that definition in front of you and ask you to push back on it. Not to agree. To tell me where it breaks. How We Got Here Four institutions took a run at defining R&D. Each one got it right for their own purposes. None of them got it right for yours. Frascati: Built for Governments In June 1963, OECD economists met at a villa in Frascati, Italy, south of Rome, and produced what became the international standard for measuring R&D across nations. Now in its seventh edition. The Frascati Manual divides R&D into three tiers: basic research (theoretical work with no application in view), applied research (original investigation toward a specific practical objective), and experimental development (using existing knowledge to produce new products or processes). To qualify, an activity must be novel, creative, uncertain in outcome, systematic, and transferable. Used by governments across roughly 75 countries. Solid for what it was designed to do: let nations compare R&D investment on consistent terms. What Frascati cannot tell you: whether a specific company's spending is creating competitive advantage. It counts the type of activity. It doesn't assess what the activity produces for the organization doing the spending. A company can satisfy every Frascati criterion investigating something every competitor already knows. The knowledge is new to them. That is enough. The accountants drew a different line, for a different reason, with a different consequence. FASB: Built for Accountants In October 1974, the Financial Accounting Standards Board issued Statement No. 2, Accounting for Research and Development Costs, now codified as Topic 730. Every public company filing under US GAAP operates under it. The rule: all R&D costs expensed as incurred. Research, development, basic, applied: one line on the income statement. Their definition: research is a planned search aimed at discovery of new knowledge. Development is the translation of research findings into a plan or design for a new product. The rationale is explicit in the original standard. Future benefits from R&D are, in FASB's language, "at best uncertain." Expense everything immediately. The standard solved the problem it was asked to solve, which was accounting treatment: when to recognize the cost, not whether the cost was strategically sound. The consequence: sustaining engineering, feature maintenance, and incremental product updates all land on the same line as genuine exploratory research. Nobody looking at the income statement from outside can see the difference. The number is technically accurate and analytically opaque. Abraham Briloff, the late accounting professor at Baruch College, put it plainly: "Accounting statements are like bikinis. What they show is interesting, but what they conceal is significant." He was talking about financial reporting broadly. He could have been writing specifically about the R&D line. Researchers at Duke and London Business School spent years tracking corporate scientific output and found that it declined steadily across industries even as headline R&D spending kept rising. The combined number was hiding a substitution. Nobody on the outside could see it. Outside the United States, a different standard governs, and it creates a comparison problem most analysts never account for. IFRS: Built for International Investors IAS 38 governs R&D under IFRS, and its treatment differs from FASB in one significant way. Research costs are ...
    Show More Show Less
    21 mins
  • R&D Spending Is the Most Misleading Number in Business
    Apr 15 2026
    Every public company's R&D number is a lie hiding in plain sight. Not because anyone falsified it. Because the number was never built to tell the truth. It was built to satisfy an accounting standard written in 1974. And for fifty years, boards, analysts, and CEOs have been making billion-dollar innovation decisions based on a number designed by accountants to solve a different problem entirely. Here's what makes this genuinely strange. The real number exists. The government has been collecting it from every major US company for decades. It would answer the question every innovation leader and investor actually needs answered. And it is locked away by federal law. Confidential. Never published. Never seen by the people who need it most. It's sitting in a federal database right now. And there's a way to estimate it for any public company, without asking anyone's permission. I know it exists because I spent years building it from the inside. Why the R&D Signal Was Blurry When I was running innovation at HP, we discovered this problem firsthand. We had a connection between R&D investment and gross margin that held up across decades of HP history. Better than anything Wall Street was using. But the signal was blurry. None of us could figure out why. The answer came from a question someone on the team asked almost as an aside. What if R&D isn't one thing? Research and Development Are Not the Same Thing Think about what actually lives inside a typical R&D budget. There's a team somewhere investigating whether a new approach could enable a capability that doesn't exist yet. No product defined. No spec written. Asking whether something is even possible. And there's a team building the next version of a product that ships in eighteen months. Spec locked. Timeline set. Engineering executing against a defined target. Both show up on the same line in the budget. Both get called R&D. Both count equally toward the number that gets reviewed every quarter. They are not the same thing. One is Research. The other is Development. Research is the work you do when you don't yet know what you're building. The output is understanding. New knowledge that might enable future products nobody has designed yet. You can't know exactly what you'll find. If you already knew, it wouldn't be research. Development is the work you do when you know exactly what you're building. The spec exists. The product is defined. The question isn't what to make. It's whether it can be made, on time, at cost, at quality. One creates the future. The other delivers the present. And for fifty years, every public company in America has been required to report them as one indistinguishable number. When we split the HP data along that line, Research on one side and Development on the other, the signal sharpened immediately. Research spend, measured against gross margin three to five years later, was a meaningfully stronger predictor than the combined number had ever been. The blur hadn't been in the gross margin data. It had been in the R&D number itself. Two fundamentally different things, averaged together, producing a number that looked precise and predicted almost nothing. But splitting R from D at the company level was only the beginning. The model was still lying to us. Just more quietly. Why Company-Level R&D Splits Still Mislead Even with the split, something was still soft. HP wasn't one business. It was dozens. Printers, PCs, servers, software, each running on different timelines, different technology cycles, different competitive dynamics. What if the R/D split meant something different depending on where it was applied? We pushed it to the product line level. Then further, to the platform level within product lines. Printers were the clearest example. HP's printer business wasn't one story. There were platforms built on established technology. Mature ink systems, proven print head chemistry, products that had been shipping for years. And there were platforms built on genuinely new core technology. New chemistry. New mechanisms. New approaches to fundamental problems that nobody had solved yet. Research investment by platform told a completely different story than Research investment by product category. The Research going into new technology platforms had a completely different relationship to future margin than Research going into mature platforms. Different time horizons. Different risk profiles. Different margin implications years down the road. Laptops told the same story. A traditional consumer laptop line and a high-performance portable workstation weren't the same investment. One was Development-heavy. Defined product, known market, engineering executing against spec. The other had genuine Research behind it. Unsolved thermal problems, new form factor constraints, and materials questions that hadn't been answered yet. When a single R&D assumption is applied across all of that, treating every dollar the same regardless of what it actually...
    Show More Show Less
    17 mins
  • The Innovation Metric Bill Hewlett and Dave Packard Used
    Apr 1 2026
    Every public company in the technology industry measures innovation spending the same way. R&D as a percentage of revenue. Why? Because Wall Street tracks it. Boards benchmark it. CEOs get fired over it. And it tells you almost nothing about whether the spending is working. Bill Hewlett and Dave Packard knew that. From the very beginning, they measured something different. Something the rest of the industry has been ignoring for seventy years. And the proof was sitting in a paper that Chuck House pulled out and sent to me after a conversation at a Computer History Museum board meeting. By the end of this episode, you'll know what that metric is, why it works, and why the one everyone else uses makes it nearly impossible to tell whether your innovation investment is building the future or just burning cash. Here's how I found it. The Question That Wouldn't Let Go In the last episode, I talked about the argument with Mark Hurd. The question was over whether HP should cut R&D as a percentage of revenue to match Acer. I knew Mark was fundamentally wrong. But I couldn't prove it. The only metric on the table was R&D as a percentage of revenue. That was what Wall Street expected. It's what shareholders expected. It's what the board expected. But I couldn't argue against it, because I didn't have the data. I needed a better metric. So I decided to go back to the beginning. HP's complete financial records dating back to the 1940s. Division by division. R&D project by R&D project. The actual operating data. I got access to all of it. The HP archive team gave me direct access to Bill and Dave's original notebooks. Now, data alone wasn't enough. It was mountains and mountains of data, and you're trying to extract the signal. What is the trigger in that data? The conversation that cracked it open happened outside HP. The Man with the Medal of Defiance I was at a Computer History Museum board meeting, standing next to Chuck House, and I shared with him the struggle I was having. A little context on Chuck. He spent twenty-nine years at HP. He was the Corporate Engineering Director and he helped launch dozens of products. He's also the recipient, from David Packard himself, of the Medal of Defiance. The Medal of Defiance was given to him because David had told him at one point to kill a product line. Chuck went around that decision, put the product into the catalog, shipped it, and it turned into a phenomenal success. When David gave Chuck the medal, the citation was something along the lines of: "for going above and beyond the stupidity of management and doing what was right." Chuck and Raymond Price co-authored a book called The HP Phenomenon, published by Stanford Press. It's the deep dive into the history of the innovation culture inside HP, all of the metrics used back in the Bill and Dave days that put in place the structure that allowed HP to be successful. By the time I'm at HP, Chuck had long since moved on. He was running Media X at Stanford, the university's research program on innovation, media, and technology. But we both served on the Computer History Museum board. At that board meeting, I shared the argument I'd had with Mark and the search for a better metric. I had a strong feeling there was something around gross margin. That R&D investment impacted gross margin. But a feeling isn't an argument. I needed data. I needed to correlate R&D spend to margin, and that's extraordinarily hard to do when you've got all these different product lines and divisions. Chuck got this little smile on his face and said, "I need to send you something." The Paper and the Whiteboard What he sent me was a paper. A journal paper he and a few of his colleagues had written decades before. And it laid out the connection between research investment and margin performance. The correlation I suspected but couldn't prove was right there on the page. I read it that night. The next morning I emailed Chuck, and I was just really excited. What they'd written decades ago matched what I was finding in the data. That email exchange turned into an invitation. I asked Chuck to come to HP Labs. We met in a conference room in Building 3, the main building for HP Labs at the time. And I'll tell you, I look back on this and it makes me smile a little, because this conference room was just down the hall from Bill and Dave's offices. HP preserved those offices exactly as Bill and Dave left them. You can walk in there today, see their desks, see their offices, just as they were on their last day. There's something about being that close to where it all started that makes the history feel less like history and more like unfinished business. Chuck walked up to the whiteboard and drew two things. On the left side: R&D as a percentage of revenue. The metric every company reports. The metric Mark used to argue HP was overspending. Chuck's point was simple. That metric tells you how much you're spending. That's it. Nothing about whether your ...
    Show More Show Less
    20 mins
No reviews yet