The biggest issue I see is Microsoft's entire mentality around AI adoption that focuses more on "getting the numbers up" then actually delivering a product people want to use.
Most of the announcements I hear about Copilot, it's always how they've integrated it into some other piece of software or cut a deal with yet another vendor to add it to that vendors product offering. On the surface there's nothing wrong with doing that but that just seems to be the ONLY thing Microsoft is focused on.
Worse yet, most of these integrations seem like a exercise in ticking boxes rather than actually thinking through how integrating Copilot into a product will actually improve user experience. A great example was someone mentioned that Copilot was now integrated into the terminal app but beyond an icon + a chat window, there is zero integration.
Overall, MS just reeks of an organization that is cares more about numbers on a dashboard and pretty reports than they are on what users are actually experiencing.
There aren't any "AI" products that have enough value.
Compare to their Office suite, which had 100 - 150 engineers working on it, every business paid big $$ for every employee using it, and once they shipped install media their ongoing costs were the employees. With a 1,000,000:1 ratio of users to developers and an operating expense (OpEx) of engineers/offices/management. That works as a business.
But with "AI", not only is it not a product in itself, it's a feature to a product, but it has OpEx and CapEx costs that dominate the balance sheet based on their public disclosures. Worse, as a feature, it demonstrably harms business with its hallucinations.
In a normal world, at this point companies would say, "hmm, well we thought it could be amazing but it just doesn't work as a product or a feature of a product because we can't sell it for enough money to both cover its operation, and its development, and the capital expenditures we need to make every time someone signs up. So a normal C staff would make some post about "too early" or whatever and shelve it. But we don't live in a normal world, so companies are literally burning the cash they need to survive the future in a vain hope that somehow, somewhere, a real product will emerge.
For most software products I use, if the company spent a year doing nothing but fixing P2 bugs and making small performance improvements, that would deliver far, FAR more value to me than spending a year hamfistedly cramming AI into every corner of the software. But fixing bugs doesn't 1. pad engineer's resumes with new technology, or 2. give company leadership exciting things to talk about to their golfing buddies. So we get AI cram instead.
I think it is more externally driven as well, a prisoners dilemma.
I don't want to keep crapping out questionable features but if competitors keep doing it the customer wants it -- even if infrastructure and bug fixes would actually make their life better.
Last time I saw results of a survey on this, it found that for most consumers AI features are a deciding factor in their purchasing decisions. That is, if they are looking at two options and one sports AI features and the other doesn’t, they will pick the one that doesn’t.
It’s possible AI just seems more popular than it is because it’s easy to hear the people who are talking about it but harder to hear the people who aren’t.
Consumers is nice, but far more important are the big corporate purchases. There may be a lot of people there too who don't want AI, but they all depend on decisions made at the top and AI seems to be the way to go, because of expectations and also because of the mentioned prisoner's dilemma, if competitors gain an advantage it is bad for your org, if all fail together it is manageable.
My job is like that, although it's mostly driven by my direct boss and not the whole company, but our yearly review depends on reaching out to our vendors and seeing if an AI solution is available for their products and then doing whatever is necessary to implement it. Most of the software packages we support don't have anything where AI would improve things, but somehow we're supposed to convince the vendor that we want and need that.
>It’s possible AI just seems more popular than it is because it’s easy to hear the people who are talking about it but harder to hear the people who aren’t.
I think it's because there's a financial motivation for all the toxic positivity that can be seen all over the internet. A lot of people put large quantities of money into AI-related stocks and to them any criticism is a direct attack on their wealth. It's no different from crypobros who put their kids' entire college fund into some failed and useless project and now they need that project to succeed or else it's all over.
I’m not sure that really explains how people get onto hype trains like this in the first place, though. I doubt many people intentionally stake their livelihoods on a solution in search of a problem.
My guess is that it’s more of a recency bias sort of thing: it’s quite easy to assume that a newer way of solving a problem is superior to existing ways simply because it’s new. And also, of course, newfangled things naturally attract investment capital because everyone implicitly knows it’s hard to sell someone a thing they already have and don’t need more of.
It’s not just tech. For example, many people in the USA believe that the ease of getting new drugs approved by the FDA is a reason why the US’s health care system is superior to others, and want to make it even easier to get drugs approved. But research indicates the opposite: within a drug class, newer drugs tend to be less effective and have worse side effects than older ones. But new drugs are definitely much more expensive because their period of government-granted monopoly hasn’t expired yet. And so, contrary to what recency bias leads us to believe, this more conservative approach to drug approval is actually one of the reasons why other countries have better health care outcomes at lower cost.
Currently if someone posts here (or in similar forums elsewhere) there is a convention that they should disclose if they comment on a story related to where they work. It would be nice if the same convention existed for anyone who had more than say, ten thousand dollars directly invested in a company/technology (outside of index funds/pensions/etc).
A browser plugin that showed the stock portfolios of the HN commenter (and article-flagger) next to each post would be absolutely amazing, and would probably not surprise us even a little.
Most of the announcements I hear about Copilot, it's always how they've integrated it into some other piece of software or cut a deal with yet another vendor to add it to that vendors product offering. On the surface there's nothing wrong with doing that but that just seems to be the ONLY thing Microsoft is focused on.
Worse yet, most of these integrations seem like a exercise in ticking boxes rather than actually thinking through how integrating Copilot into a product will actually improve user experience. A great example was someone mentioned that Copilot was now integrated into the terminal app but beyond an icon + a chat window, there is zero integration.
Overall, MS just reeks of an organization that is cares more about numbers on a dashboard and pretty reports than they are on what users are actually experiencing.
reply