Category Archives: rants

When It Comes To Gen-AI, I’m NOT Yelling Enough! Part I

Deep dive into the comments of this LinkedIn post and you’ll see a comment that we should stop yelling at the tools. I strongly disagree!

As per a previous post, until the space is ready to admit that

  • Gen-AI/LLMs are not the be-all and end-all, having very limited uses
  • real progress still requires real blood, sweat, elbow grease, and tears
  • you can’t replace people as this tech is NOT intelligent

and, more importantly

  • that these tools are not what people need and
  • these tools cannot be used as the foundation for suitable solutions (although they can be [a small] part of those solutions if care is taken)

We need to keep yelling, and do so rather loudly.

Because, to build on the metaphor, it’s not a shiny new hammer. If it was just a shiny new hammer, we could depend on one of three things happening when we use the hammer to hit the nail:

  1. the nail goes some distance into the wood, depending on how hard we swing,
  2. the nail doesn’t go, because the hammer is too light, or
  3. if the handle is weak or the head not securely attached and we hit really hard and the nail doesn’t go in, in the absolute worst case the handle will crack or the head will fall off.

However, with the fancy new hammer equivalent of Gen-AI, we also have to worry about the possibility that:

  1. the hammer is super magnetized and pulls the nail out on the backswing,
  2. the hammer splits the nail in half,
  3. the hammer super heats the nail and melts it, or
  4. the hammer is packed with C4 and explodes, ripping our arm off our body!

Because, when you use Gen-AI, you accept the possible side effects of hallucinations, decreased code/application security, bad math, fraud, lawsuits, deadly diets, extremist views, sleeper behaviour, dependency and cognitive reduction, suicide, blackmail, hit lists, and murder, with many links summarized in this LinkedIn post.

And the worst part is this technology is being shoved into every nook and cranny, even those where we have technology that has worked great for over a decade (because the new generation of college-dropout script kiddies who believe that they can prompt engineer a solution to anything don’t even know the basics anymore).

It’s not just not solving our problems, it’s creating new ones, and they are often worse than the problems we have. We need to yell about this!

A Shiny New SaaS or AI Wrapper Doesn’t Make Tech Any Better

Just like painting a hammer bright shiny pink doesn’t change it’s fundamental function, putting a new shiny SaaS wrapper on a traditional desktop application or adding a Gen-AI interface to allow for a “conversational” interaction doesn’t fundamentally change what the application can do.

What an application can do depends upon the data model it can support, the core algorithms that process that data, and the workflows that connect them together to take raw inputs and produce necessary outputs. If the data model is not sufficient, the algorithms not appropriate, and the workflow lacking, a shiny new wrapper won’t change anything … the software will be no more effective than the software that is being replaced.

Pick any significant application, and the best results usually depend on intense or complex calculations, using a proper algorithm that works on a proper model populated by the right inputs, and if any piece is missing, the solution doesn’t work. In our area, it’s Source to Pay, and that starts with sourcing. In sourcing, the right decision is that which results not in the lowest bid, but the lowest lifecycle cost of the purchase, which takes into account not just unit costs, and not just shipping and tariffs and interim warehousing costs for landed costs, but also utilization/waste costs, local warehousing and inventory costs, (amortized) service costs, disposal costs, and even carbon costs if they vary by option. It considers all of the available product/SKU options, plants, shipping routes, and localized plant/warehouse/store needs and uses optimization and analytics to identify the optimal award that minimizes the overall cost while maintaining service levels and minimizing risk. If the solution doesn’t allow you to build the right models, collect all the options, identify the plants and routes, and determine optimal mixes that meet your criteria, then it’s not a modern sourcing solution no matter how SaaSy it is, how new it is, or how much BS Gen-AI gets shoved into it. A good application solves your core problem. If it doesn’t do that, it’s not good. And at the end of the day, it doesn’t matter how slick and SaaSy it is, because if the only application that gets it right is a green screen desktop application, then that is the best solution to your problem. (We hope it’s not — but given how little there is behind many of these SaaS apps, which are built to look good by developers with little to no knowledge of the domain they think they can satisfy with simple algorithms, and sometimes just fancy interfaces to a classic desktop application wrapped in a web container which slaps on a web-friendly API interface to the classic app and classic algorithm — we can’t say it’s not going to be the case that you have to keep using that decades old green screen application.)

At the end of the day, it’s algorithms that work, and the reality is that these are often the algorithms that were developed decades ago by leading minds, stress tested and sharpened by brilliant minds, proven to work, and just waiting for the computing power to catch up to where they need it in order to shine. (The best data structures and algorithms text book ever written is over 35 years old. Most of the revolutionary developments were between the 70s and 90s.) MILP is decades old, but we really didn’t have the computing power to solve large, complex, real world models until about two decades ago (and then only if you didn’t mind waiting a few hours to a few days for a scenario to solve). But now we can solve them in minutes, if not seconds, and that allows for next-generation strategic analysis and planning, as long as you have a modern platform that uses a modern algorithm that can take advantage of multi-core cloud processing capabilities, the right data model, and the data inputs you need.

And therein lies the hitch — it all comes down to the data model, algorithm, and application design — not the UX, the intake and orchestration, or the “conversational” Gen-AI interface.

Remember this the next time someone tries to sell you a shiny new interface or an upgrade to what you have. Remember that most upgrades are because software stacks change, functionality that should have been in the last release is finally added (since many SaaS companies now release untested alphas), or major security or performance issues are resolved. Now, you need the fixes for sure, but you shouldn’t be paying any more than the maintenance fee for those. If the buyer rolls them in “functionality updates”, you should insist you get those for free. If you got buy without the missing functionality (either because you had complementary systems or added it yourself), then do you really need more untested functionality now?

And at the end of the day, the primary reason software stacks change is that if they didn’t, you’d have to buy a lot less tech, and then the investors wouldn’t make money. Not all tech stacks offer significant improvements in functionality or even security. They just allow developers to work on the new hotness and enterprises to force you into spending more money, without any guarantee of more value in what you’re delivered.

So don’t get fooled by new tech. Do your homework. Sometimes the best tech is the old busted hotness.

P.S. Yes, Joel the number 666 is ruining Procurement*, but not necessarily, or just, in the way you appear to believe it is.

* see the Mega Map

Chief Sustainability Officer: USA Edition

A version of the graphic below has been making the rounds on LinkedIn for a few months (and the doctor wishes he could point to the original source of this [on LinkedIn], but either Google mis-indexed it [as the link goes to a user’s profile page] or it’s gone), and a more recent version can be found in this post.

These are great … if you are based in the EU. However, they are not so great if you are based in the USA, as outlined in our first quarter post on how in the corporate world, sustainability/ESG is NOT a priority. So, the doctor decided to correct it for you if you are based in the USA. Enjoy!

Gen-AI is Bad for Consulting Firms … But Even Worse For You When the Consulting Firms Blindly Use It!

A recent post on LinkedIn noted how there’s a wave of AI products flooding the consultancy and advisory space and how they are, frankly mediocre, overpriced wrappers on public models with minimum innovation, if any.

This is sad, but true, and it’s not the worst of it. The worst of it is that some of the Big X firms are training tens of thousands of consultants and f6ckw@ds on these tools to generate hundred page pitch decks and three hundred page strategy and implementation guides of standard generic, meaningless, drivel to deliver to you as “highly tailored guidance and expertise from their leading partners with 20 years experience delivering high-value projects” and charge you tens of thousands of dollars for the privilege.

This is especially egregious when you can use free/cheap (and I’m talking put it on your personal credit card cheap because you won’t notice the fee that is less than your monthly coffee charge from the coffee shop) to build the exact same pitches, strategy, and implementation guides from the thousands of freely available documents on the web in a few hours with a few generic prompts over a Sunday morning coffee. (And then, when the coffee kicks in, realize it’s all a load of cr@p and put in the bit bucket, but at least you will know what a load of cr@p looks like in pitch deck, strategy guide, and implementation plan form and will recognize it the next time an overpriced Big X tries to sell it to you for a ridiculous price tag and will have learned something from the exercise.)

Now that there are companies selling overpriced “custom” products to these consultancies, the situation is only getting worse, especially when the “customization” is just a wrapper with some pre-engineered prompts that aren’t well tested, only work at a point in time, don’t really give the consultancies what they need, and sometimes translate mediocre inputs to inputs that are even worse. Moreover, when you consider the price is sometimes a 100X multiple on the products they build on top of, it’s disgusting. Consultancies are paying more for less, and, in return, you are paying even more for even less!

Which makes no sense when the current publicly available LLM tech is being offered cheap (to try and hook you on it, even though, as we’ve repeatedly explained, the tech is not ready for prime time and will never deliver more than a fraction of what they are promising), and new implementations will get a lot cheaper. Just look at how DeepSeek undercuts the cost by a factor of 100 and gets 90% of ChatGPT (as long as you don’t mind exposing all of your secrets to the CCP). LLMs are nothing more than a fancy next-gen “deep learning” Neural Networks that construct responses vs. serving up canned responses (which is why hallucinations and lies are a core function, not an error that can be trained out) which gets us closer (but no cigar) to decent natural language processing (NLP) for the express purpose of the generation of desired outputs from inputs, but not there (and now, in addition to all the false positives and false negatives, we had to deal with, we now get to deal with hallucinations and lies as well). It’s not secret magic, it’s layers and layers of interconnected statistics and probabilities that no human can understand, in rather standard models that any Theoretical CS and Applied Math PhDs can build, and implementations that are better and cheaper are going to keep appearing as time goes on.

This means three things to any consultancy thinking about using these custom “AI” solutions

  • you still have to be even more tech savvy to use them to any degree of effectiveness
  • it’s not “the art of the prompt“, it’s the art of the training (even though they don’t really learn because they are NOT intelligent) because that determines the maximum level of effectiveness you will ever reach with them (and you need to provide them with sufficient correct data, which needs to be in the high gigabytes at a minimum, and, preferably, in the petabytes)
  • you don’t have to worry about when they are right (enough), which will happen between 90% and 95% of the time with proper training and proper prompting, or when they are obviously wrong, which will happen a very low percentage of the time (say 5% to 9%), but when they are oh so wrong but the response is constructed in a way that is oh so convincing that an above average person in intellect and experience wouldn’t know otherwise (that danger zone between obviously wrong and good enough that is likely only 1% to 2% of the time).

Now remember that your consultants aren’t that tech savvy, and you should know right off the bat incorporating and using these is going to be difficult and time consuming. (There’s a reason we are constantly advising you to be very careful about using Big X for tech selection and tech projects, and that’s because, even though they say it is, it’s NOT their forte. They weren’t built on tech, and they don’t have the best talent in tech — that talent goes to the big tech companies who can offer the 500K salaries to leading devs or the wild-west startups that leading devs think are cool.)

You only have so much clean and complete data you can use for training. You can’t just throw in the 1000s of decks you’ve built as you can’t share work you’ve explicitly created and sold to past clients, and the AI won’t anonymize the decks and suggestions (even though you think it will). It won’t know that “Ford” is the name of your client and might think that “Ford Data” is another term for shallow data and copy sections from that custom strategy straight into your pitch deck for General Motors (and chances are your overworked junior consultant won’t catch it when skimming that 200 page deck with only 2 hours to go before the meeting). And we know what happens then … (and it ends with the consultancy not keeping either client).

It will take a lot of analysis to identify those 1% to 2% of cases where it is very, very wrong but so convincingly right that you will miss some. What happens when you do and give your client advice that explodes in their faces? (We’ll let you answer that one.)

And for you as a consumer, if your consultancy is using this Bogus AI tech, it means that:

  • the situation that results from solution delivered might be even worse than the situation you started with (as should be evidenced not just by the tech project failure rate that is approaching 92% but the fact that 42% of projects are being abandoned during implementation!)

A solution designed by Gen-AI is not a solution. A real solution is a solution designed by human intelligence that uses real, augmented intelligence, to research and validate that solution. Remember that if you are going to hire a consultant!

When Someone Says “Real AI”, Ask For Details!

We shouldn’t have to remind you, but since too many people are falling for, and buying into, the hype and selecting tech that does not, and can not, ever,work, we are going to remind you yet again.

Computers do NOT think!

To think is to direct one’s mind … where one is an intelligent being, not a dumb box. Computers thunk … they compute using algorithms (which are hopefully advanced and encapsulate expert guidance and knowledge, but that is far from guaranteed).

Computers do NOT learn.

Appropriately selected and implemented probabilistic / statistical / machine learning algorithms will improve their performance over time as more data becomes available, but they do not learn. Learn is to acquire knowledge (or skill), and by definition, knowledge can only be acquired by an intelligent being.

Computer Programs Can Adapt …

but there’s no guarantee the adaption is going to improve their performance under your definition, or even maintain their performance. Their performance could actually decrease over time.

What is critically important is that there are two primary types of algorithms that can be used to create an AI application:

Deterministic and Probabilistic

A deterministic algorithm is one that, by definition, given a particular input will, no matter what, always produce the same output, with the underlying machine always passing through the same sequence of states. As long as you don’t screw up the input, or the retrieval of the output, (and, of course, the hardware doesn’t fail), it is 100% reliable.

A probabilistic algorithm, in comparison, is an algorithm that incorporates randomness or unpredictability into its execution, and may or may not produce the same output given successive iterations of the same input. Nor is there even any guarantee that the algorithm will produce a correct, or even an acceptable, input a given percentage of the time. Well designed, these algorithms may allow for consistently faster computation, better identification of edge cases, or even a lower chance of error, on average, for a certain class of inputs (but with the caveat that other classes of inputs may suffer a higher error rate).

Deterministic algorithms can be relied on to execute certain tasks and functions autonomously with no oversight and no worry. Probabilistic cannot. In other words, you cannot assign a probabilistic algorithm a task for autonomous computation unless you can live with the worst possible outcome of the algorithm getting it wrong. And this is what Gen-AI, and most of today’s “AI” tech, is based on.

This is the critical problem with today’s AI-tech and AI-Hype. Especially when a probabilistic system can, by definition, use any method it likes to determine a probability (which may or may not be at all appropriate, since a model is only valid if it accurately captures the “population” dynamics) and may, or may not, be accurate. For some of these situations, it will be the case that neither the company nor the provider of the system will have enough historical data (market situation and outcome) to even attempt to make a reasonable prediction, and there definitely won’t be enough data to know the accuracy, because standard measures of model accuracy (like the Brier Score), tend to require a lot of data, especially if you have a situation where you need to accurately identify rare events as this could require 1,000 or more “data points” (which, in a typical market scenario, would require enough data to identify the market condition and then the unexpected change”).

(And this is exacerbated by the reality that, for many of these situations, one could likely employ more traditional “statistical techniques” like trend analysis, clustering, classical machine learning, etc. to solve much of the problem at hand.)

It’s important to remember that Gen-AI LLMs, which power most of the new (fake) agentic tech, are all probabilistic based (and designed in such a way that hallucinations are a core function that CAN NOT be eliminated), and much of it is complete and utter garbage for what it was designed for, and even worse for tasks it wasn’t defined for (like math and complex analyses). (Everyday we see a new example of complete and utter failure, often due to hallucinations, of this tech. For example, you can’t even get a list of real books out of it — as per a recent contribution to the Chicago Sun Times which which published its Summer Reading List of 15 books, of which only 5 of which actually exist. And then there are numerous examples of lazy lawyers getting raked over the coals by judges for using ChatGPT to do their homework and quoting fake cases!)

While we do need to augment purely deterministic tech with more adaptive tech that uses the best “statistical techniques” to more quickly adapt to situations, we need to spell out the techniques and restrict ourselves to what is now “classic machine learning” where the algorithms have been well researched and stress tested over decades (not modern Gen-AI powered agentic tech that has worse odds than your local casino). At least then we’ll have confidence and can enforce bounds on what the solution can actually do (to limit any potential damage).

Especially now that we finally have the computing power we need to effectively use tried-and-true “classic” ML/AI techniques that require large data stores and huge processing power for highly accurate predictions. The reality is that even though this tech has existed for at least 25 years, the computing power required made it totally impractical for all but the most critical situations. Twenty-five years ago, a large Strategic Sourcing Decision Optimization (SSDO) model would run all weekend. Today you can solve it in a few seconds on a large rack server (with 64 cores, GB of cache, and high-speed access to TB of storage). The fact that we finally have (near) real time capability means that this tech is not only finally usable in all situations, but finally effective.

[And if vendors actually hired real computer scientists, applied mathematicians, and engineers and built more of this tech, instead of script kiddies cobbling together LLMs they don’t understand, we would be a decade ahead of where we are today.]