Agile, Interviews, Software Development

“Functional metrics remain key in the age of AI” A conversation with Luigi Buglione

Jun, 13, 2025 | Read 5 min.

On May 16, during the 1st Metrics Event organized by GUFPI-ISMA in Rome—an event we had the pleasure of sponsoring—we sat down with one of the world’s leading experts in software measurement: Luigi Buglione. President of GUFPI-ISMA, Secretary of IFPUG, Academic Director and Vice President of ISBSG, Luigi has dedicated his career to promoting standards such as IFPUG Function Points and SNAP to improve estimation, productivity, and data-driven decision-making.

In this interview, we discussed the role of artificial intelligence in software development and estimation, the current state of Agile, the importance of measuring to improve, and how to approach the future from a quantitative and human perspective.

Interview made by Julián Gómez (CDAIO in LedaMC and Quanter). In Italian, subtitles available in English and Spanish.

Luigi Buglione: Thank you, Julián. Thanks to Quanter Smart AI Estimation.

LB: It definitely still has a lot of value to offer. What we need is to improve our skills, especially when it comes to measuring sprints and interactions.

We have to stop estimating based only on experience, using story points or t-shirt sizing, and start introducing functional and non-functional measurement, but with quantitative metrics.

Using Function Points, SNAP Points, or other ISO-compliant non-functional metrics can certainly help.

Without that, estimates in person-days remain highly variable.

LB: Mainly two things:

  • They can’t quantify the value of an asset (what isn’t measured doesn’t exist, only time worked does).
  • And more importantly, from the standpoint of organizations like ISBSG, there’s the issue of productivity.

If you don’t have a size divided by time worked, you can’t calculate productivity. You’re left with effort measured in story points or based on subjective estimates.
And without historical data, you start from scratch every time. In Italian, we say it’s like having the memory of a goldfish.

LB: That they don’t collect data. Particularly project managers, who should have studied this in guides like PMBOK, where the first thing you do in a project closing phase is collect data.

In Agile, this would be a retrospective.

The problem is that it often doesn’t happen—people say there’s no time for it. Which really means it wasn’t planned.

So that’s often the first area for improvement.

LB: The key is understanding the mindset of those who say “I don’t need them” or “They’re useless”.

In my experience, many such people are managers who don’t want to dive deep and focus only on short-term profitability.

But if you start to realize that you can’t say how long something will take without first knowing how big it is, and that you need an objective measure with very low subjectivity, then any number becomes significant.

It’s like in The Hitchhiker’s Guide to the Galaxy, where the magic number is 42.

LB: Absolutely yes, and for a very simple reason: If you don’t know how much something is worth—its quantity—you can’t know how long it will take.

And if you don’t know that, or who’s going to work on it, you don’t know how much it will cost, nor what price to charge if you’re a supplier. Or how much to pay if you’re a customer.

We call it the QTC flow: from quantities to time, and from time to cost.

Another thing we recommend to both the metrics and non-metrics communities is to use measurements to determine the correct time, considering the assigned team’s cost mix.

Consider the skill mix, the average daily rate, and then set a market price. That’s what we call the ABC model.

LB: Obviously! Having good prompt engineering practices is essential.

Today, even though it’s not recorded in this interview, we generated songs with AI. We created multiple versions using prompts with fewer than 10 words—and the effect, according to the audience, was fun, enjoyable, and effective.

Generating and structuring requirements is becoming a key asset for any business analyst who, with metrics skills, can use AI to accelerate timelines—as long as they understand how to interact with the machine.

As someone said in one of today’s talks: it’s about human-AI interaction, not just human-machine interaction. And that’s clearly the way forward.

LB: As mentioned in many talks today and in past events, we need to stay aware of AI hallucinations.

At ISBSG, we’ve proposed the hashtag #datahallucination.

It’s about making sure that unverified values for productivity, size, or effort don’t make it into tenders or technical specs if they don’t align with real-world conditions.
Training machine learning models also has a cost, which is why the full lifecycle needs to be reconsidered.

We must shift costs toward earlier phases, reducing them later during implementation.
We’re talking about cost redistribution, not small savings—that’s clear.

Also, as with any tool, AI shouldn’t be seen as a substitute for humans, but rather as a tool that remains under human control.

LB: Functional size measurement dates back to 1979 with Albrecht and is still a process metric, meaning it’s technology independent.

The mistake is to think that productivity is solely tied to size, when in fact it’s about the relationship between size and time.

So measuring functionality will continue to be a stable process.

What will change is the technique: IFPUG, COSMIC, NESMA, FISMA, Mark II…, but the approach will remain.

What will evolve, even within IFPUG, is that while developing and updating SNAP, we’ll need to consider more and more the quality of data, as defined in ISO 25012 (not yet strictly included in the model), and also ISO 25059.

So we’ll likely see evolutions within the SNAP model.

LB: That will always depend on costs and the cost-benefit ratio.

At first, some companies will invest more to create tools that, in a way, act as assistants to other stakeholders.

And it will largely depend on the volume of data and the cost to train a system that can be resold at the right price.

The challenge will always be in verifying the correctness of the input data—which isn’t about code quality or AI quality (which is still software), but rather data quality, which will still need to be verified manually.

LB: Inter. May 31, Monaco.

LB: Thank you all.

About the author

| | | |

Back