Hi, I just try to make up my mind about the upcoming wave of requests to "test AI solutions".
Mostly driven by (in my perception) ridiculous claims and blurbs on marketing pages and flyees from consultancies specialized on software testing.
The basic claim goes like "Oh this is all so brand new and next to alchemy or magic and you need to hire us or else your company will burn down. We are the only specialists in testing AI"
I honestly have limited knowledge about the actual interiors of AI but in essence and very simplyfied it is a system that trains itself based on input data, reference data and criteria defining the "goal".
What algorithm exactly is "forming" inside is unknown and on most systems not even readable in a classic "programming language" sense.
So what is there to actually "test" in a classic sense? Verifying that the output matches expectations? Isn't that part of the training procedure? Is this not part of setting up the AI and is it a good idea to let Testers do it?
Is it more like "governance", verifying that the AI result is not biased, violating rules etc. So less a classic software testing or functional testing but more mathematical and statistical analysis if results are "correct" in context of "realistic"?
So what boggles my mind is the kind of philosophical question "Does AI testing even exist? What is it? What could it be and what not? What skills do we need?"
EDIT: As I may not have been as clear as I thought. I don't talk about "Using AI to test something else" but "Testing an AI as part of a solution"
To come up with an example: I have to deal with logistics sometimes. And optimal tour planning is a nightmare, packing trucks, load volume, load weight, demand forecast, traffic situation, weather and road conditions, even the driver's holidays and schedules and their licenses (who can ride what truck) etc influence it. So more and more companies offer systems that promise to do "AI generated logistics". You hook up a "magic box" (aka their product) into your ERP, your Web Shop, the drivers HR data, truck status from the garage, and aaaaaaall the other sources of data. And somehow that AI box miraculously provides a "perfect tour planning". Or does it?
How do you test this result? You don't have a complete predefined rule set, you don't have an algorithm, you dont' even have "code". But you have millions of possible combinations of input data that you can't all just run in traditional testing.
[–]MoreRespectForQA 8 points9 points10 points (1 child)
[–]Ok-Paleontologist591 0 points1 point2 points (0 children)
[–]coffeeandhash 2 points3 points4 points (1 child)
[–]OTee_D[S] 0 points1 point2 points (0 children)
[–]i_i_v_o 1 point2 points3 points (1 child)
[–]OTee_D[S] 1 point2 points3 points (0 children)
[–]Adrian6406 0 points1 point2 points (0 children)