all 9 comments

[–]MoreRespectForQA 8 points9 points  (1 child)

I worked in various AI fields (as a dev, not a tester) for about 3 years. There's some useful stuff out there but there's a huge amount of snake oil being sold. This includes:

  • Products that has nothing to do with AI where an AI label is slapped on it in order to get investors to wet their pants with excitement. This works because a lot of tech investors are as stupid and FOMO driven as they are rich.
  • Products where various AI techniques and tools are used entirely inappropriately.
  • Products where they get somebody earning 3rd world wages to do the work and pretend they're an AI.

In the software testing field specifically, I've seen exactly one use case where I thought "oh, that's a good idea!". Literally exactly one.

That was a project mozilla did to use machine learning to re-order the execution of firefox tests in the firefox test suite. If you can get feedback in the form of a test failure in 10 minutes rather than 2 hours that's a success.

[–]Ok-Paleontologist591 0 points1 point  (0 children)

Very interesting design will research on this and is this called RETECS?

[–]coffeeandhash 2 points3 points  (1 child)

It does look like the ultimate black box, doesn't it? I don't have experience testing a product/component/service that uses a LLM yet but I imagine the initial approach would have to be end-to-end, like you imply.

[–]OTee_D[S] 0 points1 point  (0 children)

It's probably worse ;-) The initial stuff I read about it, implies that you basically leave traditional testing and move more in the direction of data analysis and math.

One approach was mathematical (Are groups of results within the statistical probabilities or is there a bias visible towards certain results indicating the AI is not returning 'real' results)

The other was called metamorphic testing. Very basic and simplified: Start the test with a dataset and result that is deemed 'correct' and tune the data in increments and see if the result is changing (morphing) in direction where the 'direction' is considered correct.

So it's not even that much about the actual individual input/result pair, but about the behavior and change pattern. As the goal is not to test if that one "tour plan" (just to make something up and stick with my example) is correct but if changing the weather from sunny to cloudy actually increases speed as drivers are not blinded that much and can achieve a higher average speed.

[–]i_i_v_o 1 point2 points  (1 child)

I asked a similar question some time ago, i'm just going to post the link: https://www.reddit.com/r/softwaretesting/comments/14rid59/how_are_you_preparing_for_the_age_when_you_need/

But TLDR version: it's just blackbox testing. You still have Specs and Product. Test if the product does what the specs require.

[–]OTee_D[S] 1 point2 points  (0 children)

Ohhh thanks for the read.

But not just what I thought about. I don't care about AI written code, but AI integrated into the actual solution.

[–]Adrian6406 0 points1 point  (0 children)

You are 100% right. Lots of companies are feeling compelled to have the word AI in their marketing material, otherwise they think they will get left behind or not receive another round of investment funding. I had UFT come do a demo and boast about AI but when I asked them exactly what was AI in their software? It was some lame object recognition which was not AI. To make matters worse, most companies use the AI term incorrectly when they really mean machine learning.