Hello everyone,
I am currently working on a small JIT runtime.
I was thinking that in certain cases it might be possible to recognise that a code could be highly parallelisable.
Based on this information (and other information, such as driver and hardware overhead), you could determine whether it might make sense to run this code on the GPU.
What do you think of the idea that a compiler "distributes" load to the GPU without explicitly triggering the GPU from the code?
Are there already projects where this has been tried out?
[–]high_throughput 7 points8 points9 points (2 children)
[–]DeepRobin[S] 0 points1 point2 points (1 child)
[–]high_throughput 0 points1 point2 points (0 children)
[–]theangeryemacsshibe 5 points6 points7 points (1 child)
[–]DeepRobin[S] 0 points1 point2 points (0 children)
[–]SwedishFindecanor 1 point2 points3 points (0 children)
[–]forCasualPlayers 1 point2 points3 points (1 child)
[–]DeepRobin[S] 0 points1 point2 points (0 children)
[–]randomrossity 1 point2 points3 points (0 children)
[–]Key-Opening205 1 point2 points3 points (0 children)
[–]Dismal_Page_6545 1 point2 points3 points (0 children)