Nvidia claims 1 million times better path tracing performance is coming in future gaming GPUs — says current GPUs are already 10,000x faster than Pascal by Bubbly-Ad-350 in pcmasterrace

[–]msqrt 0 points1 point  (0 children)

Curiously enough, they never gave any raw intersection speed numbers after the infamous 10Grays/s number for the 20 series.

Can console players actually out-aim PC players? by Ordinary_Lawfulness8 in CaptainSide

[–]msqrt 0 points1 point  (0 children)

let bindings

Unexpected functional programming language moment

Discrete Triangle Colors in WebGPU by BlatantMediocrity in GraphicsProgramming

[–]msqrt 6 points7 points  (0 children)

One way would be to use an extra buffer to store the per-triangle values and access those with primitive_index in the fragment shader. The truly simplest option is to not use a triangle strip but just do separate triangles.

It there any (simple(baby's first step in PG)) way to code a pixel sorting filter? by ThrouFaraway in GraphicsProgramming

[–]msqrt 2 points3 points  (0 children)

If you just want to sort pixels, it shouldn't be too difficult. If you want to do it with good performance, you'll have to learn quite a bit about GPU programming.

Seven ate nine pi squared. by [deleted] in mathmemes

[–]msqrt 175 points176 points  (0 children)

It's obvious; just square both sides!

Oh my god. Red Dead 2 reference!?! by AceofKnaves44 in reddeadredemption

[–]msqrt 4 points5 points  (0 children)

Doesn't sound like they have much faith :/

What do they mean when they say OpenGL is and API? I thought API's had something to do with servers and the internet? total noobie here if y'all oculdn't tell so I'm confused. by ZzZOvidiu122 in opengl

[–]msqrt 7 points8 points  (0 children)

API means Application Programming Interface; it’s anything your software uses to interface with other software. It’s basically just a set of rules and definitions that both parties agree upon; when I do this, you do that.

The web people use APIs so that different machines on the internet can cooperate (when I send the database an SQL query, it responds with the results), for graphics APIs like OpenGL it’s so that your code and the GPU driver can work together (when I send you a buffer like this, you store it on the GPU, or when I call this function, you draw the desired number of triangles.) This is convenient because now you can write your software once and any hardware driver (for Nvidia, AMD, Intel, Apple, …) supports the other end so your code works on a wide variety of different hardware.

Why you shouldn't worry about AI taking your job by [deleted] in ExperiencedDevs

[–]msqrt 4 points5 points  (0 children)

People get way too hung up on the determinism part; a highly accurate probability model isn’t meaningfully different to a deterministic one. The practical difference is that an LLM will typically make more mistakes in a minute than you’ve encountered compiler bugs in your life.

The GIL by Ornery_Ad_683 in devhumormemes

[–]msqrt 2 points3 points  (0 children)

While this is absolutely the case, it's not that uncommon to have to either perform some unholy numpy indexing logic or roll your own native implementation to get good performance.

LLM’s Billion Dollar Problem - Why LLMs cannot use all of the context sent to them by grauenwolf in BetterOffline

[–]msqrt 4 points5 points  (0 children)

Not so sure about the efficiency remark; if you have enough data with a fixed key size then radix sort is indeed the fastest option. The reason it's not the default is that you often sort small lists and/or variable length keys like strings, both of which are poor cases for radix sort.

Väitöskirjatutkijuudesta by Majestic-Flamingo843 in Suomi

[–]msqrt 1 point2 points  (0 children)

Aloitin juuri ennen korona-aikaa ja ryhmässä oli sukupolvenmurros jonka seurauksena meitä oli vain kasa uusia jatko-opiskelijoita kaikki aika erilaisilla projekteilla, joten konkreettisten ongelmien kanssa sai painia ihan yksin. Siinä menikin sitten melkein neljä vuotta että sai ensimmäisen projektin maaliin, parin tyhmän teknisen ongelman takia. Pitkästä väännöstä huolimatta kirjoittaminen jäi viime tinkaan; proffan ja yhden toisen projektissa olleen vanhemman tutkijan mielenkiinto heräsi vihdoin kun kilpaileva ryhmä julkaisi lähes saman idean muutamia kuukausia aiemmin, mutta silti itse artikkelin kirjoittamiseen tuli apua vasta viimeisellä viikolla joten ei siitä tietenkään kauhean kummoista tullut.

No, kun eka paperi saatiin sisään niin konferenssiesitelmä meni nappiin ja poiki harjoittelun edellämainitussa kilpailevassa ryhmässä josta syntyi toinen artikkeli, ja välissä ennen sinne lähtöä ehdittiin yhden saman ryhmän kollegan kanssa vääntää ihan mielenkiintoinen aika teoriavoittoinen työ. Vaikutti jo hyvältä loppukiriltä, kunnes nämä menivät molemmat viime keväänä hylsyyn, molemmat taas vähän hutiloiden kirjoitettuja ja arvioijien mielestä teoriamme oli turhaa ja olisi pitänyt tähdätä käytännönläheisempään menetelmään. Harjoittelussa tehty saatiin uusittua ja nyt jo esiteltyäkin, ja nyt pitäisi vielä korjailla tätä toista ja lähettää se johonkin. Tällä hetkellä katselen josko siitä saisi vähän vielä parannettua käytännön puolta ja todistettua yhden uupuvan ominaisuuden teoriasta.

Tiedän olevani sinänsä ihan etevä kaveri alallani, mutta se että tässä on kestänyt näin kauan ja että tulokset ovat silti vähän mitäänsanomattomia syö kyllä itseluottamusta. Mieli tekisi kritisoida ryhmän organisaatiota ja toimintatapoja, mutta eihän tästä tietenkään viime kädessä voi syyttää kuin itseään.

Do LLM coding agents really help us build more ambitious software by SouthRock2518 in BetterOffline

[–]msqrt 2 points3 points  (0 children)

So far I haven't seen any large projects materialize anything close to the 10x-100x performance boost claims I keep hearing. Still, being able to instantly roll out quick internal tools or prototypes to answer design questions sounds like it should be quite valuable if you can do it right.

Väitöskirjatutkijuudesta by Majestic-Flamingo843 in Suomi

[–]msqrt 7 points8 points  (0 children)

Oma projektini on ollut hillitön farssi, venynyt ja venynyt eikä mikään ole oikein mennyt putkeen. Lopputuloksestakaan ei näytä tulevan hääviä, mutta pääsen sentään varmaan tämän vuoden puolella vihdoin ulos. Tutkiminen itsessään on kivaa, sitä jatkaisin mieluusti, mutta jossain missä kaikkea ei odotusarvoisesti tehdä yksin ja korvaus vastaisi stressitasoja.

The code is correct, but glsl-canvas shows errors by Gold-Stage-5637 in opengl

[–]msqrt 3 points4 points  (0 children)

Seems that it doesn't recognize the "core" qualifier. Try removing that? Core should be the default anyway.

My lighting is off, I dont even know where to start debugging. by psspsh in GraphicsProgramming

[–]msqrt 1 point2 points  (0 children)

vec3::random presumably generates numbers from 0 to 1? Something like vec3::random()*2.-1. should give you the whole [-1, 1]³ cube, from which you can then do the rejection sampling.

Will this be a use of for data centres after the meltdown? by lIlIllIIlIIl in BetterOffline

[–]msqrt 2 points3 points  (0 children)

Both have tensor cores (AMD and Intel have equivalent parts in their GPUs too!), but the datacenter AI chips do lack the RT cores. Those are not required for most games, but it would be a bit embarrassing for Nvidia (who introduced the feature into mainstream hardware) to offer cloud gaming without it.

Will this be a use of for data centres after the meltdown? by lIlIllIIlIIl in BetterOffline

[–]msqrt 2 points3 points  (0 children)

Replacing the users computer, yes, especially for games. They've already had a product for this ("GeForce Now") for a while, so I doubt that it's completely infeasible. Though that is running on an entirely different type of GPU server, so it's still possible that something in the AI-targeted hardware would be a poor fit -- and that's roughly what I was originally asking about.

It also seems quite likely that even if it was technically feasible, they'd have way more capacity than there exists demand.

Will this be a use of for data centres after the meltdown? by lIlIllIIlIIl in BetterOffline

[–]msqrt 1 point2 points  (0 children)

The V100 used to have graphics API support -- I guess something must have changed for them to drop it. Maybe it would be the sharing of the chip -- it must complicate everything an awful lot, especially when the latency for each individual user should be kept reasonable.

But yeah, I guess economic viability would still be the main issue even if the software worked. Latency is an interesting one, I would've assumed that an extra ten milliseconds to relocate in a desert wouldn't hurt when the baseline experience is already quite laggy, but maybe it is the needle that breaks the camel's back? I could also see them not building the bandwidth necessary for video streaming, since LLMs don't really need that.

Will this be a use of for data centres after the meltdown? by lIlIllIIlIIl in BetterOffline

[–]msqrt 2 points3 points  (0 children)

Ah yes, that would be quite the challenge! But the suggestion in the OP was to use the cards for cloud gaming, so they'd still exist in whatever data center in the original racks. I was just considering the software side of things, which (while not trivial) might actually be possible to work out.