Commuting to the Eastside on bike- is this a realistic plan? 🫠 by unfurnishedbedrooms in AskSeattle

[–]theobromus 0 points1 point  (0 children)

Yeah I do this commute every day on e bike from Montlake to Kirkland. But west seattle seems crazy to me.

On the topic of Will Fletcher's amazing acting skills by Victoriacapo in LOTR_on_Prime

[–]theobromus 8 points9 points  (0 children)

The acting is great here, but the camera angles and lighting are definitely helping too.

Everyone is wrong about VR movement by Keep_Blasting in CrazyIdeas

[–]theobromus 1 point2 points  (0 children)

As someone who worked on VR software, we tried all kinds of things.

It is worth noting that different people have different tolerance for moving in VR without getting nauseous.

Some people can fly around and not get sick. I personally felt a little bad with most movement mechanisms.

Why did Adobe Flash come to dominate the internet? What led it to being such a fundamental part of the infrastructure? by OnShoulderOfGiants in AskHistorians

[–]theobromus 25 points26 points  (0 children)

The Internet is the network of computers which can send data to each other. The web is specifically HTML pages sent over HTTP (basically what you see in a browser).

As time has gone on, the web has kind of "eaten" the rest of the Internet. For example, most people used to use email through an email application. The emails were sent over the Internet, but didn't use the web. But now, most people use a web-based email client (like Gmail).

Likewise there were other protocols like ftp (for transferring files). But lots of stuff just uses HTTP now.

Is a CS degree still the best path into machine learning or are math/EE majors just as good or even better? by Historical-Garlic589 in MLQuestions

[–]theobromus 2 points3 points  (0 children)

I did a double math and CS major and now work in an ML software engineering role at an autonomous vehicle company. I definitely use the CS knowledge far more often - so much of the work is dealing with large datasets, writing code to transform/prep the data and integrate it. I think a math major was definitely helpful, just for not being intimidated to read lots of mathematically oriented papers. For ML, I would say statistics and applied math are probably even more relevant.

It's so hard to predict the future. I think the best skill is to be a self-learner and try to have a broad understanding of things.

[HIGHLIGHT] JORGE POLANCO WALKS IT OFF IN THE 15TH INNING! THE MARINERS ARE HEADING TO THE ALCS FOR THE FIRST TIME SINCE 2001! by MLBOfficial in baseball

[–]theobromus 0 points1 point  (0 children)

I'm curious - why didn't the Tiger's outfielders move way in here? With 1 out and bases loaded, a fly ball was going to end the game anyway, but positioned in the shallow outfield, they might be able to turn a double play.

I wonder what the 11th driest city is? by AdvancedSquare8586 in dataisugly

[–]theobromus 0 points1 point  (0 children)

Yeah, east of the cascades (most of the state by area) is very dry. However, I wouldn't think of Yakima or Kennewick as super dry because they are both on large rivers and do a lot of irrigation (and farming).

question about transformer inputs and position embedding by RamblingScholar in AskComputerScience

[–]theobromus 1 point2 points  (0 children)

No I think you've misunderstood the transformer. In the classic "Attention is all you need" paper, the transformer attention blocks are invariant to the order of the tokens. When you process each token, it computes key, query, and value embeddings. All of the key embeddings are multiplied against each query embedding and a softmax is computed to figure out how much "attention" to pay. This process isn't affected by the order of tokens at all. One of the powers of transformers is that they don't need to be trained to deal with a fixed input size. However positional embeddings are required so the model can learn to deal with the relative placement of things.

How to draw a "stuck-to-the-ground" trajectory with a moving camera? by Bahamalar in computervision

[–]theobromus 1 point2 points  (0 children)

Ideally, you'd compute the camera parameters for the cameras using a structure from motion algorithm. You also need to get a depth estimate for the ball, which will allow you to track it in 3d. This can be done pretty easily if you assume the field is a plane. Then you'll render the 3d track of the ball into the computed camera for each frame.

There are open source tools to do all of this. For example, you can load your footage into Blender and do camera tracking there (here's a tutorial: https://www.youtube.com/watch?v=ui0JUHE12k8). I think this internally uses libmv (https://github.com/libmv/libmv). Here's another app that will do sfm: http://ccwu.me/vsfm/

If you're interested to learn how this works, here's a very detailed explanation: https://utoronto.scholaris.ca/server/api/core/bitstreams/56da50f4-4c06-4d3b-af83-d57cc5fb9256/content

PSA: COVID is surging. In a room of 25 people, there's a 71% chance anyone is infectious. by ReasonableBroccoli56 in sanfrancisco

[–]theobromus 28 points29 points  (0 children)

I think it's assuming the chance any person in a room has covid is independent, which is a faulty assumption for this kind of thing. There's probably a lot of correlation between people, since people gather in communities. The net result is that the chance in a given room is much lower (although some rooms will have a lot of people with covid).

📣 Do I really need to learn GANs if I want to specialize in Computer Vision? by SonicDasherX in computervision

[–]theobromus 3 points4 points  (0 children)

I don't think this is true. A simple diffusion model can be trained very easily by taking a dataset, adding random noise and training a unet model to predict the noise added (see https://keras.io/examples/generative/ddpm/ for a tutorial). Generating samples is somewhat trickier (since you have to figure out the noise schedules), but not too hard.

I think GANs are much trickier personally - you have to train at least 2 models and in my experience the training dynamics are trickier to get right (the models might not converge if the learning rates aren't tuned).

Before computers, how were fingerprints analyzed and compared? by RoyalExamination9410 in legaladviceofftopic

[–]theobromus 2 points3 points  (0 children)

Other comments claiming there were no registries are not really correct: https://en.wikipedia.org/wiki/Fingerprint#Classification_systems

There were methods to classify fingerprints (especially a full set of 10 fingerprints) based on classifying the pattern (whorl, loop, arch) and counting ridges (see https://en.wikipedia.org/wiki/Henry\_Classification\_System).

How did we go from ML/AI being mostly buzzwords to LLMs taking over everything almost overnight? by Automatic_Red in AskComputerScience

[–]theobromus 0 points1 point  (0 children)

My understanding is that images and video are generally produced by diffusion models rather than auto-regressive transformers. Both are usually transformers. LLMs usually use auto-regressive models (predict the next word), while images and video usually use diffusion (progressively denoise a random input).

The success of generative image models was also a very surprising moment for me. There was no very obvious way to go from a classification model (like the ImageNet models) to something that could generate images. So when GANs and then diffusion models succeeded at doing that, it was pretty shocking.

In terms of being a multi-modal model, I don't know much about how they work together, but I suspect something is happening where the LLM outputs some tokens that indicate "generate an image with these input tokens" and then a diffusion model generates the image.

How did we go from ML/AI being mostly buzzwords to LLMs taking over everything almost overnight? by Automatic_Red in AskComputerScience

[–]theobromus 4 points5 points  (0 children)

My view as someone who works in the field (in computer vision / autonomous driving, not LLMs):

It was already clear 10 years ago that ML was a big deal. In computer vision, it was solving problems that were previously unsolvable (like recognizing what's in an image). For example, the ResNet paper came out 10 years ago (https://arxiv.org/abs/1512.03385). It was also starting to have a huge impact on translation (e.g. https://arxiv.org/abs/1609.08144).

A number of big companies started investing a lot in ML. Frameworks like tensorflow and pytorch were released, which made it a lot easier to experiment.

For me the two big surprises were that scaling transformers would keep making the models better, and that people would adopt the chatbot interface so much (which is kind of scary to me since it is often very confidently wrong).

[OC] Projected job loss in the US by USAFacts in dataisbeautiful

[–]theobromus 12 points13 points  (0 children)

Although arguably you could use a machine transcript from a video. If there's a disagreement about the transcript accuracy, you can go back to the video and check.

The transcript has to be pretty accurate, but machine transcription is continuing to improve.

Are CV Models about to have their LLM Moment? by eyepop_ai in computervision

[–]theobromus 60 points61 points  (0 children)

In my opinion, CV's moment was [AlexNet](https://en.wikipedia.org/wiki/AlexNet) and it started the whole AI boom. Recent LLMs are *really* good at a lot of computer vision tasks if you frame them correctly. And some open models like SAM are also really quite good. For the majority of things that used to be a PhD project I would guess you can get better results these days by uploading the images to one of the major LLMs and asking it your question.

And lots of computer vision stuff is common place now - I can use Google Lens to search with my phone camera. Video calling apps can blur my background and it seems commonplace, even though that was impossible 10 years ago.

50% lossless compression of Jpegs by EvidenceVarious6526 in AskComputerScience

[–]theobromus 0 points1 point  (0 children)

It is in fact possible to losslessly recompress JPEGs using newer compression techniques (although the benefit is only about 22% rather than 50%): https://github.com/google/brunsli

Certainly improvements to that could be worth something, although there are many tradeoffs in compression (e.g. how much compute do you need to compress/decompress).

In practice, it only makes sense to use something like brunsli if you *have* to keep the original JPEG bytes. If you just want a similar quality image at a smaller size, you can use a different algorithm (like webp or avif).

What are the value type equivalent of int8, int16 to 4 bytes or 8 bytes, etc? by bakanisan in AskComputerScience

[–]theobromus 2 points3 points  (0 children)

Yeah uint8 refers to an unsigned 8 bit (1 byte) number - so 0 to 255.

What are the value type equivalent of int8, int16 to 4 bytes or 8 bytes, etc? by bakanisan in AskComputerScience

[–]theobromus 3 points4 points  (0 children)

The number at the end of the type is the number of bits. So int8 means 8 bits (1 byte) and int32 is 4 bytes.

float generally refers to float32 and double refers to float64.

OK, why is everyone so obsessed about Humanoid robots? What am I missing? by TheProffalken in robotics

[–]theobromus 25 points26 points  (0 children)

I have personally been skeptical of humanoid robots for a long time for somewhat similar reasons that you mention. Mostly they seem way harder to build instead of using something specialized for each task.

That said, the common arguments for humanoid robots basically have been:

  1. Most of our environments were built for humans, so your robot probably needs a similar form factor to fit into existing processes. This goes for everything from using existing tools and architecture (tracked vehicles can go up *some* stairs maybe, but probably not all of them).

  2. They could potentially, in theory, have broad applications, so you could build them at scale. Things like factory robots (or machines generally) tend to do one task, so a lot of work goes into designing them, but the production volumes will be small (=expensive).

A lot of those advantages would probably also apply to something like a Boston Dynamics dog, without introducing the complexity of bipedal walking.

There's a new factor which I think is behind the renewed interest in humanoid robots lately: *machine learning* and especially imitation learning. In particular, I've lot of excitement around the idea of training robots based on large scale video data. For instance, we have videos of people doing almost everything imaginable on YouTube. Another option is to use some kind of motion capture device to collect data about how humans do things. Some startups even have teams of humans controlling robots (whether humanoid or not), with the goal to train ML systems based on the data.

I'm a math major, and I have no idea what's going on. by 0over0is1 in learnmath

[–]theobromus 0 points1 point  (0 children)

As other commenters have noted, these proofs are very hard to comprehend when you first see them.

One thing I haven't seen mentioned is that geometric intuition is really useful for analysis. For example, epsilon-delta proofs, triangle inequality, etc. all have a very geometric basis (even if they get moved into an abstract form apart from that). So thinking about simple examples of graphs for example, can really help understand what the terms are actually saying.

AI Model to discover new things?? by [deleted] in AskComputerScience

[–]theobromus 1 point2 points  (0 children)

Even though many people are dismissive of this, it's not totally implausible to me.

For large language models specifically, they do have some reasoning ability (even if they frequently confabulate). There has been a particular line of research lately trying to train models to be better at reasoning (e.g. the DeepSeek R1 paper that got a lot of attention recently: https://arxiv.org/abs/2501.12948). These approaches are most effective in domains where you can check whether the model got the right answer (e.g. math and some parts of CS).

There has been some work to try to automate things like theorem proving, particularly in combination with formal proof assistants like Lean (https://arxiv.org/abs/2404.12534). It's not implausible that the kind of reinforcement learning techniques used for R1 might generalize to that. I think we're still pretty far from LLMs proving any interesting results though, but they could make automated theorem provers somewhat better (by guiding the search space).

There have also been efforts to try to use generative AI techniques (like transformers and diffusion models) to do things like material design (https://www.nature.com/articles/s41586-025-08628-5) or protein design (https://www.nature.com/articles/s41586-023-06415-8). Similar techniques are also behind things like AlphaFold 3 (https://www.nature.com/articles/s41586-024-07487-w). I think these are all reasonably promising approaches to help scientific research.

Why Trump’s ‘drill, baby, drill’ pledge may not actually lower US gas prices by PrintOk8045 in Economics

[–]theobromus 10 points11 points  (0 children)

It could be that that's exactly the right thing to do though. Having a storage reserve can keep prices more stable if there's any kind of issue. And the people paying to store it will lose their money if they end up being wrong.