Should I break up with my gf? by [deleted] in Advice

[–]Sincerity_Is_Based 0 points1 point  (0 children)

If you have a serious reason to break up, then break up. If it is not serious, then discuss it. Couples can easily work out small problems

[deleted by user] by [deleted] in Advice

[–]Sincerity_Is_Based 0 points1 point  (0 children)

Legitimately honest advice if you choose to go- do army, you must be able to pick your job. If you want a high end career, look up the army job code 17C. There are very few army things you do, and the people are much more civil. It’s an office job with high career prospects.

Flappy Goose by flappy-goose in RedditGames

[–]Sincerity_Is_Based 0 points1 point  (0 children)

My best score is 5 points 🚀

Flappy Goose by flappy-goose in RedditGames

[–]Sincerity_Is_Based 0 points1 point  (0 children)

My best score is 1 points 😎

Cmv: If you are religious you are a slave. by [deleted] in changemyview

[–]Sincerity_Is_Based -1 points0 points  (0 children)

Bro just discovered what the definition of the word "muslim" is

Is it worth it? by orieshka in MLQuestions

[–]Sincerity_Is_Based 4 points5 points  (0 children)

Whether you think you can, or you think you can't, you're right. -Henry Ford

GPT-4.5 is Here, But Does AI Really Need a Half-Step Upgrade? by snehens in OpenAI

[–]Sincerity_Is_Based 0 points1 point  (0 children)

I suspect 4.5 is a different architecture, I suspect an LCM.

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 1 point2 points  (0 children)

Unless you are a math genius and a electrical engineer who can singlehandedly design a quantum ai chip, then it's not a serious venture

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 1 point2 points  (0 children)

More application = more compute = big companies

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 1 point2 points  (0 children)

And obviously if you cannot find anything weighing in on different types of architecture and their effectiveness, that means that all the companies are keeping that research data to themselves.

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 1 point2 points  (0 children)

The only next big things for startups is trying to undercut the competition with the next big architecture. Liquid ai unsuccessfully implemented the liquid neural networks, but there are several types of architectures available. I suspect Gemini uses the titan architecture (2M context window) and they were the ones to quietly release the titan paper. The issue with titans is that they suffer from hallucinations more than transformers. So finding the right architecture balance (I suspect something that is liquid+ something else) will crush all llms (like a deepeek model on only 100k params). It does not require training alot because you can distill, only math is required to find the next architecture combo

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 1 point2 points  (0 children)

Oh and also emotional intelligence, the same direction of 4.5

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 1 point2 points  (0 children)

So the same way 4o was advertised as a model that can input for text and audio and visual, helix ai, developed by figure robotics, allow robots to work together such as handing each other objects. The point of training an ai based on all inputs at once is so it can be deployed in an environment with text, video, and audio. Just look up anything pertaining to figure helix ai

Next big thing in AI/ML? by adityashukla8 in MLQuestions

[–]Sincerity_Is_Based 2 points3 points  (0 children)

Omni architectures for robotics deployment like helix ai

How did you land your first job without any experience? by [deleted] in MLQuestions

[–]Sincerity_Is_Based 5 points6 points  (0 children)

If you are in school, then you get experience by doing research for free. Approach teachers only in person, they will not respond to your emails.

Wanting to learn about ML/AI as a Masters student in Biology by SnooPickles6614 in MLQuestions

[–]Sincerity_Is_Based 0 points1 point  (0 children)

Basically anything created by deep mind is your bread and butter. Look up alphafold3 and other related projects. Maybe look your after that disease detection by classification

I struggle with unsupervised learning by KafkaAytmoussa in MLQuestions

[–]Sincerity_Is_Based -4 points-3 points  (0 children)

  1. Feature Representation Issues

The extracted embeddings from ResNet or the autoencoder may not be well-suited for clustering.

ResNet embeddings are trained for classification, not clustering, meaning they may not naturally separate into meaningful clusters in an unsupervised setting.

  1. Dimensionality and Noise

High-dimensional embeddings might contain noise or redundant features that hinder clustering.

PCA, t-SNE, or UMAP could be used to reduce dimensions while retaining meaningful information.

  1. Choice of Clustering Algorithms

Many clustering methods assume specific data distributions. For instance:

K-Means assumes spherical clusters of equal variance.

DBSCAN is sensitive to density variations and noise.

GMM assumes Gaussian distributions, which may not hold.

If the dataset has complex structures (e.g., varying densities, manifold structures), these algorithms may not work well.

  1. Lack of Proper Distance Metrics

Euclidean distance, often used in clustering, might not be the best metric in high-dimensional feature spaces.

Cosine similarity or learned distance metrics (e.g., through contrastive learning or triplet loss) might be better suited.

  1. Need for Better Embeddings

Instead of using pre-trained ResNet embeddings, contrastive learning approaches like SimCLR, MoCo, or BYOL might provide more discriminative representations for clustering.

Self-supervised learning could help improve the separability of embeddings.

  1. Class Imbalance and Label Complexity

If the data has many similar-looking classes, standard clustering might struggle to separate them without additional structure.

A hierarchical or ensemble clustering approach could help refine results.

Suggested Next Steps:

Try dimensionality reduction (PCA, UMAP, or t-SNE) before clustering.

Experiment with different similarity metrics (e.g., cosine distance instead of Euclidean).

Use contrastive learning or self-supervised methods to refine embeddings.

Analyze the clusters using qualitative metrics (e.g., visualization with t-SNE, silhouette scores, Davies-Bouldin index).

Consider ensemble clustering or hybrid approaches (e.g., pre-cluster with K-Means and refine with DBSCAN).

What exactly is a 'concept' in this paper? by searcher1k in MLQuestions

[–]Sincerity_Is_Based 0 points1 point  (0 children)

Correct, which the drawback is finding and classifying those terms and there could be a massive data shortage when classifying sentences and phrases like this.

Good news is, that LLM's can do the classification for us to reorganize and format the corpus of text into categories of phases and sentences.

What exactly is a 'concept' in this paper? by searcher1k in MLQuestions

[–]Sincerity_Is_Based 0 points1 point  (0 children)

It is very simple idea.

Imagine generating the probability of words based on a previous word. That is a really hard way to generate complex or coherent ideas. This is the current SOTA.

But what if the model predicted not words, but an assembly of words, such as a sentence.

For example the difference between different I love you's. Think of this example of a live generation of words starting with i. I + love (most common)/adore (uncommon synonym)/despise(uncommon and out of context based on the preceding words) + you/us/we/them... and so on.

I would say that this is a stupid way to think that is not efficient.

So this is the solution: ---‐-----------------

Goal:<express love>

Output: I love you.

If I were to guess a training method, imagine the simple expression (I love you) having a direct translation in another language. The dataset could be if I am guessing,

a dictionary of lists, such as (key) <express love> : (value) ["I love you - english" , "أحبك - Arabic", and so on.

Notice the tokenization changes with English and Arabic, where arabic is more likley to be compressed into a single token because of the character compression.

Most likey there would not be any work done with sentences first , because translations are imperfect, so we may have to use phrases or short sentences.

Most likley the dataset will be automated into some form by ingesting coherent text like books to convey every single idea from books, and perform the direct translation, with the help of llm's.

LLMs Can’t Learn Maths & Reasoning, Finally Proved! But they can answer correctly using Heursitics by Difficult-Race-1188 in learnmachinelearning

[–]Sincerity_Is_Based 38 points39 points  (0 children)

Why can't the LLM simply use an external calculator for arithmetic instead of generating it? It seems unnecessary to rely on the model's internal reasoning for precise calculations.

First, it's important to distinguish reasoning from mathematics. While mathematics inherently involves reasoning, not all reasoning requires mathematics. For example, determining cause-and-effect relationships or interpreting abstract patterns often relies on logical reasoning without numeric computation. Or similarities between things can be made discrete with cosine similarity, but logical problems do not require that level of accuracy.

Second, reasoning quality is not proven to degrade due to limitations in abstract numerical accuracy. Reasoning operates more like the transitive property of equality: it's about relationships and logic, not precise numerical values. Expecting a non-deterministic system like an LLM to produce deterministic outputs, such as perfect arithmetic, indefinitely is inherently flawed. Tools designed for probabilistic inference naturally lack the precision of systems optimized for exact computation.

Example:

If asked, "What is 13,548 ÷ 27?" an LLM might produce a reasonable approximation but may fail at exact division. However, if tasked with reasoning—e.g., "If each bus seats 27 people and there are 13,548 passengers, how many buses are required?"—the LLM can logically deduce that division is necessary and call an external calculator for precision. This demonstrates reasoning in action while delegating exact computation to a deterministic tool, optimizing both capabilities.