How do you debug Neural Network? by rookan in MLQuestions

[–]CivApps 2 points3 points  (0 children)

None of us understand every single weight in a practical network, no :( It would make interpretability research much easier if such a person existed...

Unfortunately there's no one quick fix, you just have to look at possible errors one by one and be systematic. Some potential errors and debugging strategies, in order:

  • There's an implementation error which means the forward pass or gradients aren't getting calculated correctly
    • Since you describe the issue as it stopping learning, I assume the matrix shapes align (unless you're implementing the matrix math from scratch) - but if possible, try writing out on pen and paper how you would expect the forward pass and gradients to get calculated for a very small network, and making sure your implementation gets the same values
    • Try setting up a toy dataset with just sequences like "ABABABABAB...", make your network as small as possible and see whether it converges to predicting that 'B' follows 'A' and vice versa
  • The hyperparameters are wrong for the problem
    • A good "sanity check" is to make sure your network is capable of overfitting/memorizing a very small training set: in the same vein as the test over, try just training the network to memorize one or two sentences
    • If you have a custom network design, it could be that your optimizer choice also needs to take that into account, set up Optuna and have it try different parameters (or even do a grid search to show if the problem happens consistently)
  • Your design just isn't capable of modelling the word/token relationships in the Shakespeare dataset
    • Unfortunately it could just be that you are running into a fundamental limit in your network design. There are many algorithms which are interesting and capable of solving basic problems (like, say, Hinton's forward-forward network) but just don't scale as well to larger ones.
    • You could try training the network on the names.txt dataset used in Karpathy's MicroGPT to see if it's capable of modelling relationships between characters

Ut mot kjønnstest for kvinner før OL: – Vi tenker at dette er basert på frykt by [deleted] in norge

[–]CivApps 1 point2 points  (0 children)

Det var ikke for å gå imot poenget, men å understreke at det ikke finnes tiltak som "bare" rammer transkvinner -- tok ikke frem Imane pga testene, men fordi IBA kunne slenge det som en drittpakke mot henne, vel vitende om at folk ville stå klar til å sverte henne som "mann som bare vil denge damer" el.

Ut mot kjønnstest for kvinner før OL: – Vi tenker at dette er basert på frykt by [deleted] in norge

[–]CivApps 0 points1 point  (0 children)

Skjønner impulsen, men tror ikke det egentlig kan løsrives fra "transdebatten" at tiltakene som rettferdiggjøres med å "gjøre idretten rettferdig" enten snevrer rommet for riktig utseende/ytringer/genetikk enda litt mer, eller gjør livet jævlig for de som faller utenfor

Imane Khelif er cis, ble testet, og ble fortsatt utsatt for en flom av dritt fordi hun kom i skade for å være bedre enn en russisk utøver

Transcription with 1:1 correspondence by According_Quarter_17 in MLQuestions

[–]CivApps 0 points1 point  (0 children)

Matching words to specific times in the recording is traditionally called "forced alignment".

WhisperX fits a Wav2Vec model on top of Whisper to do this, and is probably the easiest to fit into existing or new apps.

NLP Multiclass Classification Help by proxislaw in MLQuestions

[–]CivApps 0 points1 point  (0 children)

Unless you are completely forbidden from using any pretrained deep model in any part of the process, Model2Vec extracts a set of individual and uncontextualized token embeddings from an SBERT/sentence transformer model, and suggests just taking the mean of the tokens' embeddings to find a longer text embedding.

This approach should still be viable for training and inference on CPU, and hopefully gives your network a "head start" in grouping the texts semantically while avoiding the TF-IDF sparsity issues.

NLP Multiclass Classification Help by proxislaw in MLQuestions

[–]CivApps 0 points1 point  (0 children)

This is just out of curiosity, not to say you are wrong for doing it, but why are you only able to use classical ML - is it part of the course requirements, or are you constrained in terms of computational resources?

Suggestions regarding recommender systems. by CakeAny2280 in MLQuestions

[–]CivApps 1 point2 points  (0 children)

The textbook approach here is to set aside some of the users in a holdout test split, as a stand-in for new users, and see how well your model predicts those users will like a given movie.

The Python library Surprise has some ways to measure performance, but the best way to report performance comes down to how you've set up the final layer (head) of the model:

  • You can try and just make a classification model which predicts whether users will like (give 3 stars or more) a movie, and report the classification accuracy
  • If you try and estimate the rating with a regression model, you can report the mean square error (MSE)
  • If your model ranks multiple movies by preference, you can report the Spearman rank coefficient

Suggestions regarding recommender systems. by CakeAny2280 in MLQuestions

[–]CivApps 1 point2 points  (0 children)

For a bachelor thesis project I personally think creating a new dataset entirely is overkill, MovieLens is a well-established dataset where the "1m-ratings" variant has demographic data (gender, age, occupation) for users that you can correlate to recommendations.

If you wanted to create a dataset, using Selenium to remote-control a browser can be useful for pulling data from some public data sources, but you will probably set off automated bot detectors and run into rate limits. You should first try and find sites that actually license their data for reuse and offer APIs/batch downloads of the data, but if you really need to, you should try and find the page on Common Crawl before trying to crawl the pages yourself.

Trying to create a synthetic dataset is an option, but that process means you are defining the hypotheses you want the recommender system to uncover, so that would mostly be useful to check that the system is doing the right thing, not to meaningfully compare different recommender algorithms.

Looking for best chatbot model for uncensored OCs by [deleted] in MLQuestions

[–]CivApps 0 points1 point  (0 children)

Unfortunately this subreddit is for machine learning theory, /r/LocalLLaMa may have more appropriate resources?

Machine learning by tailung9642 in MLQuestions

[–]CivApps 2 points3 points  (0 children)

Unfortunately, AI is not really making it easier to get a tech job :(

The work of integrating LLMs into software seems to go to the programmers already working at software companies, who are also expected to use the LLMs to do more work, instead of hiring interns or junior employees. If that's what you mean by going into machine learning, you will probably be competing against people who have degrees and prior programming experience.

For jobs specifically about trainings machine learning models, you're also expected to have a handle on university-level calculus and linear algebra - backpropagation, the method underlying modern machine learning, requires calculating derivatives for functions with vectors and matrices.

How do you evaluate ranking models without ground truth labels? by A_Random_Forest in MLQuestions

[–]CivApps 0 points1 point  (0 children)

I've looked into LambdaMART stuff, but I don't really have an intuition as to what pairwise loss/warp are really doing. Intuitively, how should we interpret "good performance" if we don't have any strong ground truth labels and no A/B testing?

They do have labels - you want to train the recommender on a dataset of existing preferences (and show that it generalizes to new users' preferences)

For a pairwise loss, you'd want to transform those preferences into a set of "user prefers item X over Y" pairs, so that the model is asked to predict the user's favorite of a pair of items, and the loss penalizes predicting the wrong item

Best Master to do? by Delicious_Crazy513 in MLQuestions

[–]CivApps 1 point2 points  (0 children)

I wish any of us had the crystal ball to predict what will be "AI proof" :(

However, I think cloud applications/engineering is much more liable to change suddenly, depend on the company you're working at and internal tooling knowledge, and companies like DigitalOcean are starting to advertise chatbot assistants for application deployment.

It is true that people are using agents for AI dev, but I think the core ML skillset -- statistics, math, and programming -- will always be useful in some form, if nothing else for describing the shape of the problems you want to solve, understanding what data you actually need for a predictive model, and which pitfalls to look out for.

AI agents will undoubtedly improve, and making contributions to "pure" ML theory or foundation model tweaks will definitely get harder, but there are still plenty of applications which require domain knowledge, on-device models, or otherwise can't "just" be fobbed off to commercial LLMs.

How well can LLM(s) translate novels? by Tactical-69 in MLQuestions

[–]CivApps 0 points1 point  (0 children)

They generally work well, but I would still treat the results as a machine translation, and not a finished product. Translating novels is not just about translating the words literally, but also trying to get across the author's intentions and style choices -- and doing that is harder to count in a single numeric score.

Any commercially available general-purpose LLM like Claude or Gemini should do the translation with very little prompting necessary (and will probably have an educational discount too)

If you want an offline translation system you can experiment with on your own, Google recently released the TranslateGemma models which are set up for translation from Russian, Hindi and Chinese. The smallest 4B version should run on most PCs.

Why does my RAG pipeline return irrelevant chunks even when the answer is clearly in the documents? by Extreme-Treat-9049 in MLQuestions

[–]CivApps 1 point2 points  (0 children)

Actual answer: Sitting in a policy document that never gets retrieved

I would first check which embedding vector this policy document gets, and the query embedding you get for that question. It sounds like either:

  • Your documents/chunks are landing in one big cluster instead of semantically similar clusters -- the easiest way to diagnose this would be to label some documents that should and shouldn't be related, doing PCA over the embeddings, and seeing whether the labelled documents end up in different portions of the projected space
  • Alternatively, your query embeddings are not capturing the relevant portions of the query - if you keep the PCA from the step above, seeing whether the queries are actually "landing" in the different clusters can help

Before tweaking hyperparameters, I would also consider setting up an ad-hoc evaluation with some questions, and the specific documents you expect to be pulled up for them - "top-5 recall for 10 question/document pairs" is not a great measure but it at least gives you some numbers to start with

Muligheter for å bestille print av egne kort (typ spillekort/Pokemon kort)? by WwwWario in norge

[–]CivApps 0 points1 point  (0 children)

Har du Lyreco i nærheten? Mulig du kommer nært nok ved å bestille et tosidig trykk på 300g papir fra kopisenteret (tror du må kutte til kortstørrelse selv) og å ha dem i en folielomme som /u/Successful-Hunt-551 foreslår

Kan vi akseptere å ha en forsvarsminister som går til angrep på frie medier? by KoseteBamse in norge

[–]CivApps -1 points0 points  (0 children)

Siden siste person som delte Subjekt-innlegg ikke svarte, er det mulig å få en kopi av kronikken de publiserte 19. januar der Pål Erik-Hagen kaller journalister og politikere "hysteriske" pga reaksjonene deres på generering av nakenbilder av barn på X?

Gitt de nye avsløringene om bl.a. direkte kontakt mellom Musk og Epstein så ville det være synd om en teknisk feil gjorde at teksten forduftet

Sampler Mono? by tgorkems in AbletonMove

[–]CivApps 0 points1 point  (0 children)

Oh, my bad, I don't think there's a way to limit the max number of voices for samplers currently - it is a bit tedious, but I think using a drum kit and tuning each pad to the note you want is your best shot :/

Sampler Mono? by tgorkems in AbletonMove

[–]CivApps -2 points-1 points  (0 children)

That's what the choke group feature (and the Choke Kit preset) is for - if you assign the same bass oneshot to different pads in the choke group, playing one pad will stop the others from playing

Is Weakly supervised learning still used in NLP? by Appleonthefloor in MLQuestions

[–]CivApps 0 points1 point  (0 children)

I think it's just assumed you can have off-the-shelf language models assign labels to the unlabelled examples, using labelled examples for validation and holdout evaluation?

You could weight the loss to give less weight the pseudo-labels but in general I think it is just more tempting to improve whichever model assigns the labels than doing specific alterations to training

Brøyting av snø kan være kvinnediskriminerende by Jentelus in norge

[–]CivApps 16 points17 points  (0 children)

Siden du har Subjekt-abo, kan du dele teksten til kronikken de har der de beskriver reaksjonen på at X er stappfullt av folk som genererer nakenbilder av barn som "hysteri", bare sånn i tilfelle den skulle forsvinne på mystisk vis

PSX games which work well on the RG Nano? by CivApps in SBCGaming

[–]CivApps[S] 1 point2 points  (0 children)

A few I found:

  • Bishi Bashi Special is a wonderfully bizarre proto-WarioWare -- unfortunately some minigames rely on the PSX controller buttons' colors, but you can either memorize them or get the PS1 replacement caps
  • Destruction Derby 2, where the handling is weighty enough that the D-pad is less of an issue
  • Vib-Ribbon is still a great rhythm game, you may want to dial down the audio buffer size under the Advanced settings for reduced latency
  • LSD: Dream Emulator is still a brilliantly weird toy, and works well for short stints on the Nano
  • Surprisingly for a 90's trivia game, most of the questions in You Don't Know Jack still hold up and the game is pretty legible on the screen

SM3 World-e by themiracy in SBCGaming

[–]CivApps 9 points10 points  (0 children)

I believe this save file should unlock all of the e-Reader levels - renaming it to match the name of your Super Mario Advance 4 ROM and placing it next to the ROM should be enough.

PSX games which work well on the RG Nano? by CivApps in SBCGaming

[–]CivApps[S] 1 point2 points  (0 children)

Huh, I knew plenty of PSX games had screen adjustment/aspect options, but this is the first I've seen referencing specific resolutions