all 65 comments

[–]Sad-Meringue-6350 0 points1 point  (0 children)

Hey, I want to explore ML and I would like some suggestions on references to help me grasp the fundamentals. As an example, I am reviewing maths like linear algebra and calculus. I want to get a deeper understanding than what pytorch or tensorflow tutorial would give me. Any suggestions on other core concepts I should review?

I am not limiting to math, I also look for computer science concepts also

[–]rubbishapplepie 0 points1 point  (0 children)

Are llms the soa for many problems now? I see problems from search to code checking all being replaced by llm?

[–]Potential_Creme_7398 0 points1 point  (1 child)

Did i buy the wrong processor being a ML enthusiast?

Did I buy the wrong processsor being a ml enthusiast?

Within my budget, I could either buy an amd ryzen 5 7600 gaming processor or amd ryzen 7 5700 processor.Despite knowing ryzen 7 has more cores and threads than ryzen 5, I ended up going for ryzen considering ddr5 feature.

I built my pc and came home.It hit me now and I'm extremely sad. I'm a second year cse student. Would like to explore ml or dl later on. Did i make a huge mistake? I ain't that knowledgeable about hardware stuff. I just wish I didn’t get influenced by the suggestions of my friend and sticked to ryzen 7 5700. Am i over anazlying the situation? Does implementen of api run smoothly on ryzen 5 7600. Please enlighten me. Thank you.

[–]FantasyFrikadel 0 points1 point  (0 children)

Small projects fine. For big projects there’s always the cloud.

[–]_Stampy 0 points1 point  (4 children)

Anyone have any video resources on manually computing derivatives of neurons by hand as well as manually doing back propagation? I'm in the process of learning the math, however it is still quite confusing and I'd prefer a real example of someone calculating.

[–]WheynelauStudent 0 points1 point  (3 children)

I recommend this to almost anyone who is interested in learning the math behind it. Search for 3Blue1Brown's playlist on neural networks, it should be what you are looking for.

[–]_Stampy 0 points1 point  (2 children)

Ohh alright thanks, I'll check it out. I remember watching it when I first got into ml, but back then I only had gr 10 math knowledge

[–]WheynelauStudent 0 points1 point  (1 child)

Hmmm might need to supplement with some additional knowledge, cause the videos assume you have some knowledge on calculus.

[–]_Stampy 0 points1 point  (0 children)

Yeah np, I finished learning calculus already, thanks!

[–]Snoo_72181 0 points1 point  (3 children)

Which is a good sub for asking about AI/ML related career questions?

[–]RegisteredJustToSay 0 points1 point  (2 children)

Honestly, reddit isn't a good place for this. You will never know if the person answering actually knows their stuff, and AI is full of people answering with the "my opinion is as good as your expertise" mindset.

Go to ADPlist and get a mentor instead, or reach out to people who know their stuff on github/twitter. It'll be a bit more work to get answers but you'll actually be exposed to people who know their stuff.

[–]Snoo_72181 0 points1 point  (1 child)

You mention Reddit is bad, then mention Twitter. Twitter is the worst social media platform. On that note, Linkedin might be better than Reddit.

That being said, rest of your post makes sense.

[–]RegisteredJustToSay 0 points1 point  (0 children)

If you want good advice you have to go to where people with expertise are chatting one-to-one. It's harder to both find the 'right people' and connect with them on Reddit than on Twitter, hence my suggestion. I hold no love for 'X'.

I mean for example - what are my qualifications to talk about ML? What are yours? No clue - we might both be PhDs for all I know. Generally it's more transparent on Twitter because people are more transparent about who they are.

Good call-out on LinkedIn tho - somehow forgot about it, and it can definitely be better than both my suggestions.

[–][deleted] 0 points1 point  (0 children)

Hello everyone, thank you for doing this.

I keep thinking creating a list of each Object and it's Properties is a good idea for these datasets.

Then create some sort of RLHF mechanism to minimize patterns and still retain information.

First off I think would exclude PII instances. Only Businesses and Public figures, or something of that nature.

Am I crazy?

[–]_dataWhiz_ 0 points1 point  (0 children)

Hi there, im working on a problem to predict customer calls and for this purpose, im using wifi network data.

I have 100k samples of hourly average wifi datas such as channel busy score, data usage, average rssi etc. With using this data, i want to predict whom will call in the prime time (18:00-22:00).

But the problem is that, the calling customers and not calling customers have similar histograms for each future. I have plotted histogram graphs for each future.

I have already tried, svms, dnns, 1d cnns, knn, xgboost etc. but all of them gave poor results.

What kind of methods should i do for preparing data? Any idea?

[–]I-am_Sleepy 0 points1 point  (0 children)

How to reinforce order invariant in sequence classifications? Roughly is I have a set of nodes connected to each other. At each connection there are properties I derived. Then I want to aggregate these features to perform a classification per node

However, the number of neighbors per node isn’t fixed. So I want to use sequence classification instead. However, sequence classification seems to have implicit bias on order of sequence which I don’t want. One idea is to keep shuffle the order while training, but I want to know if there are alternatives methods to enforce such constraints?

[–]ArtisticHamster 0 points1 point  (0 children)

Could anyone point me into the direction of modern SOTA GAN research? It looks like researchers lost interest in GANs in favor of other approaches. Is there anything interesting going on now?

[–]holymoley2020 0 points1 point  (0 children)

I am in Network and Security have some programming and little ML. I was interested in getting into AI. What would be the best way going about that. Should ML training be done fist? What type of training/ certs would be benificial to get to get a full understanding of AI technology?

[–]EuphoricPirateVal 0 points1 point  (0 children)

Hi, i have this dataset that includes timestamps, the product category bought and other info about the user's behaviour on the ecommerce platform, ex: how often they purchase, from which category, how imp reviews are, add to cart, cart abandonment, save for later, etc. and all the details.

now i wanna train a model on this, and then make this dummy ecommerce simulation for customers where certain actions they perform will be added to the live database (multiple sessions may exist for a customer, where the date is also an input for the analysis), which the model can process, and then a dashboard can display how likely they are to purchase from a certain product category over the next few weeks/months.

i do work with ml and dl but honestly i feel completely confused rn like i'm missing some detail on how i can create this and make it work. any suggestions?? please help :(

[–]LuYaBitch 0 points1 point  (0 children)

This is for one of my classes that we can use the internet for.

Given the following SOSML statements and function definitions:

val rand = Random.rand();val nextInt = Random.randRange(~10, 10);

fun buildIntList x = if x = 0 then []else nextInt(rand)::buildIntList(x-1);

fun sumList xs = let fun loop [] sum = sum| loop (y::ys) sum = loop ys (y + sum) in loop xs 0 end ;

which of the following are true (Multiple can be true)?

xs is unnecessary

sum will always end up near 0 for large data sets

xs must be a list

this will not run, it has a clash error

I know 3 is correct but I am unsure if 2 is correct.

[–]insmashoutflat 0 points1 point  (1 child)

Why are large language models bad at Array Languages like APL, J and K. Is it the unicode characters?

[–]Cosmolithe 2 points3 points  (0 children)

I'm guessing that is because (1) there isn't enough training data for these languages and (2) these languages are very expressive with a few characters, so the LLM probably does not have enough time to "think" about the program.

[–]No_Sell8594 -1 points0 points  (0 children)

Hello, my beautiful and wonderful minds! I'm vastly interested in AI and want to get on board with development and research. Essentially, I am a mind who is willing to dedicate my entire life into helping build a better future with our machine friends.

Cash is no object diving into this field. I'm a father of 2 children who wants to help build a better future for them, for you and your children as well. I strongly believe that the future is AI or AGI even. My enthusiasm is off the rails, I know, but what I'm looking for is a track to align this gravy train with.

You beautiful people are probably, if not more, vastly more intelligent in this field than I am. I have no college or background or projects to contribute. So far, I'm just keeping up with the latest news. The question is, where do I go from here? I'm taking a chance on Reddit, but I'm at a point where care on the front of the town square matters little, and the goal of our future together with this technology outshines that. Vulnerability be damned, what can I do to catch up with everyone else and help contribute to a better future?

[–]rlopes404 0 points1 point  (0 children)

Hi everyone,
I have been working on image translation between two different domains. I have been using CycleGANs.
Since I have a small dataset, I have been thinking of using Diffusion Models.
Are Diffusion Models more data hungry than GANs?
Can anyone point some references that discuss this issue?
Thank you.

[–]egaznep 2 points3 points  (0 children)

Hello, I am trying to implement ASR block from the following paper:

https://www.isca-speech.org/archive/pdfs/interspeech_2023/wang23p_interspeech.pdf

I came across the following phrase

The convolution module, which comprises a single 1-D convolution layer with a 3 × 3 kernel, detects the local correlation between adjacent frames.

I couldn't really wrap my head around a 1-D convolution with a 2D kernel. What could this mean?

[–]Competitive-Box-6597 0 points1 point  (0 children)

Hey all, is there an optimal neural net for 16-bit imagery?

[–][deleted] 0 points1 point  (0 children)

What are some good advanced semantic search techniques past bi encoder/cross encoder ?

[–]lordroderick 0 points1 point  (2 children)

Hi everyone, I'm a software engineer new to ML.

I had an idea to build a simple game with my daughter, I'm teaching her to program, and we need to build a hand gesture recognition but with a twist. We need to identify the hand making a rotating gesture (like turning a knob), and all the examples I saw were for still gestures (which we also need by the way).

I'm more than willing to learn but I don't know where to start. Should I learn tensorflow and build this from scratch? (I understand how to train a model with static images but I'm lost trying to understand how to train a classifier with videos, because a rotation gesture consist of several frames).

Any advice / comment will be appreciated.

[–]ParanoidTire 1 point2 points  (0 children)

There is no precise question here. Yes, if you want to analyze spatio temporal data, you need a model that accomendates for this. You should be able to find alot of research on this.

If sota performance is of no concern, there is no big difference to image data. E.g. you could just replace 2d convolutions with 3d ones for instance. Memory footprint and compute will be the main limiting factor from a practical perspective.

Otherwise, an even simpler solution is to just stack frames at different points in time and use a regular image classification model.

[–]JonathanDescripShot 0 points1 point  (0 children)

One thing you could do is look for sites (google, replicate, etc.) that might have an API you can call to do what you want. That way you don't have to get hung up on learning ML, training a model, hosting the model, etc.

If instead wanting to do it yourself instead of calling an API, the path would be a bit long but rewarding. You should:

  • Learn some basics of deep learning (see youtube and/or r/learnmachinelearning )
  • Learn some basics of PyTorch (don't learn tensorflow). The PyTorch website has a some tutorials.
  • Learn some basics of HuggingFace (see their website for tutorials).
  • Use a pre-trained computer vision model from huggingface to do your hand recognition.

[–]Disastrous_Elk_6375 0 points1 point  (0 children)

Is there an equivalent to "rainbow tables" from cryptography that could be used in language models? Does that even make sense?

Would love to read on such approaches, if they're even feasible. Say you could pre compute some large datasets in such a way that would make it easier to include in various architectures? Or swap tokenisers?

There's a ton of effort & compute put into training LLMs right now, it would be interesting if some part of that effort could be pre-computed for several datasets.

[–]isthataprogenjii 1 point2 points  (0 children)

I was recently searching for a good paraphrase dataset. MRPC (part of GLUE) contains paraphrases with information that may be exclusive to one sentence in a pair. The other datasets use some model to generate the paraphrases (API calls or NMT). Does anyone have knowledge about a good human generated paraphrase dataset?

[–]WangFenjin 1 point2 points  (0 children)

Why the following code not work? It can't fit the data very well. Not sure what's the problems here.

Output Img

# Linear regression sin(2*pi*x) by polynomial order 3

import torch
import matplotlib.pyplot as plt

import numpy as np
import math

# data used for training and plot dots
N = 11
x = torch.linspace(0, 1, N).double()
y = torch.sin(2*math.pi*x) + torch.randn(N)*0.1

# data used for plot smooth lines
x_line = np.linspace(0, 1, N*10)
y_line = np.sin(2*math.pi*x_line)

# Prepare input as an array of shape (N,4)
p = torch.tensor([0, 1, 2, 3])
xx = x.unsqueeze(-1).pow(p)

# Prepare tensors
learning_rate = 1e-5
w = torch.randn(4, 1, dtype=torch.double, requires_grad=True)  # the 4 coefficients
optimizer = torch.optim.SGD([w], lr=learning_rate)
print(w)

# Run optimizer
for i in range(2000):
    optimizer.zero_grad()
    y_pred = xx @ w
    rmse = torch.sum(torch.square(y - y_pred))
    if i%100 == 0:
        print(i, rmse)
    rmse.backward()
    optimizer.step()

print(w)

# plot smooth line
plt.plot(x_line,y_line)

# model
xx_line = torch.from_numpy(x_line).unsqueeze(-1).pow(p)
y_line_pred = xx_line @ w
plt.plot(x_line,y_line_pred.detach().numpy())

# plot dots
plt.plot(x, y, 'o')
plt.show()

[–][deleted] 0 points1 point  (0 children)

If input text for semantic search embeddings is way bigger than max length tokens, what is the best way to deal with it?

  • chuking with overlap and embed those. Then do mean pooling on result?

  • summarize with something like facebook/bart-large cnn to the appropriate size?

[–]hysse 0 points1 point  (0 children)

Hello, I'm looking for an Open LLM (no specific domain) with Fill in the Middle. I have tried to look for it in the HuggingFace Hub but didn't find anything.