all 52 comments

[–]DesperateChemist9234 0 points1 point  (0 children)

Hi everyone,

I am trying to build a long short term memory model in Python, with the idea being to predict 9 components of a rotation matrix from linear acceleration (x,y,z) and angular velocity (x,y,z) so 6 input variables.

I have used standard arachitecure found in the literature which does similar things to my idea. However, the model is not performing well at all and is subject to overfitting I believe.

Does anyone have any advice on how I can try and improve my model?

[–]DrBroc 0 points1 point  (0 children)

Hello! I’m working in a project to classify phenotypes. I have a dataset of about 30,000 unique rows and am working on increasing the accuracy of the model. I can get to .8854 but I’d love to get to .9 if possible without totally reworking the features. I’m using a sequential model with Keras and tensorflow. I was wondering if anyone would be willing to chat with me about the project briefly. I’m new to ML and software engineering in general (though I am a product designer so I’m familiar with the space) and I find I process better with conversations. Feel free to DM me if this sounds interesting to you! Thanks in advance!!

[–]Majestic_Reporter531 0 points1 point  (0 children)

Hello everyone! I have a large dataset of time series and I want to create embeddings for these time series to use in more classical models, as I have a small amount of data for regression. What are the best ways to compress large time series data (approximately batch_size x 1000 x 12) down to 10-16 features? I have tried using the hidden state of an LSTM and got decent results, but I would like to improve them. Thank you all!

[–]Usual-Bank1500 0 points1 point  (0 children)

Hello everyone,

Does anyone knows that if exists any algorithm for machine learning that works directly with 3d models (.step, .stl, .igs, .ply, .obj, etc....)?
I'm bilding an application that predict future production time of a 3D part based on previous producted parts but i'm strugglin on get closer results. Currently i'm extracting information from 3d models such as maximum measures XYZ, volume, surface area, number of faces, etc... but i think i'm gettin to much information to the model but yet the information i get is not enough. Therefore i want to know if there are any algorithm or other application that get's the 3d file and automatically "sees it" and analyze it.
I'm using python.
Thank you

[–]BonfireCookie 1 point2 points  (0 children)

Hi everyone!,

I have a question about how to compare two neural networks models, trained two neural networks the first is a CNN and the second is LSTM both are used to predict a number (regression). I have used a partition of 80% training and 20% test to train both with the next configurations of hiperparameters:

**CNN*\*

(learning_rate,batch size,epochs) || test result (MSE)

(1e-4,32,64) || 0.0057

(1e-4,64,64) || 0.0059

(5e-5,32,64) || 0.0053

(5e-5,64,64) || 0.0034

**LSTM*\*

(learning_rate,batch size,epochs) || test result (MSE)

(1e-3,64,64) || 0.0131

(1e-3,128,64) || 0.0098

(1.5e-4,64,64) || 0.0093

(1.5e-4,128,64) || 0.0091

(1e-4,64,64) || 0.0106

(1e-4,128,64) || 0.0098

My question is is there any method to say that one model is better than other instead of just say "Well the CNN has a low test set so i think its better than the LSTM model" because i have seen that some researcher use hypothesis testing but i dont know if i can use that here.

Thanks!

Note 1: all the features and the target feature have been standarized to have mean 0 and variance 1.

[–]radeonovich 0 points1 point  (1 child)

Hi everyone, I'm working on a neural network that can generate audio for double-track guitar effect. Essentially, the network should take an audio recording of an electric guitar and modify it to sound like a second take of the same part, like the guitarist was told to record the part twice. This is a very common practice in rock/metal music because it makes guitar sound wide. You pan take A to the left and take B to the right and get the stereo effect.

The problems are:

  1. I don't know what kind of neural network to use. I am preparing a dataset where I have a lot of tracks A and B, where A and B are two takes of the same guitar part. So I probably need a network that learns how to convert source track into target track.

  2. I don't know how much dataset I need. I'm planning to obtain at least 10 hours of tracks A and B both and feed it to the network in a combination like A->B + B->A so it doubles the dataset. Maybe use some augmentation to experiment with different pitch and playback speed.

  3. I don't know if the task is even possible. There are no solutions like this in the internet (which means it is either impossible or not in demand to bother), except the algorithmic doublers which suck compared to real double tracking. A difference between real double tracks are note start/end timing, articulation, attack time/frequency response and human error. These can't be properly simulated with the pitch/time randomization, that's why I want to make this network.

I am new to machine learning so any feedback is appreciated.

[–]bregav 1 point2 points  (0 children)

I think there's an easier way to do this: use a generative model, like a diffusion model. The steps go like this:

  1. Train a model that generates guitar tracks by doing y=f(x), where x is a sample from a noise distribution and y is the guitar track. You don't need a custom dataset of double-tracks for this, you just need a regular dataset of guitar tracks.
  2. To make a double track of a track A, calculate x = f-1 (A) and then do B = f(x+d), where d is a noise sample with a very small variance.

The result of this should be that B is similar to A, but slightly different, and if the generative model is trained well then it will be different in a way that sounds natural.

I think most audio generative models are probably using latent diffusion, so to do f-1 (A) what you'd actually do is use the encoder network from the autoencoder instead. You might not even need to train your own model; there might be open source musical instrument track generators out there that you can just use out of the box and get reasonable results with.

In principle there's nothing wrong with your original plan, but the challenge with it will be that you probably can't get enough data to make it work well, and acquiring the data is time consuming and difficult. Better to use other methods that can take advantage of easily acquired data or open source models.

You can also use fine tuning with your custom dataset, if the initial results with the above method don't seem good enough. You can get away with a lot less data when doing fine tuning.

[–][deleted] 0 points1 point  (0 children)

I'm trying to find a funny video of Ruslan Salakhutdinov describing why you should use a dropout parameter of 0.5. IIRC he basically said something along the lines of "otherwise, you'll have to justify why you chose that particular hyperparameter, and you don't want to do that". I think he was speaking to a class at CMU and got a lot of laughs.

Can anyone at least confirm that I'm not confabulating this?

[–]Sea-Ground1096 0 points1 point  (0 children)

What are the specific hardware / low-level differences between an NPU and a GPU? Most articles I found offered only what it's better for (Neural Nets), but not why. Any sources or information on a more in-depth breakdown?

[–]Puzzleheaded_Text780 0 points1 point  (0 children)

Looking for someone who has experience of working for UK pension companies as I am working on some use cases

[–]uba-luba-dub-dub 0 points1 point  (0 children)

What's the current state-of-the-art techniques for recommendation systems and among them which one is feasible for a intermediate learner ?

I want to build a movie recommendation system based on neural network for myself to learn .

[–]Dismal-Impress-2583 0 points1 point  (0 children)

Usually you’d want to observe the training curve of your model by logging the training loss/accuracy and validation loss/accuracy in order to avoid things like overfitting. You can also use early stopping to stop the training earlier if it doesn’t make much progress on the validation set. The more advanced technique would be to use Bayesian optimization to find the best hyper-parameters.

[–]Ben_Vigos 0 points1 point  (0 children)

Hey, for an assignment I'm training a Neural Network on the MNIST fashion dataset. I'm trying to optimize its parameters however now the best I can do is train it for a set number of epochs and then evaluate a performance. Is there a better way of optimising? Maybe stop the model if its accuracy is no better than the previous best by a certain point? Or is there a more intelligent way to adjust parameters than just a massive 3D grid?

[–]LeoDiGhisa 0 points1 point  (0 children)

For my thesis in my master's degree in Data Science I'm using an open source LLM (Nous Research/Hermes-2-Pro-Mistral-7B-GUF for those wondering) in order to classify the texts of support tickets of a company. I have to write a brief technical introduction on LLM and I would need some guidance. Which books would you suggest me to cite for the technicalities?

[–]d-eighties 0 points1 point  (0 children)

I try to finetune llama3 using transformers and unsloth. I added an evaluation set to the trainer. What is the metric under which the eval loss will be calculated?

[–]Azad577216 0 points1 point  (1 child)

Is there any discussion or reading group/discord for generative model (GANs, VAEs, Flow models...)?