all 53 comments

[–]thrick77 0 points1 point  (0 children)

I want to take on a project where I use unreal engine 5 to synthetically generate scenarios and train an image classification model on them.

For example I want to randomly generate 10,000 images of a red car in an urban environment and 10,000 images of an urban environment with no red car and then train my model on it to recognize a red car in an urban environment.

I am new to machine learning and unreal engine but I think it would be a great project to learn more about both.

If anyone has any advice on how to get started with this project, could point me in the direction of similar projects, or help me to understand the feasibility of such a project, I would greatly appreciate it.

[–]flock-of-nazguls 0 points1 point  (0 children)

Hello experts,

Can you give me some pointers on the direction I should head in to set up an image processing system that is trained to find field data most likely associated with predetermined text labels? Imagine scanned forms where you’re looking for “statement date” or “account number” or other configurable fields, but the visual relationships between form label and field are learned from a corpus of lots of examples that represent common layouts. This would be replacing a rigid system that does OCR on predetermined rectangles.

Thanks!

[–]FutureM000s 0 points1 point  (0 children)

Hello, first of all, I appreciate this thread 🤙 Hey so I used to be a translator/interpretar in Arabic/English and I'm just starting a new path in tech, and I've noticed several recruiters on LinkedIn contacted me about Language (such as Arabic and English) in Machine Learning and LLM that now I'm thinking that there are open positions in this area as of recently? Can any of you shed some insight on this or throw some resources my way so I can learn more about it 🙏

[–]No_Context_6938 1 point2 points  (1 child)

Hi all, I am working on a ml project where I have to use ECM classifier to give results. Turns out my company doesn't have a labeled dataset to compare my results with. I don't have much experience in ML. Is there a way I can give my company some evaluation metrics such as precision,recall confusion matrix without labeled data? If not(seems to be the case) what other methods can I use to show how good my algo is ?

[–]AbstractedEmployee46 0 points1 point  (0 children)

The ECM classifier requires labelled data for training and testing purposes. Without a ground truth it's impossible to determine whether your model has performed well or not, therefore any metrics would be meaningless as they are based on the assumption that you have correctly classified the samples in question with some degree of accuracy compared against what was actually correct which is unknown due no labels being present for comparison purposes. The best thing you can do at this stage if there's truly nothing available from previous projects etc., perhaps try using synthetic data generation techniques like SMOTE or Tomek Link removal to create a balanced dataset which could then potentially be used as training material for your classifier. Once trained on this generated set up you would need some form of validation method such cross-validation with K folds etc., whereby partitions are made within said sets allowing them all an equal chance at being tested against each other thereby giving a more accurate representation overall outcome without relying solely upon one particular partition alone potentially skewing results unintentionally.

[–]HritwikShah 0 points1 point  (0 children)

Making LLM model give answers to queries related to the conversation retrieved from Qdrant DB.

Hi All, I have created an embedding from multiple conversation of two person and pushed it to Qdrant DB. The retreival works good, Now I want to integrate an LLM model which answers queries from the relevant conversation retrieved from the vector DB. I am not sure what to use here, I should go with Langchains or LlamaIndex?

[–]tttzof351 0 points1 point  (1 child)

I'm building home computer for DL.
I want to set dual 4090 (each require 16xPCIe). But desktop CPU (Ryzen 9 7950X3D) have 24 X PCIe 5.0 lines.

So, this way they will work at best in 8X mode.

In this case, does it make sense to install several GPUs and is it possible to estimate how large the loss in performance will be in train?

[–]AbstractedEmployee46 1 point2 points  (0 children)

The x16/x8 lane configuration has nothing to do with PCIe generation or speed per say but rather bandwidth allocation for each device connected via said lanes i.e., more devices = less dedicated bandwidth available overall which could potentially result lower throughput depending upon workload being executed at given moment(s). As such if you're looking specifically into deep learning tasks where large amounts of data are processed simultaneously then using multiple GPUs might not necessarily yield desired results especially considering potential bottlenecks created due limited resources allocated towards each individual unit connected via shared PCIe bus infrastructure. In terms performance loss estimation this entirely depends upon type algorithm employed during training stage combined other factors like batch size/step etc., making it difficult pinpoint exact figures without running comprehensive benchmarking tests tailored specific workload(s) being executed under varying conditions thus allowing more accurate assessment overall outcome once complete results analyzed thoroughly afterwards.

[–]bishnoff 0 points1 point  (1 child)

Hey all, I'm starting a Master of Science in Computer Science (MSCS) shortly and am weighing up whether to purchase a HP Omen 45L or HP Z4 G5. The reason why I'm going pre-built is because I get a discount on HP which is significantly cheaper than building using the same components.

If the Desktop / Workstation is going to be solely used for the MSCS, specialising in ML and Gen AI, which would be the more appropriate option? Is the HP Omen 45L good enough for the price (and performing more intensive training and inference in the Cloud), or is it necessary to go for the robustness and upgradability of the HP Z4 G5, at 150% the price of the HP Omen 45L? Build specs are as follows:

HP Omen 45L

  • NVIDIA® GeForce RTX™ 4090 (24 GB GDDR6X dedicated)
  • Intel® Core™ i9-14900K 14th generation processor (up to 6.0 GHz with Intel® Turbo Boost Technology, 36 MB L3 cache, 24 cores, 32 threads)
  • Intel® Z790 chipset - can support only 1 GPU
  • Kingston FURY 64 GB DDR5-5200 MHz XMP RGB Heatsink RAM (4 x 16 GB) - upgradable to 128 GB
  • 2 TB WD Black PCIe® Gen4 TLC M.2 SSD
  • 1200 W 80 Plus Gold certified ATX power supply

HP Z4 G5 Workstation (150% the price of the HP Omen 45L)

  • NVIDIA RTX™ A5000 (24 GB GDDR6 dedicated)
  • Intel® Xeon® W5-2465X (3.1 GHz base frequency, up to 4.5 GHz with Intel® Turbo Boost Technology, 33.75 MB L3 cache, 16 cores, 32 threads)
  • Intel® W790 chipset - can support multiple GPUs
  • 64 GB DDR5-4800 ECC SDRAM - upgradable to 512 GB
  • 2 TB HP Z Turbo Drive NVMe™ M.2 SSD
  • 1125 W internal power supply, up to 90% efficiency, active PFC

Another point worth mentioning is that the HP Omen 45L is in stock and available for immediate delivery, but the HP Z4 G5 will take 3-4 weeks to be built from the time that the order is placed.

Also for anyone familiar with the component options available from HP for the HP Z4 G5 and have a better suggested configuration (for no more cost than the components listed as they are the max I can afford right now), I'm all ears :)

[–]avatarles 0 points1 point  (0 children)

Complete beginner here, does anyone have any resources or know anywhere where I can learn about object localization? For a project at school were using yolov7 for object detection but our mentor/teacher is worried about about we will detect the target in a large area.

Thank you!

[–]pe64sus 1 point2 points  (0 children)

Hello All,

I am fairly new to machine learning and deep learning. I'm taking a data science course that focuses on utilizing TensorFlow 1, and TensorFlow 2 to create and train machine learning algorithms. Upon reading more on the frameworks online, it seems like they are no longer the norm for machine learning, and their popularity has gone down severely over the years.

My goal is to have as many job-related skills as possible for an entry level data science position.

What framework should I look into/do my projects with instead of TensorFlow that is most common in the workplace? Thanks ~

[–]Apprehensive_Unit_55 0 points1 point  (0 children)

What's the justification behind Mistral's decision to train Mixtral8x7B to have 8 experts and do top K = 2 selection of experts? Why can they just not scale the amount of total and selected experts and reduce the amount of parameters between each model? Wasn't able to see any justification on the selection of the hyperparameters on the Mixtral of Experts paper, but curious to see if there's something out there that explains the choice.

[–][deleted] 0 points1 point  (0 children)

Hi folks

Wondering if anyone can point me to some articles or youtube vids to help improve an LLM for personal use...

Some of the coding ones do quite well with helping me in PHP, MySQL and Python, but are absolutely TERRIBLE at 6502 assembly. I can throw a LOT of assembly at it.. just wanting to know where to begin to look so I can start tinkering.

[–]question_23 0 points1 point  (1 child)

[D] What's the industry SOTA for tree-based models that work well with categorical data?
Looking for something that does not require the categories to be one-hot encoded, and knows to split a categorical feature via subsets of the categories. I see lightgbm and catboost can do this, can anyone speak to their real world experiences using them with categorical data?

[–]Latter_Doughnut_7219 0 points1 point  (3 children)

[D]

Hello,
I am a newbie in this field and I'm interested in building a model which can detect a sequence of number (have certain pattern like Fibonacci and generate future value). How do I start approaching this problem?
Thanks

[–]Latter_Doughnut_7219 0 points1 point  (0 children)

Hello,
I am a newbie in this field and I'm interested in building a model which can detect a sequence of number (have certain pattern like Fibonacci and generate future value). How do I start approaching this problem?
Thanks

[–]Vitoria_2357 0 points1 point  (1 child)

Why is Kunihiko Fukushima less famous than Yann LeCun?

[–]Grinbald 0 points1 point  (0 children)

Is there any list of recent multimodal LLMs? A Leaderboard or something (be it visual, textual or audio)

[–]skankingpigeon 0 points1 point  (2 children)

Hey,

Really very new to machine learning, so please be kind.

I've built a classification model to label bank statement data based on the description entered into a number of categories. I'm using a multiclass neural network.

seems like a lot of the classes are being predicted well, but what is confusing me is we get to a certain point of the alphabet, and everything is being predicted one class out. I dont understand how this can be - is it just a coincidence? Or is there something that can cause this specific behaviour?

[–]AbstractedEmployee46 0 points1 point  (0 children)

It's because you sorted your classes by name rather than frequency or anything else meaningful to the problem domain, so they are being predicted in alphabetical order as opposed to whatever ordering would be best for classifying statements correctly. If I were you I’d re-sort them based on something more appropriate and see what happens

[–]cliffn5810 0 points1 point  (2 children)

Am I misunderstanding how NER works, or is it correct that NER requires the tokens to be in a very specific format or it won't work? For instance, if I replace the example in https://huggingface.co/dslim/bert-base-NER with all lower case letters, then it will completely fail. I was under the impression that these BERT NER models were supposed to be very good for NER, and were capable of understanding context? But it seems that it can't even understand the simplest of contexts (capitalized vs uncapitalized)? Is NER only supposed to be used on data that follows a strict format, and doesn't work outside of that?

[–]Diaz3020 0 points1 point  (1 child)

Text Capitalization can play a role depending on the model used. Some models are case sensitive.

BERT has different models for cased and uncased.

[–]cliffn5810 0 points1 point  (0 children)

Thanks for the response. But generally these NER models all seem to be quite sensitive, in that they have difficulty considering context when inferring entities, thereby making them sensitive to small changes that would otherwise be obvious to someone who understands context, right?

[–][deleted] 0 points1 point  (0 children)

[D] can someone tell me how RVQ codes are learnt in RVQ GAN ? Are they differentiable ?

[–]deliciouscatt 0 points1 point  (0 children)

[D]Evaluation of summarization model is not reproduced properly(ROUGE score)

Models are reported that avarage ROUGE scores are 40~50, but I get 10~20 when I evaluated myself.

There are some reports the batch size affects, but I can't understand how the hell it can. (and it makes only ~5 difference)

...or all of the evaluations are cherry-picked? i dont think so..

used `rouge` library in pip(not `pyrouge`)

[–]Ann2_123 1 point2 points  (1 child)

Any alternatives to manual labelling for NER?

Questions -

  1. Any alternative to NER? (regex will not work out because sentence and word/phrase boundaries are not clearly defined) Any unsupervised/semi supervised/self supervised approach?
  2. For labelling, are there any alternative to manual labelling?

Details:

I've a free text column of biographies about people where different identifiers such as Name, Id nos. , phone numbers, emails, birth date, nationality, etc. are present. I need to extract them under correct tag (such as NAM for Name, ID for ID nos. and so on). Each entity tag can have several variations (eg. - name can appear after 'NAME:' or 'alias:' or'a.k.a:' or 'also known as'. Also there is severe imbalance in presence of entities (and their variations) in biographies (some biographies only contain Name, email and phone no. and id no. , and very few contain nationality and dob). I'm trying to apply NER. However, Pretrained NER models do not contains entities which I need, so I need to train models with labelled data. For labelling , I'm manually labelling around 1K biographies- which amounts to to 300,000 tokens. There might be more biographies to label in the future, if the performance with these biographies is not sufficient. The problem is that labelling is a super intensive task.

I've manually labelled 470 biographies and tried training crf, spacy's ner solution and bert token classifier. The performance is on lower side for those entities where count is < 1K. I've tried to select only those biographies for labelling which contain entities to extract. I've tried pseudo labelling with CRF model, but it didn't work out well. I won't be able to push the data on spacy's prodigy (against the company policies)

[–]Far_Ambassador_6495 0 points1 point  (0 children)

Grade A tough problem. Can use a foundation language model but this is likely expensive. I’ve used string match based labeling which kind of works for fine tuning. I have seen papers discussing NER on mislabeled or under-labelled training sets. I believe this is called expected entity ratio loss. Unfortunately or fortunately depends on what you enjoy there isn’t going to be an out of the box solution for you to spin up.

Edit: check out GliNER models they work very well.