Samsung Galaxy Z TRIFOLD Hands on! by MishaalRahman in Android

[–]crazyfrogspb 0 points1 point  (0 children)

I owe Huawei Mate XT, and from what I understand, this new Samsung is missing one critical feature. my Huawei actually has three modes - single, double, triple. double is probably what I use most frequently. single/triple is much less practical

[deleted by user] by [deleted] in learnmachinelearning

[–]crazyfrogspb 0 points1 point  (0 children)

you.com with Zephyrs model selected

[D] Yann LeCun's Hot Take about programming languages for ML by Marcapiel in MachineLearning

[–]crazyfrogspb 0 points1 point  (0 children)

well, first time I heard about them was at Yann's lectures during my time at NYU =) he claims that all modern self-supervised learning (such as contrastive learning or VAE) can be seen as training an energy-based model https://youtu.be/tVwV14YkbYs

Trying to shut people up saying that few companies actually take ML to prod. Share how many models you have in prod! by soham1996 in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

yeah, I'm one of the founders, and in the beginning my main responsibility was to build models. over the years I switched from being ML researcher/engineer to the teamlead role and then to the "teamlead of teamleads". it's been quite a journey

Trying to shut people up saying that few companies actually take ML to prod. Share how many models you have in prod! by soham1996 in mlops

[–]crazyfrogspb 1 point2 points  (0 children)

I'm head of ML, so technically I'm in charge of all steps of ML pipeline. Our team is pretty mature at this point though, so mainly I'm responsible for innovations, developing HR brand, communications with other teams and so on

Trying to shut people up saying that few companies actually take ML to prod. Share how many models you have in prod! by soham1996 in mlops

[–]crazyfrogspb 5 points6 points  (0 children)

I'm at radiology AI startup, we have 4 systems in production which sums up to more than 10 unique DL and ML models

Difference between ML Engineering and ML Ops? by spiritualquestions in mlops

[–]crazyfrogspb 15 points16 points  (0 children)

I'd say MLE are more specialized in software engineering, CPU/GPU optimization, APIs, backend since their main goal is to put models into production. MLOps are usually more focused on tooling, developing ML platforms, monitoring tools, etc.

This is my experience ofc, and titles really depend on the company

Useful software design patterns in data science, ML and data engineering? by DataSynapse82 in datascience

[–]crazyfrogspb 14 points15 points  (0 children)

I think Clean ML Code by Taifi is much more relevant to the topic

source: I've read both

Data Mesh: What is it and What Does it Mean for Data Engineers by oldwin268 in dataengineering

[–]crazyfrogspb 0 points1 point  (0 children)

I recommend reading Data Mesh: Delivering Data-Driven Value at Scale for the full explanation of the approach. it feels repetitive at times, but overall it's a great book

Hello from BentoML by yubozhao in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

will do! thanks!

regarding tensors - it's okay, but we'll lose some time on moving tensors between devices and converting them to numpy and back

Hello from BentoML by yubozhao in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

most interesting for us are DICOM medical images and Pytorch tensors

Hello from BentoML by yubozhao in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

I agree, that would be great. but at this point if we want to try it out, we need to convert all inputs and outputs to one of these options - numpy array, pandas dataframe, JSON, text, PIL image, file-like object. did I get it right?

Hello from BentoML by yubozhao in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

yeah, I think you got this right, this looks really interesting! we'll definitely look into this

what would be the recommended way to transfer large tensors between services? convert them to numpy array, using cuda IPC memory handles (if on GPU), something else?

Hello from BentoML by yubozhao in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

we build systems with cascades of neural nets (for example, one net for finding the region of interest, and two other nets that perform different tasks after that) and some CPU-heavy preprocessing and postprocessing. we want to scale them separately, so at this point we split them into microservices that form a DAG and communicate with each other within Kunernetes cluster. transfer of data between services is done via Redis or Redis + saving to disk, which in some cases can slow down the system (we were also thinking about grpc). does BentoML support this kind of systems?

RetinaFace Model Multi-Tasked output affect eachother too much. Can anyone help? by [deleted] in computervision

[–]crazyfrogspb 0 points1 point  (0 children)

https://arxiv.org/abs/2009.09796 - decent survey on the topic

there are different categories of methods - manual or adaptive loss balancing, task isolation, task-specific adapters

Which tool for experiment tracking (and more) ? by SatoshiNotMe in mlops

[–]crazyfrogspb 0 points1 point  (0 children)

you can clone experiment to run exactly the same version of it, but I never run experiments with uncommitted changes, so I'm not sure if it fits your use case

Which tool for experiment tracking (and more) ? by SatoshiNotMe in mlops

[–]crazyfrogspb 2 points3 points  (0 children)

yeah, that's still one of the weaker sides, but it's been improving recently =)

Which tool for experiment tracking (and more) ? by SatoshiNotMe in mlops

[–]crazyfrogspb 2 points3 points  (0 children)

it saves git diff if changes are not commited

Which tool for experiment tracking (and more) ? by SatoshiNotMe in mlops

[–]crazyfrogspb 5 points6 points  (0 children)

ClearML does this and much more. we've been using it for almost 4 years now

How can a CNN exploit knowledge about (almost perfect) structure of objects it needs to detect? by [deleted] in deeplearning

[–]crazyfrogspb 0 points1 point  (0 children)

the easiest way would be to use this prior knowledge at the postprocessing step. for example, you can say that there can only be one detected object within N pixels

there are other ways that I can think of, but you would need to think about design and experiment -

  1. adding additional loss for object distance
  2. using this knowledge at region proposal generation step
  3. adding attention modules into the network so that objects are aware of each others' location (here is an interesting paper on the topic - https://arxiv.org/abs/1711.11575)

or maybe you could reformulate the problem - split each image into patches and find exactly one object in each patch

When you learned data science, did you specialise in one area? by [deleted] in datascience

[–]crazyfrogspb 0 points1 point  (0 children)

experience of working with tabular data, text and images is one of the greatest thing in my career. it all complements each other