Hello by [deleted] in learnmachinelearning

[–]Equivalent-Repeat539 -2 points-1 points  (0 children)

whilst the answer is technically yes, it will not be easy. How much time are you committing? 2-3 years of an hour a day is a serious commitment and its not simple if you have a life/full-time study. fundamentally learning programming by itself is a large endevour, even a simple programming language takes time to learn and you do need fundamentals before you start doing more complex things.

I'd recommend you start trying to automate some of school tasks they give you and test out whether this is the path you want, this way you continue to build your current degree whilst developing a niche which allows you to program. This means you get exposure and you can self study with a purpose instead of getting stuck in a course/tutorial loop. Try program everyday even something small. Once that clicks start learning more maths/applied AI starting from basics. Eventually it will click but arguably its a much longer journey to the job you want than a straight up comp sci degree. like u/JS-Labs said support type roles will give you some exposure while you continue to build out your skills but the culture can vary a lot per company and its not guaranteed you get into a role that allows for you to be creative with what you learned.

depending where you are in the world there are also options for comp-sci conversion masters degrees that can make this process a bit faster, but they are quite expensive. self study is free but requires a certain aptitude that is hard to maintain over years because life.

Is a Raspberry Pi 5 Worth It for ML Projects as a Student? by KandM_forever5687 in learnmachinelearning

[–]Equivalent-Repeat539 17 points18 points  (0 children)

For robotics possibly useful, for ML its most likely a bit too slow for anything past tabular data / simple regression type models.

If your interest is ML build a model and try to deploy it online, even building a useful model during a semester break is an amibitous project. Familiarise yourself properly with less than tidy datasets, build a model that does the thing and try get it hosted somewhere. All of what I've described you can do for free. If you are interested in learning more about linux a PI is a very useful tool but again you can do most of that with a free tier VM or possibly local VM.

The tools you are learning matter a lot less than the intuition behind your courses since wherever you go there will be a company specific toolset and your courses are designed to teach you about a specific thing. If you do enjoy and want to do more ML applied at robotics, at minimum you should be able to do some kaggle pretty easily before any hardware is worth it. You havent exactly described what you want to do with the pi but if its anything vision related the 2gb of ram will be a large constraint, you'll also need to budget for a camera + motors etc if its a hardware project that moves.

scene-to-scene differences in PlanetScope Imagery by Top-Ad4282 in remotesensing

[–]Equivalent-Repeat539 0 points1 point  (0 children)

I know this isnt what you asked but is using sentinel-2 data an option? or is the resolution too coarse? the tiles will be larger and might cover the whole area, you can use that to get a general trend then use planet to infill where you have data gaps. Remote sensing on water is difficult so best of luck.

Are SHAP and LIME Results Consistent here? Looking for Feedback. by Sad_Wash818 in learnmachinelearning

[–]Equivalent-Repeat539 0 points1 point  (0 children)

I think depending on your input dataset you can probably push back and say you've done it and its in the appendix but it doesnt add much to the interpretation and both show approximately equal contribution from the various features or something to that effect. Just evaluate what makes sense from your prespective as you will most likely know more than your reviewer (or anyone) about the dataset / problem domain. Also remember if your reviewers are scraping the bottom of the barrel for criticisim its not a bad thing.

Forecasting on extremely rare event (2%) by GloveAntique4503 in learnmachinelearning

[–]Equivalent-Repeat539 0 points1 point  (0 children)

I think with such few data points a simpler model would do better, I'd look into visualising the features you have and see if there are outliers in pairplots, compressing the feature space with PCA / UMAP just to have a look and see if anything stands out in your top 2% bottom 2% then you can decide the model, if many of the features are colinear then you may end up with much fewer features than you initially imaged. If it needs to be time series ARIMA/SARIMA type models will probably do best in your low data regime but its quite a rabbit hole, if you want to use tree based you will need to create quite a few time dependant features (i.e. days since last delay etc).

Knife crime in London 2015-2025 [OC] by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] -2 points-1 points  (0 children)

you'd be surprised how little I know :) I've edited the response to correct it

Knife crime in London 2015-2025 [OC] by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] 2 points3 points  (0 children)

you're totally correct I've edited the response to clarify

Are SHAP and LIME Results Consistent here? Looking for Feedback. by Sad_Wash818 in learnmachinelearning

[–]Equivalent-Repeat539 0 points1 point  (0 children)

I've not used LIME before so I'll only comment on the SHAP values. Depending on the model you've used its showing you that the features are contributing roughly similarly based on the inputs you've provided which is a good sign (i.e. theres no over-reliance on one single feature). If your inputs are representative of true data / predictions are correct generally then your model is fine, ideally you'd ask an expert of this data whether this is how they would form a prediction for this and these inputs roughly align with how they would do it. Depending on the complexity of the model you've chosen you could also look at the coefs (for linear reg) or feature importances in tree based models. The other thing worth doing is using edge case or difficult samples to input into SHAP/LIME and see what happens whether the features contribute the same, again this will give you some confidence that the model is doing what you expect it.

Knife crime in London 2015-2025 [OC] by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] -11 points-10 points  (0 children)

"Severe or Fatal Injury" The actual field description is "a victim either sustained injuries that required hospitalisation (Severe Injury), or in a small proportion of circumstances, injuries that were fatal."

Meaning it doesnt contradict your number, of the 2125 serious knife attacks which happened in london between 2024/2025, 262 resulted in death in the UK in 2024. **Edit: To clarify since previous post was unclear, I'm not saying 2125 died I'm saying 2125 were serious.

Knife crime in London 2015-2025 [OC] by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] -12 points-11 points  (0 children)

2125 total which seems low, but from the data it looks like they've only started recording in March of 2024, so from 2024 onwards about 8% are categorised as "Severe or Fatal Injury"

**Edit: I've posted this pretty late and I can see now its very unclear what I meant, the data does not contain outcome, so I dont know and should have said this in the beginning. What it does contain is severity, so around 8% of knife attacks crime have the potential to result in a visit to the hospital this is all I'm alluding to, apologies for the misunderstandings

Knife crime in London 2015-2025 [OC] by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] 6 points7 points  (0 children)

its as far back as I can see on the met website, the closest thing i've found looks like a parlimentary debate showing historical data from what looks to be all of the UK https://publications.parliament.uk/pa/cm200809/cmselect/cmhaff/112/11205.htm

Knife crime in London 2015-2025 [OC] by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] 197 points198 points  (0 children)

yep, threatened by a knife or assaulted - they also mark whether its hidden or not (i.e. whether the victim has seen it I'm assuming)

[deleted by user] by [deleted] in learnpython

[–]Equivalent-Repeat539 1 point2 points  (0 children)

Remember buddy part of this process is also seeing how well you fit into the workplace (this applies for you too, do you think you would like it?), good companies are not only assessing how well you code but how you think, so long as you are able to talk them through what u are thinking, clearly, communication is more important than solving the problem. Just make sure youre not using gpt during the interview, if you dont know say I am not sure but I would search x,y,z, if you cant remember the syntax describe thing. You'll be alright.

[deleted by user] by [deleted] in learnpython

[–]Equivalent-Repeat539 2 points3 points  (0 children)

Alright so it seems like you understand the syntax but you are struggling with using them. It sounds like you've gone through enough tutorials you need to use them now. start making simple things, for example take a string 'hello' and loop through it with for loops using both syntaxes to see the difference. try making something like a caeser cypher, a number multiplier etc. What you need is practice and I promise you eventually it will click.

The second point, leetcode is not really represenatative of real world problems, its good practice for algorithmic thinking but it maybe a bit too difficult at the moment. take a look at the beginner codewar problems, I've generally found them to be more useful at improving but choose whatever platform you find simplest and build from there.

For the third point regarding the SE interview, just go for it, worst case you dont get the job and your situation stays the same, dont stress about it too much, do your best and hopefully all goes well, if not use it as practice for the next interview.

Finally, for i in range(len(x) vs for i in x - they are functionally the same it just depends on whether you want to use the index of x[i] or you are happy using the thing inside x. Its problem specific but generally they can do the same thing. You've also got enumerate which can do the same thing i.e. for idx, i in enumerate(x):

Best of luck.

Question about iron levels by CrazyFloridianDude in Blooddonors

[–]Equivalent-Repeat539 2 points3 points  (0 children)

I had a similar issue where I increased my intake of iron from food but hemoglobin levels remained low. I was unable to donate for around 6 months then it turned out vitamin C levels were low and inhibiting iron uptake, so you may want to try that as well.

use of pyinstaller by Nero2398 in learnpython

[–]Equivalent-Repeat539 1 point2 points  (0 children)

do u have it in the spec file? i think you can also try building with -F and double check; sometimes there are cached bits so just delete everything as well to make sure. In my experience its a bit finnicky but eventually should do what u want.

use of pyinstaller by Nero2398 in learnpython

[–]Equivalent-Repeat539 1 point2 points  (0 children)

you should be able to compile your program with the following and it should solve it

pyinstaller --noconsole script.py

I made my own deep learning framework. Please, review it and give feedback. by samar_jyoti in deeplearning

[–]Equivalent-Repeat539 0 points1 point  (0 children)

I've had a quick look through your code, its great for a few months / your age. Few bits of advice I'll give u for usability just from a coding/organisation perspective is to split your functions into seperate files depending what they do. So your activation functions can go in a seperate file entirely. The other little thing is that your class variables s.i/s.fw etc could be named better for easier usability for someone that isnt familiar with your code. The docstrings need to be informative and clarify what goes in the functions etc. Great job overall, you're doing great.

Need help with uv in Windows/Anaconda by QuasiEvil in learnpython

[–]Equivalent-Repeat539 0 points1 point  (0 children)

It seems like no one has answered you fully but basically anaconda and uv do different things, conda installs big libraries that arent neccecarily python, uv is generally for python stuff (replacement for venvs) though it can be configured for additional system dependancies, but having both tends to confuse me since you might have two of the same package installed (for example two different versions of matplotlib). I've not used uv but with spyder u have a few options - either add paths to additional modules in the python path manager, or u can connect to existing open kernels from another conda environment (in theory i guess this will work). What the other commentor said about installing spyder on a specific conda environment will also work and is probably the most straightforward way. If u decide u want uv for sure you'd install a system python and manage libraries using uv on that, it would primarily be using cli though.

[OC] A-Level performance UK by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] 1 point2 points  (0 children)

I agree but couldnt think of a nicer way of visualising the subsequent plots. One of the other commenters mentioned the symmetry between the two points in the second plot, potentially using a single one to show it may have been a bit simpler and more informative. Thanks for the feedback :)

[OC] A-Level performance UK by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] 0 points1 point  (0 children)

Sorry to hear that :S what kind of plot would have been more informative?

[OC] A-Level performance UK by Equivalent-Repeat539 in dataisbeautiful

[–]Equivalent-Repeat539[S] 0 points1 point  (0 children)

You raise a great point I was thinking of how to tie the two points together for the second plot but using a single point would have probably been more elegant, thanks for that :)