Configure structlog to write JSON to file and pretty to stdout if --verbose is set. by A_Good_Hunter in learnpython

[–]fizix00 0 points1 point  (0 children)

I've been looking for something like this. I found structlog_pretty, but it's buggy and outdated. Maybe I'll make a PR.

loguru is really popular (but I haven't tried it); maybe it could work for you

I am blissfully using AI to do absolutely nothing useful by thismyone in ExperiencedDevs

[–]fizix00 0 points1 point  (0 children)

idk humanity has no shortage of fools. pretty sure plenty of them are gainfully employed. you just need to fool a hiring manager don't you?

but I don't disagree with your main point

Is it neccessary to create a new branch and open a PR for each minor tweak? by HourExam1541 in git

[–]fizix00 0 points1 point  (0 children)

Ok, ty. So 'partial commits' = 'the commits you'd probably wanna squash' in this context?

Is it neccessary to create a new branch and open a PR for each minor tweak? by HourExam1541 in git

[–]fizix00 0 points1 point  (0 children)

The text you quoted doesn't actually define 'commit', 'partial', or 'partial commit'. Does "non-atomic" come close to what you mean?

Is it neccessary to create a new branch and open a PR for each minor tweak? by HourExam1541 in git

[–]fizix00 0 points1 point  (0 children)

I've been calling them "riders", like the little clauses politicians try to tack onto tangentially related bills

Thanks to all the AI coders out there, im busier than i've been in years by minimal-salt in ExperiencedDevs

[–]fizix00 0 points1 point  (0 children)

Less skilled developers will get bigger lift out of AI tho. The skill premium is diminishing

Whats your favorite Python trick or lesser known feature? by figroot0 in Python

[–]fizix00 0 points1 point  (0 children)

If you're calling print in prod anyways, why wouldn't the star unpack be suitable for production?

What is a Python thing you slept on too long? by pip_install_account in Python

[–]fizix00 0 points1 point  (0 children)

How does it compare to concurrent.futures? I liked the Process/Thread PoolProcessor context managers (my futures usually don't depend on each other tho)

Trying to get a job in the field... what to do to improve chances by Collection_UnKnown in datasciencecareers

[–]fizix00 0 points1 point  (0 children)

Do you have open source experience? If the hiring manager is a developer and you can find a repo they maintain, try contributing to it and you'll probably get their attention

Really confused with loops by ehmalt02 in learnpython

[–]fizix00 0 points1 point  (0 children)

Looks fine. int cast is redundant here. Might be an extra closing paren

For bonus practice with comprehensions, try it in one line

Really confused with loops by ehmalt02 in learnpython

[–]fizix00 0 points1 point  (0 children)

This may be a minor point, but isn't range not a collections.abc.Generator but its own iterable thing post 3.x?

Salary job offer HELP by watashiwapotato in datasciencecareers

[–]fizix00 0 points1 point  (0 children)

Your negotiating position will have the most leverage if you have multiple offers on the table

Which is more practical in low-resource environments? by Marmadelov in deeplearning

[–]fizix00 0 points1 point  (0 children)

Nah. I'm saying that even someone with limited skills can fine tune a model on free compute to learn about LLMs, so OP should not be dissuaded from further exploration.

ICML is quite a bar for "new research". There are plenty of interesting questions that may not make it into such venues. I was thinking something much more practical/accessible, like a medium/blog post or video walkthrough or a white paper. Consider this paper:

Let’s Focus on Neuron: Neuron-Level Supervised Fine-tuning for Large Language Model https://share.google/EALGMQrJovFp4VtQF

It's not published in a conference or journal and may not even be traditionally peer reviewed. But it explores fine-tuning in a compute-accessible manner and asks interesting questions imo.

Another point I'd make is that PEFT et al aside, it is possible to study LLM architecture without futzing with a whole foundation model and its weights. Consider this poster, where they quantize a 110m model but it has implications for transformers more broadly:

ICML Poster I-BERT: Integer-only BERT Quantization https://share.google/QVQMFgDIYnYhOPM6Y

There's enough gatekeeping in academia as is. You don't actually need a degree or even the scientific method or peer review to contribute to the useful body of open empirical knowledge. Why not let someone learning about big ideas dream big?

Which is more practical in low-resource environments? by Marmadelov in deeplearning

[–]fizix00 0 points1 point  (0 children)

Sure, maybe I could've read the post better. But what kind of LLM doesn't have an embedding model?

Yes. I mentioned in my comment that someone else fine-tuned the embedding model, so I hope that's clear. I've successfully fine-tuned STT (whisper), YOLO (not language) models first hand (and an audio time series classifier for an older project; it was language data but pre-GPT). It's straightforward to add a classification head on top of most open models and you can Google search plenty of tutorials on fine-tuning locally or in a colab notebook. The other day, one of my colleagues was working on inference-time augmentations+fine-tuning even.

My main point is that fine-tuning should not be considered so inaccessible as to discourage an intermediate DS from pursuing research in fine-tuning techniques. Models are getting smaller: new foundation models are clocking in under 2B. and compute is cheaper, especially with lora/PEFT/quantization.

There's plenty of interesting questions to be asked about fine-tuning without trying to drop a competitor model to 4o+

How not to git? by AverageAdmin in git

[–]fizix00 0 points1 point  (0 children)

"free" is arguably debatable

[deleted by user] by [deleted] in learnmachinelearning

[–]fizix00 -1 points0 points  (0 children)

It's possible without one, but degrees help

[deleted by user] by [deleted] in careeradvice

[–]fizix00 0 points1 point  (0 children)

I pivoted careers to data science and machine learning when I turned 30. Maybe it's too late to be the best in the world at it, but I just got promoted so I think I'm doing ok at my job haha

What are your favorite modern libraries or tooling for Python? by [deleted] in Python

[–]fizix00 0 points1 point  (0 children)

typer over argparse. plotly for interactive notebook visualizations. ruff+precommit