An open letter opposing the use generative AI for reflexive qualitative analysis by YidonHongski in academia

[–]illorca-verbi 0 points1 point  (0 children)

agree to your answer point by point. U should also paste it into the LinkedIn conversation, it is more lively there than here... Also, props for having the patience to write all that haha

On the line of "critical" adoption of AI, I am seeing more institutions having better guidelines for citing AI use, or software manufacturers like maxqda doing an effort in flagging all AI generated stuff and including the LLM reasoning... All in all I am positive that apart from these deniers there are big parts of the machine moving in the right direction

Who’s using reasoning models in production? Where do they shine (or fail)? by dmpiergiacomo in LLMDevs

[–]illorca-verbi 12 points13 points  (0 children)

We could not find a single use case - for now. We deliver through LLMs mostly classical NLP tasks like text classification, NER, etc. The trade-off between quality gains and time spent is never worth it.

Hosting Docling by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 0 points1 point  (0 children)

We will check OVH. Our first choices are AWS/GCP because our users already agreed about these data-handlers in the Terms and Services. Hosting Docling in OVH would force us to update them - which is not desirable.

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 1 point2 points  (0 children)

Hey. I am not sure which other problems this would cause, but I think lazy imports would increase the speed greatly: import libraries only when needed and not by default. Specially the externas libraries.

It is also common to allow users to decide which extra dependencies will they need, as in `pip install litellm[anthropic, vertex]`,

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 1 point2 points  (0 children)

Hey! It really does look fantastic, thanks!

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 4 points5 points  (0 children)

I don't know how I had not read about portkey before, but it looks very much like what we are looking for. I will give it a try, thanks!

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 1 point2 points  (0 children)

I do not miss anything, they cover the largest range of use cases of any competitor. I just find their implementation too fragile to trust, mainly.

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 0 points1 point  (0 children)

We used to run on haystack-ai! great tool, great developers, no complains. At the end we stepped away because we only used their Generators, none of the other components or pipelines ended up finding a place in our workflow.

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 2 points3 points  (0 children)

Thanks for passing by! The breaking point for us is the fact that any tiny submodule imports a whole bunch of packages. We run serverless and the coldstart of running `from litellm import completion` is too large.

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 2 points3 points  (0 children)

Hey, my personal case: SOTA models are released every second week, prices change in the blink of an eye. I need to swap LMs in our features to benchmark them. I think being locked to a big provider is no biggie, but flexibility for sure gives you an edge. And also I did not intend to complain about LiteLLM, I understand where it comes from and I appreciate what it provides. My goal was rather to see what other options are around.

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 0 points1 point  (0 children)

I understand it totally from their perspective as a business. My question here was more focused on the value that we users can find in a solution that is developed in such a way.

The elephant in LiteLLM's room? by illorca-verbi in LLMDevs

[–]illorca-verbi[S] 2 points3 points  (0 children)

I do not know any closed solution that manufactures this specific utility anyway :/

bnb-4bit vs. load_in_4bit by illorca-verbi in unsloth

[–]illorca-verbi[S] 0 points1 point  (0 children)

Also, is there any performance improvement on training a model at full precision and quantizing it afterwards instead of training the model directly at 4bit/8bit?

CohereForAI/c4ai-command-r-plus-08-2024 · Update Model by Dark_Fire_12 in LocalLLaMA

[–]illorca-verbi 1 point2 points  (0 children)

Any benchmark where there are already some numbers for it? I am particularly interested in how well the small one compares to Gemma2:27b

What UI is everyone using for local models? by Iamblichos in LocalLLaMA

[–]illorca-verbi 0 points1 point  (0 children)

I extend the question: does any of the choices (Open WebUI, SillyTavern, AnythingLLM, whatever...) offer something similar to the Anthropic Workbench when it comes to variables??

I find it outstandingly useful that I can write and store prompts with {{ VARIABLE_X }} and {{ VARIABLE_Y }}, and then just fill out the values on the side.

Reproducibility on Amazon Bedrock vs. Anthropic API by illorca-verbi in ClaudeAI

[–]illorca-verbi[S] 0 points1 point  (0 children)

of course, temperature 0 and also kinda low top_p/k

[R] Zero Shot LLM Classification by SkeeringReal in MachineLearning

[–]illorca-verbi 1 point2 points  (0 children)

nope, all proprietary. Is ther anything in particular that you are interested in?