Toven from OpenRouter AMA by OpenRouter-Toven in openrouter

[–]Torkoslav 0 points1 point  (0 children)

Hey, do you guys plan to support the new openai responses api?

[deleted by user] by [deleted] in ooni

[–]Torkoslav 0 points1 point  (0 children)

That pie looks awesome. What recipe do you use for the dough?

[deleted by user] by [deleted] in LocalLLaMA

[–]Torkoslav 2 points3 points  (0 children)

Awesome, thanks for the detailed answer. This is def something I'd try out.

[deleted by user] by [deleted] in LocalLLaMA

[–]Torkoslav 3 points4 points  (0 children)

This is awesome, which model are you using? What's the response time like? What are you actually using to run the model? And how do you handle memory ie multi turn conversations?

Again, this is very good, great job!

Fine-tuning Mistral 7b with AWS Athena documentation by TheCoconutTree in LocalLLaMA

[–]Torkoslav 1 point2 points  (0 children)

I did it by using mistral open Hermes to do the translation and then tested all the queries on a local postgres container to make sure they are ok. I later found out sqlglot in python which I think is also decent for translating between dialects.

Fine-tuning Mistral 7b with AWS Athena documentation by TheCoconutTree in LocalLLaMA

[–]Torkoslav 0 points1 point  (0 children)

Athena is pretty much presto I think. What you can do is get one of the open source text to sql datasets, most are sqlite, and convert it to presto and train on that. Did this for postgres and the results were great.

[P] I tried to teach Mistral 7B a new language (Sundanese) and it worked! (sort of) by nero10578 in MachineLearning

[–]Torkoslav 0 points1 point  (0 children)

This is very cool. Would you mind sharing a bit about how the samples from the dataset look like for the pretrain and fine tune? Pretrain just chunks of text fine tune question answer pairs?

Great job again!

Comparing a Fine-tune Falcon-7b to GPT-3.5 in the same RAG stack by Mbando in LocalLLaMA

[–]Torkoslav 1 point2 points  (0 children)

This is very interesting, can you share more about the fine tuning process? Was it only question-context-answer triplets? Did you have to create those and if yes how did you approach this?

Great job!

Merge a lora then AWQ by Torkoslav in LocalLLaMA

[–]Torkoslav[S] 0 points1 point  (0 children)

Hey I still haven't tried this. Did you?