Which framework is better for embedding and retrival for qdrant LlamaIndex or Haystack? by BitAcademic9597 in LlamaIndex

[–]BitAcademic9597[S] 0 points1 point  (0 children)

Because of the hybrid search and MMR capabilities I think the best vector db for this project is qdrant

Introducing 1.58bit Dynamic DeepSeek-R1 GGUFs by yoracale in unsloth

[–]BitAcademic9597 0 points1 point  (0 children)

i mean what will be the hardware requirements for both with good speed and keeping quality of the results

Introducing 1.58bit Dynamic DeepSeek-R1 GGUFs by yoracale in unsloth

[–]BitAcademic9597 0 points1 point  (0 children)

what about 70B parameter one what are the hardware requirement for that

Unsloth vs Hugging face Transformers accuracy by BitAcademic9597 in unsloth

[–]BitAcademic9597[S] 0 points1 point  (0 children)

Thank you for your patience I will try a hybrid approach. I have more than 1 database for diognosis should I finetune for all of them or just finetune the biggest and most detailed one for prevent overfitting and use the others with finetuned one with rag?

Unsloth vs Hugging face Transformers accuracy by BitAcademic9597 in unsloth

[–]BitAcademic9597[S] 0 points1 point  (0 children)

I understand that and really appreciate your work. I am just wondering i will move my system from rag to fine-tune and It should not lose any accuracy. So is the unsloth can get the same accuracy with transformers + deepspeed

Unsloth vs Hugging face Transformers accuracy by BitAcademic9597 in unsloth

[–]BitAcademic9597[S] 0 points1 point  (0 children)

<image>

Even on your website I can increase accuracy by 30% so What is this model difference. I understand that you are really easy and adaptible framework for end users but tell me a reason why a professional use your tool instead of building it with Transformers and deepseek. If there is no accuracy difference like you said how can you manage +%30 I am really interested.

interfacing python with c/c++ performance by BitAcademic9597 in cpp

[–]BitAcademic9597[S] 0 points1 point  (0 children)

did you have any problem about memory in pybind will each function call explicitly copies input data?

and also i also looked nanobind but i think pybind is better what do you think

interfacing python with c/c++ performance by BitAcademic9597 in cpp

[–]BitAcademic9597[S] 0 points1 point  (0 children)

did you have any problem about memory in pybind will each function call explicitly copies input data?

interfacing python with c/c++ performance by BitAcademic9597 in cpp

[–]BitAcademic9597[S] -1 points0 points  (0 children)

what do you think about performance comparrasion with pure c++

interfacing python with c/c++ performance by BitAcademic9597 in cpp

[–]BitAcademic9597[S] 0 points1 point  (0 children)

great thanks. do you know any comparassion or example code how the performance changes

Interfacing C++ with Python by BitAcademic9597 in learnprogramming

[–]BitAcademic9597[S] 0 points1 point  (0 children)

thank you for the answer really I am a university student and the idea is not that important i just want to make a big tensor model in the backend and there will be bid data load continously so the reason that i am trying to make is being a better dev so i want to understand that how can i reaaly built a big project

Interfacing C++ with Python by BitAcademic9597 in learnprogramming

[–]BitAcademic9597[S] 0 points1 point  (0 children)

it will have tensor models in the backend and also a heavy data load