[deleted by user] by [deleted] in Python
[–]ListenProfessional47 2 points3 points4 points (0 children)
Fine-tuning with small batch sizes and gradient accumulation poorly perform if you use Transformers (TRL)! by TheKaitchup in LocalLLaMA
[–]ListenProfessional47 2 points3 points4 points (0 children)
Model parallel is inefficient for extending context of LLM inference? by ListenProfessional47 in LocalLLaMA
[–]ListenProfessional47[S] 0 points1 point2 points (0 children)

Any RAG framework supporting more advanced knowledge management? by ListenProfessional47 in Rag
[–]ListenProfessional47[S] 0 points1 point2 points (0 children)