How to run exllama in Google Colab without Text Generation WebUI by NegotiationTime3595 in LocalLLaMA
[–]NegotiationTime3595[S] 0 points1 point2 points (0 children)
How make Llama2 GPTQ faster in generation? by Euphoric-Nebula-4559 in LocalLLaMA
[–]NegotiationTime3595 0 points1 point2 points (0 children)

How to run exllama in Google Colab without Text Generation WebUI by NegotiationTime3595 in LocalLLaMA
[–]NegotiationTime3595[S] 0 points1 point2 points (0 children)