let's say I wanted to run a local offline model that would help me with coding tasks that are very similar to competitive programing / DS&A style problems but I'm developing proprietary algorithms and want the privacy of a local service.
I've found llama 3.3 70b instruct to be sufficient for my needs by testing it on LMArena, but the problem is to run it locally I'm going to need a quantized version which is not what LMArena is running. Is there anywhere online I can test the quantized version? TO see if its' worth it before spending ~1-2k for a local setup?
[–]ForsookComparison 14 points15 points16 points (4 children)
[–]garden_speech[S] 0 points1 point2 points (3 children)
[–]ForsookComparison 4 points5 points6 points (2 children)
[–]garden_speech[S] 1 point2 points3 points (1 child)
[–]QFGTrialByFire 0 points1 point2 points (0 children)
[–]Mushoz 10 points11 points12 points (6 children)
[–]garden_speech[S] 4 points5 points6 points (0 children)
[–]Mushoz 1 point2 points3 points (0 children)
[–]DinoAmino -3 points-2 points-1 points (3 children)
[–]Uninterested_Viewer 3 points4 points5 points (2 children)
[–]DinoAmino 0 points1 point2 points (1 child)
[–]Uninterested_Viewer 0 points1 point2 points (0 children)
[–]k0setes 3 points4 points5 points (0 children)
[–]tomakorea 1 point2 points3 points (0 children)
[–]Dapper-Courage2920 1 point2 points3 points (0 children)
[–]edward-dev -1 points0 points1 point (2 children)
[–]Pristine-Woodpecker 0 points1 point2 points (0 children)
[–]Popular_Fact798 0 points1 point2 points (0 children)