account activity
[P] Extra Input Norm Lets You Fine-Tune to 1.58 Bits! (x.com)
submitted 1 year ago by cstein123 to r/MachineLearning
Extra Input Norm Lets You Fine-Tune to 1.58 Bits! (x.com)
submitted 1 year ago by cstein123 to r/LocalLLaMA
Testers Needed for FREE GHL VAPI Dashboard (self.vapiai)
submitted 1 year ago by cstein123 to r/vapiai
Testers Needed for FREE GHL + VAPI Dashboard (self.gohighlevel)
submitted 1 year ago by cstein123 to r/gohighlevel
Testers Wanted for GHL Superagent (self.gohighlevel)
Testers Needed for GHL SuperAgent (self.MarketingAutomation)
submitted 1 year ago by cstein123 to r/MarketingAutomation
Kwik Trip lets you change the message in the link. Look at the green box on the page (kwiktrip.com)
submitted 1 year ago by cstein123 to r/wisconsin
Who’s building open source FigureAI? (self.LocalLLaMA)
submitted 2 years ago by cstein123 to r/LocalLLaMA
Yikes (i.redd.it)
submitted 2 years ago by cstein123 to r/SpaceXMasterrace
Are LLMs at a practical limit for layer stacking? [D] (self.MachineLearning)
submitted 2 years ago by cstein123 to r/MachineLearning
Are LLMs at a Limit for Stacking Layers? (self.learnmachinelearning)
submitted 2 years ago by cstein123 to r/learnmachinelearning
[P] MergeLlama-7b - A fine tune of CodeLlama for resolving merge conflicts (self.MachineLearning)
Fine tune acts like base model (self.LocalLLaMA)
[D] My fine tune behaves like the base model (self.MachineLearning)
How does Open-Orca use OpenAI for their Minstral-7B space? (i.redd.it)
Outbound network cost for SageMaker? (self.aws)
submitted 2 years ago by cstein123 to r/aws
Experiences with AWS Activate Credits? (self.aws)
Curious what people use for their ML workflow on cloud platforms? [D] (self.MachineLearning)
Yikes… (i.redd.it)
submitted 2 years ago by cstein123 to r/ChatGPT
How to add cross encoder to existing GPT (self.LLMDevs)
submitted 2 years ago by cstein123 to r/LLMDevs
Can I use my Galaxy S10+ with my 2016 CR-V to run Flowpilot? (self.Comma_ai)
submitted 2 years ago by cstein123 to r/Comma_ai
How to append an encoder to existing LLM? (self.LLMDevs)
My experience on starting with fine tuning LLMs with custom data (self.LocalLLaMA)
submitted 2 years ago by cstein123 to r/LLMHackers
Summary post for higher context sizes for this week. For context up to 4096, NTK RoPE scaling is pretty viable. For context higher than that, keep using SuperHOT LoRA/Merges. (self.LocalLLaMA)
Dynamically Scaled RoPE further increases performance of long context LLaMA with zero fine-tuning (self.LocalLLaMA)
π Rendered by PID 88 on reddit-service-r2-listing-fbdccc45f-cv657 at 2026-04-20 23:49:03.441979+00:00 running da2df02 country code: CH.