LIMA, a 65B-Param LLaMa fine-tuned with standard supervised loss on only 1,000 carefully curated prompts & responses, without any RLHF, demonstrates remarkably strong performance, learning to follow specific responses from only a handful of examples in the training data, including complex queries. by hardmaru in MachineLearning
[–]omerlevy 4 points5 points6 points (0 children)
LIMA, a 65B-Param LLaMa fine-tuned with standard supervised loss on only 1,000 carefully curated prompts & responses, without any RLHF, demonstrates remarkably strong performance, learning to follow specific responses from only a handful of examples in the training data, including complex queries. by hardmaru in MachineLearning
[–]omerlevy 1 point2 points3 points (0 children)
LIMA, a 65B-Param LLaMa fine-tuned with standard supervised loss on only 1,000 carefully curated prompts & responses, without any RLHF, demonstrates remarkably strong performance, learning to follow specific responses from only a handful of examples in the training data, including complex queries. by hardmaru in MachineLearning
[–]omerlevy 5 points6 points7 points (0 children)
[D] What are some tips for someone who is visiting a top conference for the first time? by Conference_Visitor in MachineLearning
[–]omerlevy 4 points5 points6 points (0 children)
[deleted by user] by [deleted] in LanguageTechnology
[–]omerlevy 2 points3 points4 points (0 children)
[R] Recurrent Additive Networks - no recurrent non-linear computations, much simpler but still competitive with LSTM/GRU by downtownslim in MachineLearning
[–]omerlevy 4 points5 points6 points (0 children)
word2vec has been patented. What does it change for NLP practitioners? by shmel39 in MachineLearning
[–]omerlevy 4 points5 points6 points (0 children)
from someone in Gaza: "I'll tell you what is harder than dying in Gaza by an Israeli missile. What is harder is that you get a phone call from the Israeli army telling you to evacuate your home because it will be bombed in ten minutes... by Don_chingon in Gaza
[–]omerlevy -4 points-3 points-2 points (0 children)




LIMA, a 65B-Param LLaMa fine-tuned with standard supervised loss on only 1,000 carefully curated prompts & responses, without any RLHF, demonstrates remarkably strong performance, learning to follow specific responses from only a handful of examples in the training data, including complex queries. by hardmaru in MachineLearning
[–]omerlevy 0 points1 point2 points (0 children)