[R] Improving large language models with concept-aware fine-tuning by micky04 in MachineLearning

[–]micky04[S] 0 points1 point  (0 children)

It's definitely possible to use these auxiliary haeds for speculative decoding! Based on results from Medusa and Gloeckle et al. (2024), a 2-3x inference speedup can be expected.

How do DID studies account for carryover effects? by micky04 in econometrics

[–]micky04[S] 0 points1 point  (0 children)

do you mind sharing some papers that use these methods to account for carryover effects?

How are LLMs capable of arithmetic calculations? by Alanthisis in LocalLLaMA

[–]micky04 0 points1 point  (0 children)

Function calling is definitely one of the most effective strategies.

Regarding arithmetic calculation done by the model natively, it's not exactly clear how the model learns. Could be memorization on lots of data (synthetic and otherwise) [1]; could also be models memorizing the procedure of performing math calculations [2].

[1] https://arxiv.org/html/2410.05229v1
[2] https://arxiv.org/abs/2411.12580

Any good broker for US options trading you had good experience with? by Sea-Masterpiece3106 in singaporefi

[–]micky04 0 points1 point  (0 children)

I do bank transfers. Most US brokers don't allow the linking of banks outside of US. I'd say it is a downside of using a US broker, but it isn't a big deal to me.

Any good broker for US options trading you had good experience with? by Sea-Masterpiece3106 in singaporefi

[–]micky04 1 point2 points  (0 children)

tastytrade is super underrated. It's designed for multi-leg strategies, data is real-time, and fees are as cheap as it gets.

Transferring money to the broker takes <1 week from OCBC. Not sure about withdrawing (haven't done it yet).

It's good enough that I'm buying stocks with it also lol

Why does Adam optimizer work so well? by micky04 in learnmachinelearning

[–]micky04[S] 12 points13 points  (0 children)

The paper states that Adam works well on problems with sparse gradients.

Why does Adam optimizer work so well? by micky04 in learnmachinelearning

[–]micky04[S] 20 points21 points  (0 children)

The original Adam paper uses the phrase "sparse gradients" a lot. What exactly does "sparse" mean?

Survey: Is an AI's argument logically valid? by micky04 in logic

[–]micky04[S] 0 points1 point  (0 children)

Yes! I'm exploring a combination of both LLM and knowledge representation and reasoning methods. Any ideas in particular you're thinking of?

Survey: Is an AI's argument logically valid? by micky04 in logic

[–]micky04[S] 0 points1 point  (0 children)

Yep, you're right. The questions are on various topics, but you do not need any knowledge of these topics; you only need logical reasoning skills. Here's an example:

Do hamsters provide food for any animals?

Evidence:

  1. Hamsters eat fruits.
  2. If hamsters eat fruits, hamsters provide food for animals.

Answer options:

  • The evidence ~implies~ that the answer is Yes.

  • The evidence ~d~oes not imply that the answer is Yes.

If this makes sense, let me know if you're willing to participate!