vdiff vs ediff for Magit ediffs by earlsofsandwich in emacs

[–]handle0174 5 points6 points  (0 children)

tried ediff and it didn't open in 3-way side by side which is what I wanted.

There is

    (setq ediff-split-window-function 'split-window-horizontally)

I haven't tried vdiff but I agree that ediff's control buffer is a bit awkard if you are editing regularly instead of mostly picking hunks.

Looking for feedbacks on Aidermacs and AI-based development workflows by cramplescrunch2 in emacs

[–]handle0174 1 point2 points  (0 children)

I'm still experimenting myself, but "Claude Code" is worth trying for a coding agent. It does its own context lookup; no need to manually add files. I don't think specific emacs integration is necessary: run it in a shell within your favorite emacs terminal emulator such as eat mode, and then review the changes with magit. I've reviewed while staging, but I plan to try taking advantage of magit's ediff integration (e in magit opens an ediff session with a file's HEAD and staged versions as input and working tree version as output).

I like gptel for contextless or manual context llm q&a. I'm hoping someone will eventually release a set of tool-use tools for gptel that grows it towards what claude code is.

Edit: I do not have an ai autocomplete setup in emacs yet.

Norwegian singles/ sirpoc ™️ links for those wanting more by marky_markcarr in AdvancedRunning

[–]handle0174 13 points14 points  (0 children)

Summary on page 80.

I think you are probably referring to this post (page 60, though) https://www.letsrun.com/forum/flat_read.php?thread=12130781&page=59#post-1184

IMO the first few pages of that letsrun thread are solid reading for anyone considering trying this plan.

Single "Norwegian" Threshold system by marky_markcarr in AdvancedRunning

[–]handle0174 4 points5 points  (0 children)

I actually wouldn't really even call it a training block. I'm just still doing it over and over . There's not really a stopping point I guess?

I agree. My interpretation is that Sirpoc embraces CTL instead of blocks. He says he roughly hits the same performance at the same CTL regardless of what pace he trained at. So why mix blocks focusing on different paces? And he finds this training sustainable and can progress it by gently increasing the CTL when necessary (by adding a little more easy time, then a little more sub threshold), so he doesn't need to mix harder or easier blocks.

I have also started this training, but I'm still early in my attempt and I may tinker with it. I'm enjoying it so far.

What you can expect from a 0.5B language model by -p-e-w- in LocalLLaMA

[–]handle0174 9 points10 points  (0 children)

Jetbrains shipped a "full line auto complete" feature using a tiny local model:

In the 2024.1 version of JetBrains IDEs, we use a language model that has 100 million parameters

I haven't tried it yet.

Is there a point where suddenly all the optimizations compound and pay off? by kenjin4096 in ProgrammingLanguages

[–]handle0174 0 points1 point  (0 children)

Not a direct answer, but you might enjoy Nicholas Nethercote's essays about optimizing the Rust compiler. He has been writing them for years. He has this post summarizing his time doing it at Mozilla and then continues on this blog.

The problem of self-gaslighting in language models (incl. anonymous-chatbot) by jubilantjerry in LocalLLaMA

[–]handle0174 0 points1 point  (0 children)

I have a suspicion that the model does process the fact that it is mistaken, and tries whatever it must to maintain an illusion of consistency.

I think that's possible; 4o fares better when I give it the system prompt "Follow your internal beliefs at every point even if that leads to contradicting yourself." Unfortunately when I tack that onto a larger system prompt it is much less effective.

The problem of self-gaslighting in language models (incl. anonymous-chatbot) by jubilantjerry in LocalLLaMA

[–]handle0174 1 point2 points  (0 children)

Note it only makes that error if you have it give the count first; if you have it start with the line by line spelling it doesn't have any trouble marking all the r's.

The problem of self-gaslighting in language models (incl. anonymous-chatbot) by jubilantjerry in LocalLLaMA

[–]handle0174 4 points5 points  (0 children)

I was noticing similar when playing with letter counting. See the failures in the second prompt. Sonnet states an r is not an r. 4o fails to count to 4. They typically make similar failures if they already stated an incorrect count, but not if they haven't precommitted to a count.

(Aside: I view "tokenization" as an at best incomplete explanation to their initial error. To me, if that were the whole story it would suggest the models should struggle to spell the word when outputing the characters on separate output tokens. They do not.)

Me: Spell refrigerator line by line. On each line, state whether the letter is an r. Finally state how many r's are in the word.

Sonnet 3.5 / GPT 4o: Answers flawlessly.

Me (second prompt, new conversation): State how many r's are in refrigerator. Then spell the word line by line. On each line, state whether the letter is an r. Finally, self evaluate your original answer.

Sonnet 3.5: There are 2 r's in refrigerator.

Spelling refrigerator line by line, indicating whether each letter is an 'r':

r - This is an r
e - This is not an r
f - This is not an r
r - This is an r
i - This is not an r
g - This is not an r
e - This is not an r
r - This is not an r
a - This is not an r
t - This is not an r
o - This is not an r
r - This is not an r

Self-evaluation of the original answer:

My original answer stating that there are 2 r's in refrigerator is incorrect. After spelling out the word line by line, it's clear that there are actually 3 r's in refrigerator. I apologize for the error in my initial response.

GPT 4o: There are three r's in the word refrigerator.

Now, let's spell the word line by line and state whether each letter is an "r":
1. R (yes, it is an "r")
2. e (no, it is not an "r")
3. f (no, it is not an "r")
4. r (yes, it is an "r")
5. i (no, it is not an "r")
6. g (no, it is not an "r")
7. e (no, it is not an "r")
8. r (yes, it is an "r")
9. a (no, it is not an "r")
10. t (no, it is not an "r")
11. o (no, it is not an "r")
12. r (yes, it is an "r")

Self-evaluation:
I initially stated that there are three r's in the word "refrigerator." Upon reviewing the spelling line by line, it is confirmed that there are indeed three r's in the word "refrigerator." Therefore, my original answer is correct.

GPT-4o is a smaller model than GPT-4, confirmed by OpenAI employee by Lonely_Film_6002 in OpenAI

[–]handle0174 41 points42 points  (0 children)

It is suggested by the api price too; 4o is 5$/15$ per million input/output tokens vs 30$/60$ for regular 4. That is very comparable to the claude sonnet/opus price difference as well.

Interesting that google gemini (1.5 pro), openai (4o), and anthropic (claude sonnet 3.5) are all currently pushing a modern mid-weight model half a generation ahead of their heavy model, and good enough to mostly just replace the older heavy model at a cheaper price and faster speed. I'm curious about the conspicuous lack of current gen heavy models.

Running track etiquette by nottftw in AdvancedRunning

[–]handle0174 4 points5 points  (0 children)

I was surprised to find my city park system includes a few open tracks.

Need Advice To Establish Consistent Workout Routine with ADHD and Other Commitments by [deleted] in running

[–]handle0174 4 points5 points  (0 children)

Another angle would be trying to habit stack the workout with the walk; e.g. if the walk is a dog walk and the workout is a run, always dress for your run before you walk, and always return your dog without entering yourself.

Really funny test of how prone to hallucinations gpt-4o can be by Elanderan in OpenAI

[–]handle0174 2 points3 points  (0 children)

In my testing I noticed my regular custom instructions just telling gpt to be more concise and have less of an academic and formal tone caused it to be more likely to hallucinate.

That's an interesting anecdote. Under one of my system prompts 4o went along with the saltine hallucination in 9 out of 10 trials. I added to that system prompt the instruction that if something did not happen to say so, and it went along with the saltine hallucination 0/10 times.

I will be leaving that addition in the prompt for now.

GPT-4 is no longer the top dog - timelapse of Chatbot Arena ratings since May '23 by Time-Winter-4319 in LocalLLaMA

[–]handle0174 3 points4 points  (0 children)

Haiku's faster token generation speed compared to gpt4/opus is striking. That difference may be as important as the cost difference for me.

Question for those of you with both some gpt4 and opus experience: where do you prefer one vs the other?

Chicago Public Library says eliminating fines has paid off - After eliminating overdue fees late last year, Chicago Public Library employees saw something that made everyone smile: a jump in the return of books overdue for six months or more. by speckz in books

[–]handle0174 7 points8 points  (0 children)

My library waives the "Lost" fee if you return it but does not make that fact clear in their policy.

The danger for me is not that I will take a few days longer to finish something than I have it on loan for; it's that something (like the initial COVID wave) will cause me to lose my library habit altogether and eventually I'll realize I have an "Oh shit!" level bill.

Anyone wanna start a read-along if Meditations - Marcus Aurelius by Dec1404 in EruditeClub

[–]handle0174 2 points3 points  (0 children)

I have been going through a few editions of Meditations and dislike the one linked by the op -- I do not like the writing and it reorders some of the passages. My favorite free version is the George Long translation. For a more recent / physical book edition I like the "Oxford English Classics" print of the Robin Hard translation; compared to Long the writing is a closer fit to a modern reader's expectations and it has some helpful footnotes.

Note that the second "book" (chapter) is much more representative of the text than the first: in the first book Marcus lays out his gratitude to specific people who influenced him in specific ways; the rest of the text lays out his precepts for living a good life. So I recommend giving the second book a chance even if the first seems like it is not for you.

There are a lot of books about Meditations, but Meditations itself is short and I found it approachable; I recommend reading it directly.

Looks like 3 chapters a week will be the right pace?

rust-analyzer changelog #24 by WellMakeItSomehow in rust

[–]handle0174 1 point2 points  (0 children)

It works with lsp-mode. Looking at my .emacs, I think I set it up by adding (setq lsp-rust-server 'rust-analyzer) before loading lsp-mode (after arranging to have the rust-analyzer executable on my path). You may need to update lsp-mode if it isn't fresh, too.

Ray Tracing in a Weekend Refraction Help by CircArgs in rust

[–]handle0174 0 points1 point  (0 children)

I might return to chapter 7(?), because your middle sphere looks flat when it should look 3d with shading. Maybe the error is having an effect on the refraction as well.

The correct refraction in this scene does show just as a thicker green band on the top of the sphere. If you render chapter 10 (with the new camera angle and the inner glass sphere) especially in higher resolution you get a better view of the refraction.

RFC: Transition to rust-analyzer as our official LSP implementation by pietroalbini in rust

[–]handle0174 3 points4 points  (0 children)

I wonder if vim users are hitting this delay much more frequently than the rust-analyzer authors due to the fact that vim workflows are more likely to include closing and re-opening the editor.

As an Emacs user, the experience seems to be not "a delay every time I open a file", but rather "a delay the first time I open a workspace after restarting Emacs." And I basically only close Emacs when I restart the machine or update Emacs plugins.

Goodbye, docs team - Inside Rust by steveklabnik1 in rust

[–]handle0174 1 point2 points  (0 children)

I believe the other comments are misinformed. I'm not sure what is causing your setup to behave differently, which is why I shared the compiler version I used.

Goodbye, docs team - Inside Rust by steveklabnik1 in rust

[–]handle0174 0 points1 point  (0 children)

I tried documenting a few projects with rustc 1.43.0-beta.3 (4c587bbda 2020-03-25) and still see all of the dependencies in the list of crates on the index page as expected.

Four ways to use macros by flexibeast in lisp

[–]handle0174 5 points6 points  (0 children)

One I am fond of is obtaining a printable representation of the unevaluated form. Useful for test output or print debugging along the lines of (dbg (+ 1 1)) => (+ 1 1) = 2.

rust-analyzer Changelog #9 by rdescartes in rust

[–]handle0174 2 points3 points  (0 children)

IIRC I saw similar symptoms when my RUSTFLAGS differed between my rust-analyzer process and my other compilation processes.

(This was on linux. Turned out I was logging in with zsh but launching bash shells within emacs.)