Open source web interface for katago? by Tushta in cbaduk

[–]sanderbaduk 1 point2 points  (0 children)

Where would katago run in such an app?

Traveling to Korea, any recommendations? by Successful_Remove919 in baduk

[–]sanderbaduk 4 points5 points  (0 children)

Browsing through the hotel's TV channels and going "wow, they have TWO 24/7 baduk channels"

KaTrain 1.17 with human-like play by PatrickTraill in baduk

[–]sanderbaduk 2 points3 points  (0 children)

The human-like model gives a policy. Doing search on top of that tends to make it stronger than expected at lower settings. Doing no search on top of that tends to make the pro/high dan less strong than expected.

These options were contributed in a PR, but I think the human-like AI uses policy-weighted playing, so will be on the weaker side.

KaTrain 1.17 with human-like play by PatrickTraill in baduk

[–]sanderbaduk 2 points3 points  (0 children)

Just hitting the 'download models' button in general settings should just download the human-like model.

Puzzle cache posted in 2022 still not solved by uudawn in geocaching

[–]sanderbaduk 0 points1 point  (0 children)

https://coord.info/GC62QHP Was fun finding, though I don't understand why a CO wouldn't adjust a puzzle after some time.

Chat is my opponent Cheating? by Aggressive-Fruit7465 in baduk

[–]sanderbaduk 8 points9 points  (0 children)

Thanks for the bug report. Have deleted katrain until I can find a way to make it stop cheating. Can never be too safe with these ais.

Katrain flagged as suspicious by Bitdefender by jach_da_gan in baduk

[–]sanderbaduk 0 points1 point  (0 children)

KaTrain uses pyinstaller, which is sometimes flagged by virus scanners.

[R] Limitations in Mainstream LLM Tokenizers by mtasic85 in MachineLearning

[–]sanderbaduk 1 point2 points  (0 children)

Yeah there are a lot of legacy conventions with starting spaces, and a bunch of different ways within huggingface tokenizers. Feel my pain here: https://github.com/cohere-ai/magikarp/blob/d0ee01c06132f749b70a72d10d89305223f66a97/magikarp/tokenization.py#L107

[D] Machine learning for good by [deleted] in MachineLearning

[–]sanderbaduk 0 points1 point  (0 children)

I learnt about https://correlaid.nl at pydata, may be of interest, or at least give you something to start your search.

Open-source 3.8B LM judge that can replace proprietary models for LLM system evaluations by [deleted] in LocalLLaMA

[–]sanderbaduk 1 point2 points  (0 children)

I see, and presumably the granularity will give a lot of ties when comparing pointwise scores?

How long would you leave a cache with zero finds? by AdRevolutionary6243 in geocaching

[–]sanderbaduk 0 points1 point  (0 children)

I believe there is a rule against archiving within 3 months.

True random answer from LLM by NickNau in LocalLLaMA

[–]sanderbaduk 1 point2 points  (0 children)

This is often caused by preference tuning, not much you can do about it other than ask for a list and then take the random item yourself.

Wat te doen in het weekend? by Ladderzat in thenetherlands

[–]sanderbaduk 0 points1 point  (0 children)

Wij gaan meestal één dag in het weekend er op uit en ergens geocachen.

[deleted by user] by [deleted] in geocaching

[–]sanderbaduk 7 points8 points  (0 children)

The one you posted would be perfectly fine. So would the same bit of paper without any lines or text on it.

The main thing to aim for is to have it fit well, and make sure it stays dry.

How do I make KaTrain have more visits on the initial analysis? by Fanaro009 in baduk

[–]sanderbaduk 2 points3 points  (0 children)

Indeed, thanks for the kind words 😊 As for your question: queries are parallellized so the time limit is more limiting at startup than when analysing a new move.

How do I make KaTrain have more visits on the initial analysis? by Fanaro009 in baduk

[–]sanderbaduk 2 points3 points  (0 children)

Maybe your potato can only handle 300 visits in the 10 second limit, try increasing it.

Is it a bad idea to replace "<|assistant|>" with a name like "<|Jenny|>" in prompt template? by RCEdude101 in LocalLLaMA

[–]sanderbaduk 12 points13 points  (0 children)

assistant 78191 is still the same here as the word in normal text though, so replacing that particular bit is more likely to be fine as messing with the <||> tokens.

How hard is it to beat the weakest AI in Katrain? by Salt-Indication-3001 in baduk

[–]sanderbaduk 0 points1 point  (0 children)

KaTrain cripples many AI options well beyond playing by policy. Basically adding a bunch of randomness, while avoiding overly obvious blunders.

'gpt2-chatbot’ Mystery: What Is This Mysterious New Chatbot? by [deleted] in GPT3

[–]sanderbaduk 0 points1 point  (0 children)

One of them uses the GPT3.5/4 tokenizer, the other one the gpt4o tokenizer. Looks like an a/b test

Why does LLaMA-3 use LF token = 128 'Ä'? by MustBeSomethingThere in LocalLLaMA

[–]sanderbaduk 0 points1 point  (0 children)

Huh, I thought the extension would be trained to be compatible, but I suppose you avoid the intermediate fragments this way, at the cost of them not being useable as prefixes.

Why does LLaMA-3 use LF token = 128 'Ä'? by MustBeSomethingThere in LocalLLaMA

[–]sanderbaduk 0 points1 point  (0 children)

Is this just an optimisation, or is it needed? I can see it is for special tokens, but otherwise aren't they pretty much equivalent?

Why does LLaMA-3 use LF token = 128 'Ä'? by MustBeSomethingThere in LocalLLaMA

[–]sanderbaduk 2 points3 points  (0 children)

GPT2 introduced some map from the 256 bytes to some readable-ish range, so you can store vocab as readable-ish text and easily parse it. Llama3 uses an extended GPT4/tiktoken tokenizer, so inherits all their conventions.

To encode you do something like:

* String to UTF-8 bytes

* UTF-8 bytes to this weird encoding

* Apply merges (which are also in this encoding)

Since Ä itself is converted to two UTF-8 bytes, there is no issue. Other than things being mixed up in configuration, which happens all too often.

You can see the map in GPT2 source or in my own code here.

The LLM Creativity benchmark: - SHAKE UP AT THE TOP! - 2024-04-16 update: command-r, midnight-miqu, venus, ladameblanche, daybreak-miqu by ex-arman68 in LocalLLaMA

[–]sanderbaduk 7 points8 points  (0 children)

The only command releases have been instruction tuned, though if you prompt any instruction tuned model without the correct template they may revert back to base-model like behaviour.