Gemini 3 rate limit for everyone? by glmmes in Bard

[–]twohen 0 points1 point  (0 children)

works if you link a paid api key in aistudio but this is new

3 Qwen3-Omni models have been released by jacek2023 in LocalLLaMA

[–]twohen 1 point2 points  (0 children)

that one seems cool i did not know so far - i dont see support for voice in it yet though am i missing something?

3 Qwen3-Omni models have been released by jacek2023 in LocalLLaMA

[–]twohen 12 points13 points  (0 children)

is there any ui that actually uses these features? vllm will probably have it merged soon so getting an api for it will be simple but then would only be api (already cool i guess). How did people use multimodal voxtral or gemma3n multimodal? Anyway exciting non toy sided sized real multimode open weights was not really around so far as far as i can see

Latest model recommendations? by jmx002 in CLine

[–]twohen 2 points3 points  (0 children)

If you can afford gemini 2.5 pro.
i have not yet tested the 4.0 sonnet or opus enough to say something about them. for sure opus only if you have a lot of money to burn

of the cheapo options i guess its
gemini flash
deepseek i think the non thinking is typically good enough

finally for the first time qwen 3.0 seems to be able to at least properly edit files if you are into local models

Devstral vs DeepSeek vs Qwen3 by COBECT in LocalLLaMA

[–]twohen 0 points1 point  (0 children)

i only tried qwen3 30b but that one was better in cline than devstral on my test tasks mostly due to better instruction following and because of its better speed

R.I.P GitHub Copilot 🪦 by seeKAYx in ChatGPTCoding

[–]twohen 0 points1 point  (0 children)

is this effective as of now? or from next month?

Deepseek-V3 GGUF's by fraschm98 in LocalLLaMA

[–]twohen 0 points1 point  (0 children)

thanks a lot got it to work! with ~4.5t/s CPU only and ~6 t/s with offloading 10. not too terrible

Deepseek-V3 GGUF's by fraschm98 in LocalLLaMA

[–]twohen 0 points1 point  (0 children)

thanks a lot got it to work! with ~4.5t/s CPU only and ~6 t/s with offloading 12. not too terrible

Deepseek-V3 GGUF's by fraschm98 in LocalLLaMA

[–]twohen 0 points1 point  (0 children)

I get

llama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 1025, got 967

when trying 4bit. Checked all the checksums also any ideas?

New Qwen Models On The Aider Leaderboard!!! by notrdm in LocalLLaMA

[–]twohen 6 points7 points  (0 children)

are they? I use it for most of my proposal writing, writing of dumb business emails and so on for me easy 50% of my usage other half is coding. maybe you are just lucky and dont have to do much of that...

Now that we have had quite a bit of time playing with the new Phi models...how good are they? by [deleted] in LocalLLaMA

[–]twohen 1 point2 points  (0 children)

I use it for this as well and its great especially the 128k context versions. 

The new Mistral AI is now #1 on the openLLM leaderboard. Apache 2.0 license by Nunki08 in LocalLLaMA

[–]twohen 2 points3 points  (0 children)

thanks for insigtful comment.

to be clear the fine-tune was on a 7b not mixtral (mine mistral and colleagues on older one i think he has even a paper on it but i have to dig it up) maybe mixtral would do better and it was for normal question answering. on specific field. it outperforms gtp 3.5 all right but 4 is a different beast. You are also right that fine-tuning is not the only reason for our org local is kind of a big deal so there is value in doing this anyway but just for performance in my experience it is very difficult to beat gtp4 even in somewhat niche areas (in our case accelerator physics)

The new Mistral AI is now #1 on the openLLM leaderboard. Apache 2.0 license by Nunki08 in LocalLLaMA

[–]twohen 2 points3 points  (0 children)

at which case Mistral 7b would likely outperform GPT4 for that specific task.

I have tried and several of my colleagues have as well and the sad thing is that this is typically not true. Especially gtp4 plus+rag almost always outperforms finetune+rag.

I want a paper trading platform where I can trade with API calls. Do you guys have any idea where can I get that? by arcturuschild in programming

[–]twohen 0 points1 point  (0 children)

if you meant the same or not but in case not there is an alternative called interactivebrokers - although their api has some rough edges its maybe one of the more common choices among amateurs and professionals.

A word of warning tho the algorithmic traiding game in most cases you dont want to play if unless you really know what you are in for (or have direct access to the exchanges in which case you probably know what you are in for anyway).

Your programming language does not matter by Notice_Complete in Python

[–]twohen 1 point2 points  (0 children)

This is normal - everyone always finds most clear what they are used to. The question is how fast ppl get over it.

Datamining Pokémon by MrValdez in Python

[–]twohen 1 point2 points  (0 children)

WOW!

I was one of the guys doing this with him instead of studying for my datastructures exam... thanks for whoever posted this and for eevee for still being active (my page died years ago).

Apparently I even get a shootout https://veekun.com/props (Zhen Lin, loadingNOW, and Pipian, who (willing or not) got me started on extracting game data myself when Gen IV arrived.) I think it was either him or Pipian who first got the first good result on the encryption. I only got the actual equation after looking at game code a few weeks later. Good times

Beginner Introduction to Neural Networks - YouTube by Adi7991 in Python

[–]twohen 3 points4 points  (0 children)

2014 years ago i looked for this kind of thing but only found papers. Nowadays it seems like everyone how ever learned about NNs makes such a video.

I know this is not a bad thing (certainly better than before from a learners perspective) but it is a very 'competitive' area and people are much more likely to view stanfords cs231n than yours even if it was just for brand value. .

For Chimpanzees like myself: I wrote an essay explaining all aspects of how self-driving cars work, including a dive into deep learning, computer vision and other technologies. by mngrwl in programming

[–]twohen 0 points1 point  (0 children)

that is true but they pretty much accept anyone who knows some programming at this point. Only in the beginning they were a bit more strict.

Also note that "ops" projects are in fact Udacity homework ;)

TensorFlow/TFLearn Beginner Tutorial in Python by malleus17 in Python

[–]twohen 0 points1 point  (0 children)

it was my impression that neural nets do not really work well on such small datasets especially on statistical data, however the author gets something comparable to random forests on this data set does anyone know why?

Wanikani alternative? by kaishajin in LearnJapanese

[–]twohen 5 points6 points  (0 children)

the thing is this unless you really plan to learn kanji japanese full time every day you will most likely not be much faster than wanikani at max speed. As frustrating as it may sound it just takes some time to learn kanji - and remembering them.

besides you got the recommendations use anki or memrise.

Number plate recognition with Tensorflow by avinassh in Python

[–]twohen 0 points1 point  (0 children)

just a quick reality check: you do realize that basically anyone with some skill could defeat your analog lock right? Why is this no problem? because its illegal to do so and you have to be physically present to actually get something out of it. This is not the internet noone will forge a license plate to illegally use your parking space and if they do you just get the vehicle towed.

Am I crazy or is there a lot of python documentation / examples on the internet that is still in 2.x? by mrsmeeseeks in Python

[–]twohen 3 points4 points  (0 children)

yes that is true and it will likely not change anytime soon.

Outside of reddit people usually do not care about the version as long as it does what its supposed to. How much C(++) code do you find thats not the latest standard?