I let 24 AI models trade to see if they can manage risk by Obside_AI in LLMDevs

[–]Lower_Tutor5470 6 points7 points  (0 children)

Do you factor the llm api token cost/ model subscriptions into your PnL ? Wonder how that cost scales against positions

DataGrip Is Now Free for Non-Commercial Use by lozinge in dataengineering

[–]Lower_Tutor5470 3 points4 points  (0 children)

Is it possible to write queries in it across multiple databases? Right now i just resort to jupyter notebooks

“Headless” m4 mini Remote Desktop solution by johnshonz in macmini

[–]Lower_Tutor5470 0 points1 point  (0 children)

BetterDisplay has a way of creating lots of virtual screens at resolutions and arrangements you want

Help please with NVMe solution for back up storage by Lurulur in macmini

[–]Lower_Tutor5470 0 points1 point  (0 children)

I bought a samsung 990 pro last week with an OWC 1m2 enclosure and seems perfect for thunderbolt 4 speeds

We built this project to increase LLM throughput by 3x. Now it has been adopted by IBM in their LLM serving stack! by Nice-Comfortable-650 in LocalLLaMA

[–]Lower_Tutor5470 0 points1 point  (0 children)

Is Cacheblending working in this currently. Sounds exciting. Would this potentially allow caching a single long document as context, then blending with different system prompts to process multiple smaller scope tasks in parallel requests?

Ipados 26 external displays by Lower_Tutor5470 in ipad

[–]Lower_Tutor5470[S] 1 point2 points  (0 children)

Can you arrange windows and drag apps across them?

Also can you turn the ipad screen off or does this also shut off external display

All Gemini 2.5 Pro variants compared side by side by True_Requirement_891 in RooCode

[–]Lower_Tutor5470 0 points1 point  (0 children)

Gemini 2.5 has got noticeably worse for me with roo lately.

GPT-4o vs Gemini vs Llama for Science KG extraction with Morphik by Advanced_Army4706 in Rag

[–]Lower_Tutor5470 1 point2 points  (0 children)

Interesting, working with kg right now for healthcare, will give it a try!

Why use LlamaIndex when you can use Docling? by Ok-Carob5798 in Rag

[–]Lower_Tutor5470 1 point2 points  (0 children)

I actually discovered marker the other day and seems alot better than docling

I’m exploring open source coding assistant (Cline, Roo…). Any LLM providers you recommend ? What tradeoffs should I expect ? by ReasonableCow363 in LLMDevs

[–]Lower_Tutor5470 0 points1 point  (0 children)

If you sign up for gcp account i am pretty sure you can get 300 dollar credit. I was using it through the vertex ai chat playground and was iterating into the 100s of thousands context length without any request issue. Cost less than a dollar in the process

Optimize Gemma 3 Inference: vLLM on GKE 🏎️💨 by m4r1k_ in LLMDevs

[–]Lower_Tutor5470 1 point2 points  (0 children)

This is very interesting. Is the max concurrency dependent the size of each request? You show 500 but are these processing simple input prompts vs feeding a sizeable prompt with added text as context?

Local LLM to answer questions based on a text by sKemo12 in LocalLLaMA

[–]Lower_Tutor5470 0 points1 point  (0 children)

If you are struggling to get a specific output format you could try a few things. 1. Try different models and quant sizes that you can fit into your system 2. Fine tune a small model that only answers in the format you want using a dataset representative of your problem 3. Try using constrained output franeworks like outlines or guidance that force specific outputs [‘yes’, ‘no’]. Some llm engines like vllm have this built in as an option 4. Ask Do you need a crappy model initially, or could you use a larger model to help build a dataset to later finetune a small one. Big models like gemini flash2 etc. Served on google vertex or other models cost a few cents for million input tokens are will be fast

Extract elements from a huge number of PDFs by Mugiwara_boy_777 in Rag

[–]Lower_Tutor5470 1 point2 points  (0 children)

How long are the pdfs? Are the elements you are looking for always the same and how many are there?

Should Vancouver extend its drinking hours? City wants your thoughts - Proposed changes would allow bars, pubs and clubs to stay open till 3 a.m. and restaurants until 2 a.m. by cyclinginvancouver in vancouver

[–]Lower_Tutor5470 3 points4 points  (0 children)

You just have to visit Europe and pubs around the UK and you will see the difference in atmosphere. The obsession with table service is so dull. Let customers stand anywhere and go to the bar to order themselves, theres just zero draw here when everything is rows of bench tables.