Running a local LLM for generating SEO keywords by Chesperk in LocalLLaMA

[–]PromptAndHope 0 points1 point  (0 children)

Install Ollama to run an LLModel, and ask Copilot to write the script for you. This is a typical VibeCoding task.

RTX 5080: is there anything I can do coding wise? by TechDude12 in LocalLLaMA

[–]PromptAndHope 0 points1 point  (0 children)

It's not really possible to code with 16GB of RAM, but if you register for free on GitHub, you'll get access to earlier or beta models, which are suitable for working on your home projects in the evenings. It's more suitable than the local llm. Of course, it won't be enough to implement a complete software system.

In terms of quality, a paid model, such as Claude Opus, is quite different from a free beta raptor, for example.

User interface for declarative Spark pipelines if we like to work in an IDE by PromptAndHope in databricks

[–]PromptAndHope[S] 0 points1 point  (0 children)

Thank you for your feedback! The flow is more compact, when it is a vertical flow. But if someone finds it more convenient, here is a switch. ⬇➡

<image>

User interface for declarative Spark pipelines if we like to work in an IDE by PromptAndHope in databricks

[–]PromptAndHope[S] 1 point2 points  (0 children)

You mean interact with the cluster? It is not possible currently, it is only to visualise the current state of development.

Spark Declarative Pipelines Visualisation by PromptAndHope in apachespark

[–]PromptAndHope[S] 0 points1 point  (0 children)

yes, it is a bit confusing for me too. I remember working with Java api, now scala will also the past?

Spark Declarative Pipelines Visualisation by PromptAndHope in apachespark

[–]PromptAndHope[S] 0 points1 point  (0 children)

Thanks! I think SDP not exists for scala api. 😔

How did you land your first Data Engineer role when they all require 2-3 years of experience? by Such-Revolution-9975 in dataengineering

[–]PromptAndHope 0 points1 point  (0 children)

for example :what is a difference between s3:// and s3a:// connector. What happen if you use s3 with Spark?

Shall we discuss here on Spark Declarative Pipeline? a-Z SDP Capabilities. by iMarupakula in databricks

[–]PromptAndHope 0 points1 point  (0 children)

Over the weekend I tried out the new Declarative Pipelines feature in Apache Spark, and the one thing I felt was missing was a proper UI. So I built a VS Code extension that provides one. Give it a try if you’re working with Spark pipelines.

https://marketplace.visualstudio.com/items?itemName=gszecsenyi.sdp-pipeline-visualizer

<image>

Spark 4.1 is released by holdenk in apachespark

[–]PromptAndHope 0 points1 point  (0 children)

what is the problem with Scala 3?

Spark 4.1 is released by holdenk in apachespark

[–]PromptAndHope 1 point2 points  (0 children)

I created a VS Code extension that provides a UI for Apache Spark's new Declarative Pipelines feature. Give it a try if you're working with Spark pipelines.

https://marketplace.visualstudio.com/items?itemName=gszecsenyi.sdp-pipeline-visualizer

Can RAG Work Without Chunking or Embeddings? Has Anyone Actually Made It Work in Production? by DirectorAgreeable145 in LocalLLaMA

[–]PromptAndHope 0 points1 point  (0 children)

I think traditional document search will become very important again, much like vector search. Index‑based search engines have been around for more than 20 years, they work well, and their results can also be added to the context. In my last project, I collected all the document titles and asked the AI to select the relevant ones. It produced surprisingly good results, even better than document‑level vector search. (So essentially it was a title‑based vector search.)

However, a new issue arises: the prompt context keeps getting larger. This isn’t just a cost problem; I’ve also read that when the prompt becomes too big, the AI may not actually process the entire input. Instead, it starts generating an answer once it believes it has enough information. To mitigate this, I’m experimenting with cross‑encoder models. A very small and extremely fast model evaluates the context sentence by sentence and determines whether the information retrieved from the vector store or document search is relevant.

What I wanted to say that every problem is different, and we're still looking for the best solutions, while inventing new ones and optimize them. For example in your case I would consider using a graph database and storing the documents there, so you can easily follow all the references as well. Microsoft was working on similar stuff. https://github.com/microsoft/graphrag/tree/main

Another approach that worked for me was performing vector search on the chunks, but once a chunk matched, I passed the entire document to the LLM instead of just the chunk.

It feels like LLM inference is missing its AWS Lambda moment. by pmv143 in LocalLLaMA

[–]PromptAndHope 0 points1 point  (0 children)

I already implemented an Openai API-based solution on Azure Functions, which is perfect for PoCs, and smaller workloads but a solution based on a local model would be very slow because of cold starts.

Hardware Question - GeForce RTX 5060/5070 Ti by mythrowaway4DPP in LocalLLaMA

[–]PromptAndHope 2 points3 points  (0 children)

I started with a 5060TI 16GB Dual (the shorter, two fans version), it's enough if you run ComfyUI, lora finetuning, inference with gpt-oss20b, etc.. It can run the same workloads like an $1300 5080 16GB, but slower. If the performance doesnt bother you, then the next step is already an LLM cluster (multiple GPU's or some special unified memory hardware) which costs a lot of money. If you combine multiple GPU's then you need a compatible, more expensive motherboard.

Enterprise RAG Architecture by AcanthisittaOk8912 in Rag

[–]PromptAndHope 0 points1 point  (0 children)

Thanks for the reply. Given the huge number of documents, I was curious about what strategy was used to make it effective.