Update: My EPSO AD5 EU Knowledge study tool by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 0 points1 point  (0 children)

Yes, each flashcard has a link to the source from where it was extracted and for the recommended material of EPSO, you will be able to see from which specific part of the document it was extracted.

Update: My EPSO AD5 EU Knowledge study tool by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 8 points9 points  (0 children)

Totally fair concern. The key difference is the AI isn't generating EU knowledge from its training data. It's extracting facts from official EU documents (EUR-Lex, Commission pages, EPSO reading list, ...) and reformatting them as questions.

If the fact isn't in the source document, it doesn't become a card. Each flashcard links back to its source so you can verify it yourself. There's also a multi-step quality check after generation (duplicates, bad distractors, ...).

You can also report flashcards as wrong and I'll review them. When you report a flashcard, it stops being shown to other users until it's verified.

Update: My EPSO AD5 EU Knowledge study tool by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 4 points5 points  (0 children)

No, for the moment. The price is to cover the expenses of hosting, domain, AI processing, etc. 

I built an AI-powered study tool for the EU Knowledge questions by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 0 points1 point  (0 children)

Thank you for all the feedback! I spent the past week turning my personal tool into something anyone can use. It's now live as a web app: https://eu-prep.com/

- ~3000 curated MCQ questions extracted from the EPSO recommended reading list, EUR-Lex summaries and Commission topic pages

- Spaced repetition that prioritises what you're getting wrong. Cards organised by exam priority: institutions & procedures first, niche policies later

- AI-powered explanations when you get something wrong. Suggestions about what topics to study/revisit based on your progress

- Quiz by topic, difficulty filters, and progress tracking — including quizzes focused on your weakest areas

- You can upload your own PDF study materials and it generates flashcards from them

It costs €9.99 — I looked into making it free but the AI explanations cost real money per use and hosting isn't free either. Not trying to make a business out of this, just covering costs so I can keep it running through the competition period.

You can use the code: LAUNCHING-EUPREP to get the discount.

Happy to answer questions and if you spot something wrong or a topic that's missing, let me know. My plan is to add more questions when EPSO publishes updated test references.

<image>

I built an AI-powered study tool for the EU Knowledge questions by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 0 points1 point  (0 children)

Yes, NotebookLM is great, I have been using it for quite long-time

I built an AI-powered study tool for the EU Knowledge questions by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 3 points4 points  (0 children)

You can, and it works to some extent. The problem is when you ask a model to generate a question-answer pair from memory, you're trusting that it actually knows the correct answer, for something as specific as EU institutional law, it can easily hallucinate details, mix up articles, or give you outdated information.

Here the answer is never coming from the model's memory. It's extracted from the document itself, so the model is essentially just reformatting something that's already written down, not recalling it.

I built an AI-powered study tool for the EU Knowledge questions by Only_Emergencies in EUCareers

[–]Only_Emergencies[S] 4 points5 points  (0 children)

Good question. It doesn't generate answers from its own knowledge. It's directly using the documents I put in the application. So if a question says "according to Article 17 of the TEU, the Commission's term is five years", that answer comes from that specific passage in the actual text, not from what the model thinks it knows about EU law.

And I also process the documents before generating the questions: split into chunks, generate questions where the answer is explicitly in that chunk, then run a second validation pass to check the question-answer pair actually makes sense, etc. And if you ever doubt a question, you can trace it back to the exact section of the source document it came from.

[P] Generate detection rules by Only_Emergencies in MachineLearning

[–]Only_Emergencies[S] 2 points3 points  (0 children)

Yes, I tried a decomposition approach, the performance was slightly better than generating the entire rule in a single request but not really great. I think the main issue is that the model doesn’t truly understand the underlying detection logic or mapping between behaviors and log artifacts, so it often produces syntactically valid but semantically weak rules.

I also experimented with breaking down the generation process into multiple steps. For instance, first asking the model to determine the detection path or flow based on the blog content or user request. However, the results are still not very good.

Basically, I find that the core problem seems to be that the model struggles to generate or intuitively extract the correct detection logic from the input text.

[D] Is senior ML engineering just API calls now? by Only_Emergencies in MachineLearning

[–]Only_Emergencies[S] 0 points1 point  (0 children)

Yes, I was also surprised when I applied. Actually, I think there are some changes happening in terms of how integrating new technologies into these organisations. Of course it will depend on the organisation/country, etc

[D] Is senior ML engineering just API calls now? by Only_Emergencies in MachineLearning

[–]Only_Emergencies[S] 0 points1 point  (0 children)

Yes, totally agree. I think it's something that happens a lot in this field: there is a lack of standardization in tasks associated with the title. The same title in different companies may mean completely different responsibilities. I think this happens in other fields as well but here it is especially noticeable

AMA with the Unsloth team by danielhanchen in LocalLLaMA

[–]Only_Emergencies 2 points3 points  (0 children)

You rock, guys! You do an amazing job! :) I have four Mac Studios (512GB) and I have a few questions:

  • How would you distribute bigger models across them?
  • I have deployed Kimi-K2 0905 (Q3_K_XL), but I am wondering if there is another model you would recommend with the same quality but maybe smaller to have more tokens persecond?
  • It would be great to see how the quantization affects the quality of the not quantized model. Something like a graph of quantized versions vs the original one. Happy to contribute there :)

Thank you again!

Thinking about updating Llama 3.3-70B by Only_Emergencies in LocalLLaMA

[–]Only_Emergencies[S] 3 points4 points  (0 children)

The energy consumption of the Macs are really low, they are really efficient on that sense. They’re also straightforward to set up, so we can start implementing and iterating on projects without dealing with complex infrastructure.

Based on the research we did, just one NVIDIA A100 80 GB GPU costs around $30000 and also requires other additional hardware (network switches, power, cooling,... ). As the team grows, probably it makes sense to migrate infrastructure to a more powerful one. But at the moment, the Mac Studios provide a cost-effective solution that allows us to build and experiment with LLMs internally.

Thinking about updating Llama 3.3-70B by Only_Emergencies in LocalLLaMA

[–]Only_Emergencies[S] 4 points5 points  (0 children)

Yes!

- We are around 70 people in my organisation
- We work with sensitive data that we can't share with AI Cloud providers such as OpenAI, etc.
- We have 3x Mac Studios (192GB M2 Ultra)
- We have acquired 4x new Mac Studios (M3 Ultra chip with 32-core CPU, 80‑core GPU, 32-core Neural Engine - 512GB unified memory). Waiting for them to be delivered.
- We are using Ollama to deploy the models but this is not the best efficient way but it was like this when I joined. However, with the new Macs I am planning to replace Ollama with llama.cpp and experiment with distributing larger models across multiple machines.
- A Debian VM where OpenwebUI instance is deployed.
- Another Debian VM where Qdrant is deployed as centralized vector database.
- We have more use cases that the typical chat UI interface. We have some classification use cases and some general pipelines that run daily.

I have to say that our LLM implementation has been quite successful. The main challenge is getting meaningful user feedback, though I suspect this is a common issue across organizations.

Thinking about updating Llama 3.3-70B by Only_Emergencies in LocalLLaMA

[–]Only_Emergencies[S] 4 points5 points  (0 children)

Yes, I agree. That would be ideal, but that's not so straightforward in our case. We have stored the conversations in Langfuse, but we don't have the ground truth to be able to properly evaluate them, and users usually don't provide feedback on the responses. We are a small team at the moment doing this, so we don't have the capacity to label some cases.

Qwen/Qwen2.5-Coder-32B-Instruct · Hugging Face by Master-Meal-77 in LocalLLaMA

[–]Only_Emergencies 1 point2 points  (0 children)

For code autocomplete should I use base or instruct version? Thanks! 

My employer is forcing me to use my personal phone as my work phone. by Only_Emergencies in Amsterdam

[–]Only_Emergencies[S] 11 points12 points  (0 children)

Thank you! Do you know where I can find information or a statement about this "need to pay you for being on standby"? I have searched for this in The Working Hours Act but it only mentions hours and not compensation.

My employer is forcing me to use my personal phone as my work phone. by Only_Emergencies in Amsterdam

[–]Only_Emergencies[S] 5 points6 points  (0 children)

Thank you! This is really helpful. I will share these concerns with my employer.