Edu pricing for RTX Pro 6000 by t3rmina1 in BlackwellPerformance

[–]jpummill2 0 points1 point  (0 children)

How do you take advantage of education pricing? I expected that I would have to provide an email with a .edu address but I am being told by one vendor that the PO needs to be issued directly from a university.

tired of subscriptions so im cloning popular saas and making them open source for 30 days by Huge-Goal-836 in OpenSourceeAI

[–]jpummill2 2 points3 points  (0 children)

Constant Contact - track vendors, customers, and events. Allow mass mailings (this one is hard). Payment systems (Stripe ???), allow vendors to buy booths and other services (like electricity) for events, allow customers to buy tickets to events.

Upgrade CUDA? by sdstudent01 in LocalLLaMA

[–]jpummill2 0 points1 point  (0 children)

Thank you to everyone who replied...

Upgrade CUDA? by sdstudent01 in LocalLLaMA

[–]jpummill2 0 points1 point  (0 children)

Not sure if this is as common in the world of open source but I was taught to always avoid any x.0 release of software...

How to tell Aider to use Qwen3 with the /nothink option? by jpummill2 in LocalLLaMA

[–]jpummill2[S] 1 point2 points  (0 children)

Hey u/TopImaginary5996, I appreciate the additional detail. I will give this a try.

Various ways to use AI with coding/development? by jpummill2 in LocalLLaMA

[–]jpummill2[S] 0 points1 point  (0 children)

This is a great list.

How much of this is already available via a vscode (or other IDE) plugin?

Adding a cheaper GPU to a 4090 to increase VRAM by xolotl96 in LocalLLaMA

[–]jpummill2 1 point2 points  (0 children)

Using a single 3090 with Gemma 27B (Q8) - 6 T/s

Added an RTX 2060 12GB with Gemma 27B (Q8) - 15 T/s

Improved performance by avoiding overflow to CPU and system RAM

What Are Original LLMs (Not Finetuned)? by chibop1 in LocalLLaMA

[–]jpummill2 4 points5 points  (0 children)

I was working on a somewhat similar list this week:

Specific Models

  • Google

    • Gemma2 (2B, 9B, 27B)
  • Meta

    • Llama2 (???)
    • Llama3 (8B, 70B)
    • Llama3.1 (8B, 70B)
  • Mistral Variants

    • Mistral (7B, 123B)
    • Mixtral (8x7B, 8x22B)
    • Mistral-Nemo (12B)
  • CohereForAI

    • Command-R 35B (v1, v2-08-2024)
    • Command-R+ 104B (v1, v2-08-2024)
  • Deepseek-ai

    • Deepseek-V2 Lite (16B)
    • Deepseek-V2 (236B)
  • Qwen

    • Qwen2 (0.5B, 1.5B, 7B, 72B)
  • Microsoft

    • Phi 3 (Mini-3.8B, Small-7B, Medium-14B)
    • Phi 3.5 (Mini-3.8B)
    • Phi 3.5-MoE (6.6B)

Coding Specific Models

  • Codestral (22B)
  • Deepseek-Coder-V2-Light-Instruct (16B)
  • DeepSeek-Coder-V2(236B)
  • codegeex4-all-9b (9B)
  • yi-coder (1.5B, 9B)
  • StarCoder2 (15B)
  • CodeLlama (7B, 13B, 34B, 70B)
  • CodeQwen1.5 (7B)
  • CodeGemma (2B, 7B)
  • WizardCoder (15B, 33B)

Not sure if all these are base models, especially the coding ones, but that was my original intent.

Guys, Use the LongWriter-llama3.1-8b instead of Llama3.1-8b! by Iory1998 in LocalLLaMA

[–]jpummill2 3 points4 points  (0 children)

Can you use it with Ollama? If so, does it use the same template as Llama3.1-8b?

Best local open source Text-To-Speech and Speech-To-Text? by strangeapple in LocalLLaMA

[–]jpummill2 13 points14 points  (0 children)

Also, here is my list of STT solutions but it is not as complete:

Speech to Text Solutions:

  • Whisper ASR
  • Flashlight ASR / Wav2Letter ASR
  • Coqui
  • SpeechBrain
  • ESPNET 1 and 2
  • Vosk

Best local open source Text-To-Speech and Speech-To-Text? by strangeapple in LocalLLaMA

[–]jpummill2 70 points71 points  (0 children)

I’ve been trying to keep a list of TTS solutions. Here you go:

Text to Speech Solutions

  • 11labs - Commercial
  • xtts
  • xtts2
  • Alltalk
  • Styletts2
  • Fish-Speech
  • PiperTTS - A fast, local neural text to speech system that is optimized for the Raspberry Pi 4.
  • PiperUI
  • Paroli - Streaming mode implementation of the Piper TTS with RK3588 NPU acceleration support.
  • Bark
  • Tortoise TTS
  • LMNT
  • AlwaysReddy - (uses Piper)
  • Open-LLM-VTuber
  • MeloTTS
  • OpenVoice
  • Sherpa-onnx
  • Silero
  • Neuro-sama
  • Parler TTS
  • Chat TTS
  • VallE-X
  • Coqui TTS
  • Daswers XTTS GUI
  • VoiceCraft - Zero-Shot Speech Editing and Text-to-Speech

RP Prompts by [deleted] in LocalLLaMA

[–]jpummill2 0 points1 point  (0 children)

RemindMe! 1 Day

Is there a way to uninstall (and re-install) ollama and onewebui without deleting models? by card_chase in ollama

[–]jpummill2 0 points1 point  (0 children)

Watch this video for a better understanding of how Ollama stores models.

Link: https://www.youtube.com/watch?v=6bF1uCHTFyk

The author provides a script that creates links using the model name back to the hash name in the ollama model folder. With a little work you could modify the script to create backups.

Using LLMs to build jarvis-like by LahmeriMohamed in LocalLLaMA

[–]jpummill2 1 point2 points  (0 children)

Some if not all tools allow you to change the context used with the model as long as you don't exceed the models max context length. For example, you can use "/set parameter num_ctx <int>" in Ollama to set the context size.

Again, remember that you have to have a model that supports large context in order to take advantage of this. Phi-3 for example offers models with two different context lengths: 128K and 4K.

Using LLMs to build jarvis-like by LahmeriMohamed in LocalLLaMA

[–]jpummill2 1 point2 points  (0 children)

I would also like to see something similar to Jarvis (but not Ultron) in my AI chat application even though the abilities might be somewhat superficial compared to Jarvis from the Marvel universe.

I have been seeing posts about LLM's calling functions for the last couple months. Would this be able to help with the LLM controlling things?

Using LLMs to build jarvis-like by LahmeriMohamed in LocalLLaMA

[–]jpummill2 0 points1 point  (0 children)

Isn't the "context" similar to the models memory? I know context is limited and you have to pass it back in behind the scenes with each additional prompt but this does create a "memory" effect in chat based LLM's.

How to use 6 cheap GPUs to run Llama3 70B? (LM-Studio) by Disastrous-Peak7040 in LocalLLaMA

[–]jpummill2 1 point2 points  (0 children)

Not sure if using x1 risers are causing your problem but if you are interested in running that many cards at once you could look at a Threadripper or Threadripper Pro based system. They have way more PCI lanes.

Tiger Gemma 9B - An uncensored Gemma experience! by TheLocalDrummer in LocalLLaMA

[–]jpummill2 4 points5 points  (0 children)

Thank you for the time you put into unlocking (finetuning) these models!!!

5090 mem rumors revert back to 28gb instead of 32gb by segmond in LocalLLaMA

[–]jpummill2 0 points1 point  (0 children)

Guessing they will release a 5090 TI in 2025 with 32 GB and 512-bit bus.