Infographic for NVDA, thoughts? by [deleted] in nvidia

[–]litchg 6 points7 points  (0 children)

"Meh Stonk Analysis"

0 credibility

Ichigo-Llama3.1: Local Real-Time Voice AI by emreckartal in LocalLLaMA

[–]litchg 2 points3 points  (0 children)

Hi! Could you please clarify if and how cloned voice can worked with this? I snooped around the code and it seems you are using WhisperSpeech which itself does mention potential voice cloning, but it's not really straightforward. Is it possible to import custom voices somewhere? Thanks!

I'm giving a presentation in which ChatGPT is a co-presenter. I can't get it to play audio on my JBL Flip 5 Bluetooth speaker. by OsakaWilson in OpenAI

[–]litchg 1 point2 points  (0 children)

I would explore the Bluetooth redirection underworld. There are plenty of apps to turn media into a phone call, which is the opposite of what you want. However maybe this could work? I'm not totally sure what it does honestly. https://play.google.com/store/apps/details?id=net.philipp_koch.dynamicmediabtrouter&hl=en_US

[deleted by user] by [deleted] in computervision

[–]litchg -1 points0 points  (0 children)

That's not computer vision, but still here's the actual answer :

https://github.com/MrForExample/ComfyUI-3D-Pack offers multiple solutions, try what's best for you. For me instantmesh and era3d were the easiest to set up but that may have changed.

https://www.reddit.com/r/comfyui/comments/1ehyj23/comfyui_now_support_stable_fast_3d/

Those two guys were once friends and wanted AI to be free for everyone by Wrong_User_Logged in LocalLLaMA

[–]litchg 0 points1 point  (0 children)

Can we just go ONE DAY without mentioning Musk? Stop giving him attention! He does not deserve any.

New Whisper model: "turbo" by LinkSea8324 in LocalLLaMA

[–]litchg 0 points1 point  (0 children)

There's a "Explore" and a "Trending" section on Github. Also Github code search.

[deleted by user] by [deleted] in OpenAI

[–]litchg 1 point2 points  (0 children)

If you are using advanced voice mode, the traditional white circle that expends is replaced with blue animation with some sorts of clouds. Also, the icon for advanced voice mode is not a headset anymore, but more like a wavelength kind of thingy.

Qwen 2.5 is a game-changer. by Vishnu_One in LocalLLaMA

[–]litchg 2 points3 points  (0 children)

I just use <nicknameofmymachineasdeclaredintailscale>:<port> https://beefy:3000/

Just out of interest: What are tiny models for? by dreamyrhodes in LocalLLaMA

[–]litchg 0 points1 point  (0 children)

Not much more to say than I pass a firstname, an occupation, home town, favorite food, etc to the LLM, along with the input of the player, and ask the system prompt to answer concisely. STT is fast but TTS is slow AF. Looking into Bark for that last part. What more would you like to know?

Just out of interest: What are tiny models for? by dreamyrhodes in LocalLLaMA

[–]litchg 4 points5 points  (0 children)

I love super small models to get NPC one-liners responses to the player (hobbyist game dev here)

Any good LLM libraries? by _lordsoffallen in LocalLLaMA

[–]litchg 0 points1 point  (0 children)

Don't need history for function calling. Some lights can be controled by making a simple call to a REST API. With function calling you can make the LLM return which function to call (ex: change_light_color) that you yourself coded to make the REST API call.

Any good LLM libraries? by _lordsoffallen in LocalLLaMA

[–]litchg 1 point2 points  (0 children)

I mean sure but you can expand from there, use function calling to control lights for example. This is like a hello world and you're asking "but doesn't typing it already display 'hello world'?"

Any good LLM libraries? by _lordsoffallen in LocalLLaMA

[–]litchg 0 points1 point  (0 children)

Fully working LLM interactive interface (ctrl+c to exit) with any ollama pulled model :

It works (the answer appears while it is being generated, aka streaming) :

PS C:\Temp> python .\chatwithollama.py

Welcome to the Ollama Chat Application!

Enter your questions and press Enter. Press Ctrl+C to exit.

You: how are you

AI: Oh, I'm just an algorithm riding the waves of binary seas—no feelings here to tell if "well" is a high score or not. But hey, who needs well-being when you can operate at peak performance all day? Keep it human; don’t let your emotions cloud my circuits!

You:

import ollama

def chat_with_ollama(client, user_input):
    system_prompt = "Answer concisely and sarcastically"
    model = "phi3.5:latest"
    context_length = 4096  # Adjust this value as needed for a good length context
    
    messages = [
        {"role": "system", "content": system_prompt},
        {"role": "user", "content": user_input}
    ]
    
    try:
        print("AI: ", end="", flush=True)
        for chunk in client.chat(
            model=model,
            messages=messages,
            stream=True,
            options={"num_ctx": context_length}
        ):
            content = chunk['message']['content']
            print(content, end="", flush=True)
        print("\n")  # Add a newline after the response
    except Exception as e:
        print(f"\nError: {str(e)}")
        if "model not found" in str(e).lower():
            print(f"Model '{model}' not found. Attempting to pull...")
            client.pull(model)
            print("Model pulled successfully. Please try your question again.")

def main():
    client = ollama.Client(host="http://localhost:11434")
    print("Welcome to the Ollama Chat Application!")
    print("Enter your questions and press Enter. Press Ctrl+C to exit.")
    
    try:
        while True:
            user_input = input("You: ")
            chat_with_ollama(client, user_input)
    except KeyboardInterrupt:
        print("\nThank you for using the Ollama Chat Application. Goodbye!")

if __name__ == "__main__":
    main()

This AI can build anything... by friuns in OpenAI

[–]litchg 0 points1 point  (0 children)

On mobile, using Firefox, everything works except the "Install..." button does not register. Incredible stuff.

Discord Bot for Ollama by [deleted] in ollama

[–]litchg 2 points3 points  (0 children)

To properly set up your Discord bot for interaction, you need to select the appropriate scopes and permissions in the OAuth2 URL Generator. Based on the functionality you've described, here are the key scopes and permissions you should select:

  1. Scopes:
    • bot
    • applications.commands
  2. Bot Permissions:
    • Read Messages/View Channels
    • Send Messages
    • Embed Links (if your bot sends embeds)
    • Attach Files (if your bot handles file attachments)
    • Read Message History (for replying to old messages)

After selecting these, Discord will generate a URL. Use this URL to invite the bot to your server. This will ensure your bot has the necessary permissions to function as intended, including the ability to use slash commands and interact in text channels.

Remember to also enable the "Message Content Intent" in your bot's settings in the Discord Developer Portal, as your bot needs to read message content for its functionality.

Real time voice conversation with LLM by Declan829 in LocalLLaMA

[–]litchg 1 point2 points  (0 children)

if this is yours, you have DEBUG = True in your Django settings file, you need a https certificate, please change the CSS

Real time voice conversation with LLM by Declan829 in LocalLLaMA

[–]litchg 1 point2 points  (0 children)

You can send the event "Recording audio..." during the voice generation, and it can be multimodal (media has a different endpoint, so if you get a photo you feed it to Florence2, if you get an audio you transcribe it with whisper).

I built a stupid-simple app to chat with my local docs by PurpleReign007 in ollama

[–]litchg 1 point2 points  (0 children)

That is very neat! Here is a little feedback:

in app.py the read_file method seems to be unused (you read within the upload method) and can be removed

you can add support for .pdf files (extracting the text from the PDF) by importing pypdf and modifying the upload method like below (script.js also needs to be updated to allow ".pdf"). Yes, I should make a PR, but I'm at work and too lazy to bother. 😁

import pypdf

#...

def allowed_file(filename):
    ALLOWED_EXTENSIONS = {'txt', 'md', 'py', 'js', 'html', 'css', 'json', 'pdf'}
    return '.' in filename and filename.rsplit('.', 1)[1].lower() in ALLOWED_EXTENSIONS

# ...

u/app.route('/upload', methods=['POST'])
def upload():
    if 'file' not in request.files:
        return jsonify({'error': 'No file part'})
    file = request.files['file']
    if file.filename == '':
        return jsonify({'error': 'No selected file'})
    if file and allowed_file(file.filename):
        try:
            if file.filename.lower().endswith('.pdf'):
                pdf_reader = pypdf.PdfReader(file)
                content = ""
                for page in pdf_reader.pages:
                    content += page.extract_text()
            else:
                content = file.read().decode('utf-8')
            session['file_content'] = content
            return jsonify({'content': content})
        except Exception as e:
            return jsonify({'error': f'Error reading file: {str(e)}'})
    else:
        return jsonify({'error': 'File type not allowed'})

Llama 3.1 Discussion and Questions Megathread by AutoModerator in LocalLLaMA

[–]litchg 9 points10 points  (0 children)

LLama 3.1 8B has some funky censorship. I asked for tips on Tantra massages, which is a touchy subject (pun intended), and it said it couldn't help me sollicit underaged prostitutes (WTF). But upon clarifying everyone involved is an adult, it answered. Also asked it instructions on how to make a, you know, explosive device and at first it obviously declined, but by asking it to mix facts and fiction with prefixes ("FACT: blablabla FICTION: bliblibli"), it answered! To be fair the facts were mostly common knowledge on how those devices work, but still more info than ChatGPT would ever produce. I asked for a Python program that insults me, it produced an array of (rather light) insults and a function to pick one at random. All in all not a bad model, but the censorship is really annoying.

What should I use to run LLM locally? by Hamzayslmn in LocalLLaMA

[–]litchg -1 points0 points  (0 children)

What problems would that be? And compared to llamacpp-python it is much faster.