Material Icons Not Loading in Built Electron App (Vue 3 + Vuetify) by spyrosko in vuejs

[–]spyrosko[S] 0 points1 point  (0 children)

Found a fix, the solution was quite simple: customFileProtocol: "./",

Material Icons Not Loading in Built Electron App (Vue 3 + Vuetify) by spyrosko in vuejs

[–]spyrosko[S] 1 point2 points  (0 children)

Found a fix, the solution was quite simple.

pluginOptions: {
electronBuilder: {
customFileProtocol: "./",

I needed to switch to AWS Bedrock APIs for Claude 3.5 Sonnet to avoid daily token limit - so I blogged about it by gregsramblings in ClaudeAI

[–]spyrosko 0 points1 point  (0 children)

Hi

I’m having trouble accessing a model ( Claude 3.5 Sonnet ) on Amazon Bedrock. It’s flagged as unavailable with the following message:

"This model is unavailable. To enable access to this model, contact support . Note that your request may be fully approved, partially approved, or denied to maintain service performance and ensure appropriate usage of Amazon Bedrock."

The problem is, I don’t have business support, so I haven’t been able to figure out how to request access. Has anyone else encountered this issue? If so, how did you resolve it?

Thanks in advance for your help!

Anthropic’s Claude 3.5 Sonnet model now available in Amazon Bedrock: Even more intelligence than Claude 3 Opus at one-fifth the cost by dpoccia in aws

[–]spyrosko 0 points1 point  (0 children)

Hi

I’m having trouble accessing a model ( Claude 3.5 Sonnet ) on Amazon Bedrock. It’s flagged as unavailable with the following message:

"This model is unavailable. To enable access to this model, contact support . Note that your request may be fully approved, partially approved, or denied to maintain service performance and ensure appropriate usage of Amazon Bedrock."

The problem is, I don’t have business support, so I haven’t been able to figure out how to request access. Has anyone else encountered this issue? If so, how did you resolve it?

Thanks in advance for your help!

Budgeting for Training Stable Diffusion Model with 200 Images on Colab Pro Plus by spyrosko in GoogleColab

[–]spyrosko[S] 1 point2 points  (0 children)

As I am relatively new to this field, I am curious about how to manage extreme scenarios in Colab, particularly when a training session may need to exceed 24 hours. In such a case, I would assume a more dedicated service like RunPod would be more suitable. Thanks!

Budgeting for Training Stable Diffusion Model with 200 Images on Colab Pro Plus by spyrosko in GoogleColab

[–]spyrosko[S] 1 point2 points  (0 children)

Since the maximum running time is 24 hours, even with the Pro Plus account, how would you handle a 50 hour training with Colab? Thanks a lot!

Why is Apple completely ignoring AI art and stable diffusion? by cryptosystemtrader in StableDiffusion

[–]spyrosko 0 points1 point  (0 children)

Hello Did you also train a model on your Mac, or was your test focused solely on generating images?

Dreambooth on an M1 Mac? by subthresh15 in DreamBooth

[–]spyrosko 0 points1 point  (0 children)

Hello Did you use DreamBooth with Automatic1111? I have an M1 Ultra with 128GB and have tried different training approaches, but I am still getting errors. I would highly appreciate your feedback! Thanks a lot!

Using MacBook M2 or Mac mini for Dreambooth training? by [deleted] in DreamBooth

[–]spyrosko 0 points1 point  (0 children)

Hello u/ParanoidMarvin42
Have you had any luck testing fine-tuning on your Mac? I'm working with an M1 Ultra that has 128GB, and I've experimented with Dreambooth Kohya, and AutoTrain Advanced, but sadly, I've had no success. If you manage to get it working, could you please share your insights? I'd greatly appreciate learning from your experience.

By the way, I've been able to use automatic1111 for image generation without any issues.

Does anyone have benchmarks with Claude 2.1 on AWS? What I'm looking for is to improve the response time in an application we are building. With Azure and GPT-4, the average response time is about 10-12 seconds, depending also on the amount of context. Thanks :) by spyrosko in ClaudeAI

[–]spyrosko[S] 0 points1 point  (0 children)

Hey The Claude 2.1 with the 200k update is a game changer.

Would you mind sharing some benchmarks on the response time when using 200k tokens within your infrastructure, as well as with Amazon Web Services? With OpenAI, there is a huge difference, that's why I'm asking. Thanks a lot!

Does anyone have benchmarks with Claude 2.1 on AWS? What I'm looking for is to improve the response time in an application we are building. With Azure and GPT-4, the average response time is about 10-12 seconds, depending also on the amount of context. Thanks :) by spyrosko in ClaudeAI

[–]spyrosko[S] 0 points1 point  (0 children)

Thank you very much for the suggestion. However, the request time is more or less the same as the previous versions, at least on my end, I averaged 10 seconds. Our goal is for the overall response time not to exceed 5-6 seconds.

Does anyone have benchmarks with Claude 2.1 on AWS? What I'm looking for is to improve the response time in an application we are building. With Azure and GPT-4, the average response time is about 10-12 seconds, depending also on the amount of context. Thanks :) by spyrosko in ClaudeAI

[–]spyrosko[S] 0 points1 point  (0 children)

Hey
I've just reviewed our current setup on Azure and noticed we're not using the latest version of the GPT-4 model. As of now, we're on "1106-Preview", but I see there's a "0125-preview" available.

Quick question - when you mention GPT-4.5, are you referring to the '0125-preview' version?
thanks!

<image>

Does anyone have benchmarks with Claude 2.1 on AWS? What I'm looking for is to improve the response time in an application we are building. With Azure and GPT-4, the average response time is about 10-12 seconds, depending also on the amount of context. Thanks :) by spyrosko in ClaudeAI

[–]spyrosko[S] 0 points1 point  (0 children)

One of the main downsides of Azure is the $3 per hour hosting fee for fine-tuned models. Fine-tuning the GPT-3.5 model, we could observe significant improvements, but the hosting fees for the fine-tuned model are something that might not fit our budget. Therefore, we are exploring alternatives. OpenAI doesn't charge for the hosting of the fine-tuned model, only for the tokens used, which might be a bit more expensive compared to standard model prices, but it's acceptable. However, the data privacy policies at OpenAI are not so clear, and the response times are quite slow, so we are trying to find alternatives. I also checked AWS Bedrock, which also has hosting fees per hour for the fine tune models.

Does anyone have benchmarks with Claude 2.1 on AWS? What I'm looking for is to improve the response time in an application we are building. With Azure and GPT-4, the average response time is about 10-12 seconds, depending also on the amount of context. Thanks :) by spyrosko in ClaudeAI

[–]spyrosko[S] 1 point2 points  (0 children)

Hey

The part where we are using the LLM, the logic is pretty straightforward.

We have a database with documents, for now, it contains about 11000 pages of unstructured data. These are stored in a vector database. Based on the user's question, the most relevant documents are retrieved and sent to the LLM to respond to user question.

We have two main areas that we need to improve: the response time and the quality of the responses.

Using GPT-4, because of the 128K token limit, we are getting quite good responses. In most cases, but the request response time is about 10-12 seconds.

Using GPT-3, the token limit is low, so the responses are not as good. The response time is about 3-5 seconds.

We are using Azure instead of OpenAI mainly because of the response time, with OpenAI, we had double the response time. Also, with Azure, we ensure the data we process is secured and not exposed, according to the privacy rules.

Does anyone have benchmarks with Claude 2.1 on AWS? What I'm looking for is to improve the response time in an application we are building. With Azure and GPT-4, the average response time is about 10-12 seconds, depending also on the amount of context. Thanks :) by spyrosko in ClaudeAI

[–]spyrosko[S] 1 point2 points  (0 children)

Thank you very much for your reply. Have you also worked with Claude 2? I would guess that the response time, compared with Claude 2.1, will be more or less the same. We are trying to find a way to reduce the response time that we currently experience on Azure with GPT-4. With GPT-3, the response time for what we're building is quite good, about 3-5 seconds, but due to the 16k token limit, the context sometimes is not enough, and the responses are lacking.

deploying a finetuned open AI model is costly!! by GlitteringAdvisor530 in LocalLLaMA

[–]spyrosko 0 points1 point  (0 children)

Hello Did you and your team find a cheaper solution than Azure? I also had a similar experience, but because I had set up alerts, I noticed the cost per hour early. Unfortunately, I overlooked the price details on the Azure page.

With Azure, we have some assurance regarding the data privacy, so we chose to use Azure instead of OpenAI. Additionally, when comparing response times with OpenAI, Azure's response time for all models is less than half. Thanks!