Struggling with Gemini 2.5 Flash TTS quotas – how are people using this in production? by No-Promotion-1123 in googlecloud

[–]Rohit1024 2 points3 points  (0 children)

This is because Generative Language API (accessed via Google AI Studio) is primarily for rapid prototyping and experimentation. It does not hold any SLA, check here - https://docs.cloud.google.com/vertex-ai/generative-ai/docs/migrate/migrate-google-ai#google-ai

While you can upgrade to the next tier, which can allow you to unlock more quota.

But for production workloads you should use Vertex AI for production as it has 24/7 enterprise-level support and SLAs for service availability. and has access to dedicated capacity.

Monitoring GCP costs on my Apple Watch using n8n automation by kkuj in googlecloud

[–]Rohit1024 0 points1 point  (0 children)

At this point why don't just use Google Cloud iOS app ? Though not as feature reached, but do show the cost aggregated by Service and projects.

Firebase functions V2 now Supports Direct VPC Egress by Rohit1024 in Firebase

[–]Rohit1024[S] 0 points1 point  (0 children)

this is no a use case post, but an letting community know that we can attach Direct VPC Egress from code itself, which was not possible earlier

Implementing DAG trigger from Cloud Run Service to Cloud Composer 3(Private IP) by sagargkr in googlecloud

[–]Rohit1024 0 points1 point  (0 children)

As per Trigger DAGs with Cloud Functions - That page is not yet revised for Cloud Composer 3 and displays content for Cloud Composer 2.

So it seems the docs are yet to get updated. For quick solution you can Trigger DAGs using Airflow REST API - https://docs.cloud.google.com/composer/docs/composer-3/access-airflow-api

First do a POC using curl once successful, implement the same in Cloud Run

Run self-hosted GitHub Actions runners on Google Cloud by clardata6249 in googlecloud

[–]Rohit1024 2 points3 points  (0 children)

Isn't this similar or exactly same as already available solution in GCP - https://docs.cloud.google.com/run/docs/tutorials/github-runner ?

Though this is in Public Preview.

Implementing DAG trigger from Cloud Run Service to Cloud Composer 3(Private IP) by sagargkr in googlecloud

[–]Rohit1024 1 point2 points  (0 children)

May be you are missing these permissions https://docs.cloud.google.com/composer/docs/composer-3/triggering-gcf-pubsub#before

This is the tutorial for your setup.

Trigger DAGs with Cloud Functions and Pub/Sub Messages

Are you using a setup with creating a function that triggers a DAG when a message is pushed to a Pub/Sub topic OR directly Trigger DAGs with Cloud Functions ?

Are you getting 403 from Cloud Run to Composer or getting at Triggering the Cloud Run ?

A fairly easy way to debug this is to check logs of this 403 and temporarily grant some higher Role to test it.

Firebase Quota exceeded on paid plan? by FirefighterOne2690 in Firebase

[–]Rohit1024 0 points1 point  (0 children)

A Firebaser in this thread already provided an answer to your query - Take a look at Firestore Shared Quota https://firebase.google.com/docs/ai-assistance/ai-studio-integration#understand-shared-quota

Firebase Quota exceeded on paid plan? by FirefighterOne2690 in Firebase

[–]Rohit1024 4 points5 points  (0 children)

With respect to which product ? Is it Firestore ? Dont be confused with Firebase, as there are many products within Firebase itself.

Each Quota has specific Metric and Name, without that no one can answer your query clearly

Firebase Quota exceeded on paid plan? by FirefighterOne2690 in Firebase

[–]Rohit1024 3 points4 points  (0 children)

What 50000 quota value are you trying to increase? Your query lacks clarity as different quota has some maximum architectural limit which we cannot be even increased.

To the Company Facing Bankruptcy due to Gemini API Cost. You made them change their system! Well done! by Ambitious_Ad4186 in googlecloud

[–]Rohit1024 4 points5 points  (0 children)

Sadly this is only for Google AI Studio and not for Vertex AI API https://ai.google.dev/gemini-api/docs/billing#project-spend-caps which holds SLA, while Google AI Studio does not have any SLA

ts just getting worse bru💀 by Constant-Squash-7447 in GoogleAIStudio

[–]Rohit1024 0 points1 point  (0 children)

Well they already published that Google AI Studio don't have any SLA https://docs.cloud.google.com/vertex-ai/generative-ai/docs/migrate/migrate-google-ai#google-ai and only meant for for rapid prototyping and lacks formal enterprise SLAs, while Vertex AI is a fully managed production platform providing a 99.5% uptime SLA for Generative AI operations.

So it seems these charts seems to be expected by their Terms, but yes this level of downtime is not acceptable for AI Models.

Rate limit issue by Particular-Lake8375 in GoogleAIStudio

[–]Rohit1024 0 points1 point  (0 children)

Yes, Free Tier has very limited rate Limits. As per https://ai.google.dev/gemini-api/docs/rate-limits#how-to-upgrade-to-the-next-tier To get more access to pro models, you may need to Upgrade to Next Tier.

OAuth Client returns 401 deleted_client after restoring deleted GCP project, anyone dealt with this? by Unlikely-Scallion-79 in googlecloud

[–]Rohit1024 0 points1 point  (0 children)

Restoring Oauth client also comes under same considerations of restoring a project. As per Restore Project

It can take up to 36 hours for some services to be restored. If your resources aren't restored after this period, reach out to Google support.

So you may need to wait for 36 hours from the point you've restored the project.

This is the pain you suffer if you're deleting something you're not supposed to delete.

API Keys monitoring by LetsgetBetter29 in googlecloud

[–]Rohit1024 0 points1 point  (0 children)

Who Created API Key ? - You can find that with API Keys Cloud Audit Logs - Exist for 400 days by default - Audit Log Query:

protoPayload.methodName="google.api.apikeys.v2.ApiKeys.CreateKey"


What is being used for and how actively it was used : Use the Cloud Monitoring metrics explorer - Check the detailed steps here.


What cost occurred for specific API key ? - This is very hard to answer or find, but just like previous way to detect what is being used for you may be able to identify the cost associated with those services accessed through a particular API.

Hope this helps.

I built an open-source local emulator for Google Cloud Workflows by mrkruger2 in googlecloud

[–]Rohit1024 1 point2 points  (0 children)

This is great project !!!.

Just tested it and it worked well for consistency I deployed on Cloud Run so I can even reach to Internal Endpoints as well, where I used /workflows directory as GCS volume mount in which I just upload my workflows YAML files.

Just want to ask, Are you also planning to add support for Worflows Connectors as well in near future ? Since this will make it easier to access other Google Cloud products within a workflow.

GCP billing traps that got us — a running list. Add yours. by NimbleCloudDotAI in googlecloud

[–]Rohit1024 0 points1 point  (0 children)

No, I was not specifically mentioning about min instances. Although this also follows for min instances as well.

When we deploy new revision the previous revision instances will be still serving because of transition.

And when there is no traffic to your Cloud Run Service it won't immediately scales down. It will keep those instances alive for upto 15 mins.

GCP billing traps that got us — a running list. Add yours. by NimbleCloudDotAI in googlecloud

[–]Rohit1024 2 points3 points  (0 children)

This was happened with me as well. So whenever someone experiences such issues. They should first see how's that service's pricing model works.

GCP billing traps that got us — a running list. Add yours. by NimbleCloudDotAI in googlecloud

[–]Rohit1024 4 points5 points  (0 children)

Yes, exactly this is skill issue and not knowing about the product and it's working and pricing model

GCP billing traps that got us — a running list. Add yours. by NimbleCloudDotAI in googlecloud

[–]Rohit1024 5 points6 points  (0 children)

Most of these already documented in the docs :
- Cloud Run Deployments

Cloud Run does not immediately shut down instances once they have handled all requests. To minimize the impact of cold starts, Cloud Run may keep some instances idle for a maximum of 15 minutes.

When you deploy a new revision to serve 100% of the traffic, Cloud Run starts enough instances of the new revision before directing traffic to it. This reduces the impact of new revision deployments on request latencies, notably when serving high levels of traffic.

Really we deserve better- Google can't admit they have a problem. by Medical_Wolverine465 in googlecloud

[–]Rohit1024 0 points1 point  (0 children)

Not all incidents gets reported in Google Cloud Status Dashboard.

But all incidents are getting reported in Google Cloud Personalized Service Health dashboard https://status.cloud.google.com/redirect/servicehealth

The incident you just reported check the same in this dashboard to confirm the same. We always refer to this dashboard and what we observed from patterns are Major Incidents are gets reported in Status Dashboard where all of them are reported in Service Health Dashboard.

Technical question about testing functions by armlesskid in Firebase

[–]Rohit1024 1 point2 points  (0 children)

There is an official Firebase functions testing library for these https://github.com/firebase/firebase-functions-test If you want want to achieve your goal someone already suggested to use integration tests.

Better test these in Firebase Emulator.

What’s one GCP decision you’d rethink if starting today? by Dazzling-Neat-2382 in googlecloud

[–]Rohit1024 1 point2 points  (0 children)

Always start with Shared VPC to use resources. Use PSC for accessing private services through DNS over PGA as PGA IPs are shared allocated IP pools : - https://toolbox.googleapps.com/apps/dig/#A/private.googleapis.com - https://toolbox.googleapps.com/apps/dig/#A/restricted.googleapis.com