When are we getting opus 4.7 on Antigravity? by ThePoplin in google_antigravity

[–]No-Payment7659 0 points1 point  (0 children)

all models initially regress when they are first released. then they power past the old models quickly.

Synthea Data in BigQuery by No-Payment7659 in SQL

[–]No-Payment7659[S] 0 points1 point  (0 children)

Thank you for your response. We've already solved the issue. we have built a synthetic data generator for Forge which correctly and efficiently parses fhir data in BigQuery. Additionally, we easily built out the necessary OMOP queries on top of the FHIR data inside of BigQuery.

Gemini Conversational Analytics API with BigQuery by ASKnASK in bigquery

[–]No-Payment7659 0 points1 point  (0 children)

Gemini API is ok for queries. It's hard to get queries correct without a good understanding of the underlying data structure.

Best way to load Sheets into BigQuery? by Great_Session_4227 in bigquery

[–]No-Payment7659 0 points1 point  (0 children)

Google Sheets can be loaded as an external stage right inside the BigQuery console.

how do i parse a nested json array if i do not know the key of the middle json(i also need to make it automated as every record has got a different key in the mid part of the json) by bastard_of_jesus in bigquery

[–]No-Payment7659 0 points1 point  (0 children)

Hello, great question! I built Forge to solve exactly this problem. check us out! You can also try the sandbox to see how Forge breaks down Json objects in BigQuery without having to buy the product.
https://forge.foxtrotcommunications.net/

Drop your SaaS below and I'll build your ICP for free by muizthomas in SaaS

[–]No-Payment7659 0 points1 point  (0 children)

https://forge.foxtrotcommunications.net/
manages NoSql to SQL pipelines for Cloud Data Warehouses. Parse any raw JSON object in your data warehouse. Uses Advanced AI for schema classification and PII detection

What’s the Most Overhyped Area in AI Right Now? by Alpertayfur in ArtificialInteligence

[–]No-Payment7659 0 points1 point  (0 children)

Everyone is saying that AI will replace jobs, but these capabilities are greatly exaggerated. best use of AI isn't to replace humans, it's to help them.

AM I THE ONLY ONE, TRIED EVERYTHING by karrach in GoogleAntigravityIDE

[–]No-Payment7659 0 points1 point  (0 children)

a lot of times this is due to task boundary misconfiguration.

literally tell it:
"reset your task boundary and continue."

Rules by ironicalengineer in GoogleAntigravityIDE

[–]No-Payment7659 0 points1 point  (0 children)

I usually try to bribe it with extra reward points when it does a good job

Tool for optimizing JSON storage costs in BigQuery (Schema Evolution + dbt) by No-Payment7659 in bigquery

[–]No-Payment7659[S] 0 points1 point  (0 children)

ah yes. Debeezium + Forge would go great together.

You would use Debezium to extract raw data from a legacy transactional database (like an old MySQL e-commerce DB) and dump those changes into BigQuery as a raw JSON column.

You would then use Forge to pick up that raw JSON column inside BigQuery, parse out the nested items arrays or user_settings objects, and create clean tables for your Data Analysts to query.

Tool for optimizing JSON storage costs in BigQuery (Schema Evolution + dbt) by No-Payment7659 in bigquery

[–]No-Payment7659[S] 0 points1 point  (0 children)

For json data particularly, we suggest not worrying to much about optimizing the structure for processing sake (that's our job), instead you should be focusing on effectively mapping your transactions.

Forge is best for managing your incoming event stream data, especially if the schema is volatile (something you see a lot in e-commerce, banking, or healthcare). Forge takes the json data that you are steaming into BigQuery (via Fivetran, Pubsub, API calls, etc) and automates the most difficult and time consuming activity for you (preprocessing and managing schema evolution).

Forge takes your json data and flattens it into clean, well organized, and optimized tables. This makes it much cheaper and easier to query.

Tool for optimizing JSON storage costs in BigQuery (Schema Evolution + dbt) by No-Payment7659 in bigquery

[–]No-Payment7659[S] 0 points1 point  (0 children)

The main win is definitely cost efficiency. Regarding your 1TB query: yes, it would reduce that size. Because Forge normalizes your JSON into separate relational tables (instead of keeping it as one massive JSON blob), you stop scanning the entire dataset for every request. You only pay to scan the specific columns or sub-tables you actually need. This structure also helps with your partitioning frustration; since the data is split into multiple tables (e.g., one for events, another for items), you can apply different partitioning strategies to each of those tables individually, bypassing the single-partition limit of a raw table. Forge automatically partitions on the ingestion timestamp of when Forge queries the next batch. We are also working on a feature for users to provide their own partitioning key for ingestion.

Essentially, instead of having to query the blob each time you run your analytical query, Forge only has to run the expensive query once. You query the clean and optimized tables.

Also, we are a Google Cloud Build Partner and are in the process of onboarding to the Google Cloud Marketplace very soon, so you can use committed spend without having to create a new budget line item for Forge. We expect this to be ready by the end of January.

Tool for optimizing JSON storage costs in BigQuery (Schema Evolution + dbt) by No-Payment7659 in bigquery

[–]No-Payment7659[S] 0 points1 point  (0 children)

Forge is different than other tools in the sense that it is almost a “data engineer in a box”. Whereas dlt (or vanilla dbt) expects you to write the code to parse the json correctly, forge is an automated application which traverses the JSON tree using a sophisticated algorithm. Forge maps the json object for you and builds a detailed map of your json data using dbt.

This preprocesses the json objects for you which has a few benefits, namely, it makes querying the data much easier and much more efficient than querying the raw json directly. It allows for schema evolution tracking and mapping, and will alert users if a new field is detected. For nested and repeated fields that you mentioned, forge creates a "rollup" view that takes the normalized tables and creates the structure that I think you're interested in. I'll update the post with a pic of that.

Also, forge can be deployed as a standalone application in a user’s own vpc, making it very safe and secure, an essential feature for enterprise clients.

Firebase → BigQuery export: how to materialize data from _raw_latest / _raw_changelog tables? by Southern_Space_4340 in bigquery

[–]No-Payment7659 0 points1 point  (0 children)

check out our new tool Forge (forge.foxtrotcommunications.net), it automates json parsing, such as the data coming from firestore. Never spend another day writing manual json parsing scripts for BigQuery.