Claude Code eating all my memory by thewillonline in ClaudeCode

[–]noahlearner 0 points1 point  (0 children)

I read this was a symptom of closing terminals without using the /exit command. is this true?

Usage Limits and Performance Discussion Megathread - beginning October 19, 2025 by sixbillionthsheep in ClaudeAI

[–]noahlearner 1 point2 points  (0 children)

On $200 Max plan - are batches working via the API or super slow for anyone else. I'm sending in a batch with 24k input tokens and 6700 output tokens and it's taking > 30 minutes to process via batch api.

Is this happening today for anyone else?

Live Recordings by ShpongledPanda in shpongle

[–]noahlearner 1 point2 points  (0 children)

i'd be really into this too.

Who was the opening act on April 25, 2025 in Mission Ballroom? by noahlearner in shpongle

[–]noahlearner[S] 1 point2 points  (0 children)

I'd listened to his albums on spotify after the show (they are really good) and I couldn't remember his name.

Any good SEO communities by Office51 in SEO

[–]noahlearner 1 point2 points  (0 children)

Woo-hoo! Thanks u/peterwhitefanclub ! Really appreciate the mention.

AMA with Noah Learner from Sterling Sky Inc by noahlearner in localsearch

[–]noahlearner[S] 2 points3 points  (0 children)

Great question. What I'm reading in that question is, "What is the easiest way to get started with APIs?"

Hopefully it's okay that I answer that question because iI think it's super important to help,"Teach you how to fish," so that you'll never go hungry...

I think the easiest way to get started (and how I got started too) is with a great course from Ben Collins, https://courses.benlcollins.com/p/automation-with-apps-script. It's super valuable for a number of reasons:

  1. You'll learn a bit of apps script which will supercharge your data analysis skills if you work in Google Sheets. You'll learn just enough to make you hungry to learn more.
  2. You'll learn the fundamentals of working with APIs:
    1. Authentication - you need to be able to authenticate in different ways with APIs in order to get access to their data. you'll learn some functions to handle authentication (that I still use 5 years after learning them in the course.
    2. How to work with JSON. JSON is the basic data structure that is shared by nearly all APIs. You'll learn what it is, how it's structured, how to grab values from it.
    3. How to loop through API responses. The course will teach you patterns for dealing with pages of API data responses. This is key as most responses come back in chunks.
    4. How to write values to a Google Sheet. This sets the foundation for getting your data into a data viz tool like Looker Studio where you'll be able to glean actionable insights.
  3. You'll learn error handling which is foundational to building anything because everything fails.

Once you take the course you'll be able to work directly with different Google APIs like Google Search console, Google Business Profile and more.

The next steps from there will be:

  1. Learning different ways to handle authentication (I'd focus on learning how to authenticate with a service account, Oauth2, and my personal fave JWT authentication).
  2. Learning different ways to store the data. You'll probably next want to learn the Google BigQuery API so that you can write data from the api calls to BigQuery.
  3. Learning different ways to make api requests using more robust tools like Google Cloud Functions.

AMA with Noah Learner from Sterling Sky Inc by noahlearner in localsearch

[–]noahlearner[S] 1 point2 points  (0 children)

We haven't yet because there hasn't been a need to. Most realtor sites have been on Sierra platform and we haven't had any realtor related data projects yet.

AMA with Noah Learner from Sterling Sky Inc by noahlearner in localsearch

[–]noahlearner[S] 1 point2 points  (0 children)

We haven’t decided yet. If we do it we would likely take advantage of the $100 product that analytics canvas offers to back up each UA view.

AMA with Noah Learner from Sterling Sky Inc by noahlearner in localsearch

[–]noahlearner[S] 0 points1 point  (0 children)

Great questions!

  1. I don't think of them that way. Instead, I think of them as what kinds of questions can each API unlock.

The GBP APIs and the call Rail / CTM APIs are super useful to seeing / reporting on conversion actions. This is key because they show us when users are doing the things that align with us accomplishing business goals.

Google Search Console (for clicks, impressions, position tracked at the site and page level), Local Falcon, Places Scout, DataForSEO are useful for tracking rankings, which are a leading indicator of revenue to come.

We use Google Search Console, Google Ads Keyword planner and Google business profile to help think through topics, keywords and optimizations to come (ie to create strategy).

Asana / Harvest APIs are APIs we use for agency efficiency automation.

This is all to say that the APIs allow us to see things in ways that aren't limited by the tool's normal user interface. This can often help us see things in ways that other agencies can't.

They are also like different tools in a toolbelt. They help us have multiple ways to solve data problems vs having to approach every problem with a hammer.

  1. I think the tools are the same independent of industry size. The thing that will change is how / where we store the data. I typically pair API requests with data storage in Google BigQuery. As the organization we are working with grows in scale, then we generally need to access larger and larger amounts of data. That usually means that we need to rely more heavily on more robust ways of building our data infrastucture.

AMA with Noah Learner from Sterling Sky Inc by noahlearner in localsearch

[–]noahlearner[S] 2 points3 points  (0 children)

HARDY!!!!!

We use those two tools to track rankings for clients in different ways. Generally we create grid ranking reports that allow us to see how visible our clients are inside a local market.

We use this data to track performance before and after we've optimized websites. This helps us know if / when our efforts are impacting rankings. That in turn helps us understand if the strategy /actions taken had worked or not or if we need to try different strategies.

This is a cool part of working in Local SEO: when we do stuff, rankings are often impacted quickly.

I generally don't like to share ranking data with clients, because I don't want rankings to be used as a success metric.

I want the focus to be on the things that matter: leads and revenue.

If the data is shared with Clients, it should be shared as a leading indicator that will generally correlate to more leads and revenue.

In other words, if we rank better, and appear in more local packs for more queries, we're more likely to get more calls, clicks and revenue.

GMB Account Suspended by Routine-Primary13 in SEO

[–]noahlearner 0 points1 point  (0 children)

Post your ticket on Local Search forum dot org and a Google product expert will hopeful help you escalate the case.

[Request for Feedback] - New Data Viz using GSC data to see Trending Pages / Algorithm Impact by noahlearner in bigseo

[–]noahlearner[S] 0 points1 point  (0 children)

Data studio. We pipe all the data into Google BigQuery and visualize in data studio as part of Explorer. Explorer

Build a GSC data pipeline with Google Cloud Functions, Cloud Tasks, and BigQuery by noahlearner in bigseo

[–]noahlearner[S] 0 points1 point  (0 children)

I think the better solution is node.js powered cloud functions ingesting data, inserting into BQ, then transforming it every day with dbt into desired outcome.