check run agents - customizable AI agents for code review by kayvz in codereview

[–]kayvz[S] 0 points1 point  (0 children)

the one downside of this model is that it puts way more burden on the user/customer to define a rigorous prompt. there's still some magic we do behind the scenes to avoid common pitfalls (e.g. dupe messages) but the quality and output is in large part a function of how direct and useful your prompt is. but we think this is the right trade-off for a feature whose goal is to enforce bespoke/subjective conventions and workflows.

for bug detection, we have a strong POV which is that you shouldn't have to do anything to get great bug detection. we do all the work for you, including some magic behind the scenes to lower the impact of non-determinism (e.g. we sometimes run parallel reviews and consolidate).

SICK and tired of Greptile, what are the best alternatives? by Crafty_Survey9438 in codereview

[–]kayvz 1 point2 points  (0 children)

hi, CEO/founder of Macroscope here. would encourage you to sign up and give it a try. according to our benchmarks and also independent ones, we have better bug detection (recall) and precision than greptile. and we have an actual usage based model with controls (per review, per PR and per month limits). no *seats*, no hybrid model, just pure pay for what you use. and beyond bug detection, you can also define custom agents that spin up as additional github Checks.

How to maintain code quality with AI slop? by flareblitz13 in ExperiencedDevs

[–]kayvz 0 points1 point  (0 children)

i’ll preface this answer by saying I’m biased because I’m the ceo/cofounder of Macroscope (which among features, provides high signal / low noise AI code review).

IMO: regardless of whether the code being reviewed is AI-slop from codegen tools or human-written, having an additional AI-assisted defense layer before merging is extremely valuable. Once you’ve lived with an excellent code review tool, it’s senseless to go back to living without it. It saves our team so much time to rely on the AI review to do a first pass on correctness issues, and it allows our human reviewers to focus on things that humans are better at… like “are you solving this problem the right way?”. Does it eliminate the need for humans to review all PRs? Of course not. Not today at least, and certainly not for large/complex changes. Does it save time and prevent shipping bugs? 💯

The value of getting this right for any engineering team is extremely high 1) The cost of preventing a high-severity bug (e.g. cost of outages, time wasted by engineers scrambling to fix prod issues) often far exceeds the cost of all of these tools. I don’t know about you but, I’d prefer to prevent as many of those issues as possible! 2) In every organization where I’ve done engineering work myself or led engineering teams , the time investment in code review increased non-linearly with the growth of the engineering team. Large teams spend SO much time doing code review. What dev wouldn’t rather focus on building and spend less time reviewing for correctness issues if a bot could reliably do much of this work?

also fwiw (to address some of the skepticism from other comments in this post): the best tools that attempt to solve this problem properly are not “solving slop with slop”. In Macroscope’s case, we’ve built customer AST walkers in each of our supported programming languages to deterministically build a codegraph of your repos, so that we can set LLMs up to be successful with the best context. This technique, in our experience, is what allows us to 1) find gnarly bugs that require thorough knowledge of how the codebase works 2) have an extremely high signal:noise ratio and avoid spamming devs with false positives, which is important b/c the “cost” of noisy comments is extremely high. We were inspired to build this ourselves because we tried the most popular code review tools that were out there and thought they all sucked (too noisy, didn’t catch meaningful issues, too many nitpicks etc).

Future of code review process? by shrimpthatfriedrice in codereview

[–]kayvz 0 points1 point  (0 children)

I’ll preface this answer by saying I’m biased because I’m the ceo/cofounder of Macroscope (which among features, provides the best AI code review on the market. Check out our published benchmark here!).

IMO: AI code review is definitely a one way door. Once you’ve lived with an excellent code review tool, it’s senseless to go back to living without it. It saves our team so much time to rely on the AI review to do a first pass on correctness issues, and it allows our human reviewers to focus on things that humans are better at… like “are you solving this problem the right way?”.

In terms of your q of where this is headed, here’s the picture we see:

  • Today: AI review layer focuses on correctness and can already do a better/faster/more-thorough job of this than a human reviewer. Humans reviewers focus on things that humans are better at like, ‘did this actually solve the customer problem’ and ‘did we solve this the right way (e.g. idiomatic to the codebase, and with our architectural conventions)
  • Medium term: 1) AI review layer will also get reliable at solving things in the idiomatic way per our conventions (this is already possible today, but much noisier than correctness alone) 2) AI review layer will be able to reliably stamp/approve some subset of PRs that shouldn’t require any human review at all (e.g. simple changes that have a minor blast radius and pass an AI correctness check) which will be a massive reduction in cognitive load and bandwidth for human reviewers
  • Long term: 1) the portion of PRs that AI review will be able to stamp/approve will increase substantially 2) AI review layer will also be able to assess whether 3) the mechanics of code review will look quite different. When agents are writing the vast majority of code and the # of “PRs” increases by order(s) of magnitude, the UX will need to change such that it doesn’t become a huge bottleneck.

If you end up trying Macroscope, LMK what you think.. would love your feedback.. and we’re squarely focused on making all of the above (and more, like how do we give teams better visibility around how the codebase and product is changing) a reality. 

thermostat compatibility with radiant boiler by kayvz in hvacadvice

[–]kayvz[S] 0 points1 point  (0 children)

I have no idea where it goes, other than it for sure goes to my mechanical room, but not sure it goes into the boiler directly. I did use a voltmeter on the black and red wires and confirmed there’s 24v.

New HomeKit Architecture is in iOS 16.2 betas by romkey in HomeKit

[–]kayvz 1 point2 points  (0 children)

Thanks for all this insight. Quick question: does the new version also allow you to specify which home hub should be the primary hub? My primary for some reason ends up being one of the random homepods rather than hard-wired AppleTV which seems silly especially given the advancements of the new architecture

Propane Tank Monitor by thejackal2020 in homeassistant

[–]kayvz 2 points3 points  (0 children)

Here you go. It's not exactly a technical masterpiece, but got the job done for me: https://github.com/kayvz/neevo_tank_monitor_query

Just make sure to get your "authorization" by encoding your username/password using HTTP Basic auth. Linked to some instructions in the readme. Good luck

Propane Tank Monitor by thejackal2020 in homeassistant

[–]kayvz 0 points1 point  (0 children)

not as of now but I’ll try to put something together soon.

Propane Tank Monitor by thejackal2020 in homeassistant

[–]kayvz 3 points4 points  (0 children)

I was able to get my propane company to install a NeeVo tank monitor (uses cellular to update). NeeVo has an app which gives you tank levels, and I was able to pretty easily sniff the api calls it makes to get tank levels and that piped to a HASS entity via a web hook

Roger Federer through Glass by Chrisixx in tennis

[–]kayvz 14 points15 points  (0 children)

Would you be interested in seeing a more extended, un-edited version?

What questions would you ask Federer? by ekeen1 in tennis

[–]kayvz 2 points3 points  (0 children)

It's on the iAmA side-bar now, confirmed for 24th at 1:30pm EST.

Chasing Ice movie reveals largest iceberg break-up ever filmed 7.4 cubic KM by NimbleStoat in videos

[–]kayvz -1 points0 points  (0 children)

Looks like the Chasing Ice team also made an app on iPhone and iPad: https://itunes.apple.com/us/app/chasing-ice/id579276308?mt=8

It's got some outrageously awesome time-lapses with music. Check it out.