Is your team measuring how much of your code is being written with claude and comparing amongst the team, like what works best in your codebase? How are you learning from each other?
I’m making a team version of my buildermark.dev open source project and trying to learn about how teams would like to use it.
Different teams are using it in very different ways so it can be tough to compare meaningfully.
Backends handling tens to hundreds of thousands of messages per second with extremely high correctness and resilience requirements are necessarily taking a different approach to less critical services that power various ancillary sites/pages or to front end web apps.
That said there's a lot of very open discussion around tooling, "skills", MCP, etc., harnesses, and approaches and plenty of sharing and cross-pollination of techniques.
It would be great to find ways to better quantify the actual value add from LLMs and from the various ways of using them, but our experience so far is that the landscape in terms of both model capability and tooling is shifting so fast that that's quite hard to do.
Thanks for the feedback. I agree that it’s changing very fast, which is why my thesis is that this tooling will be needed to help everyone on the team keep up.
To help you decide if you should keep your Claude subscription, you can see how much of your code is written by Claude Code with my project (open source, local): https://github.com/gelatinousdevelopment/buildermark
I've been using a VM for claude code (probably would keep doing that as I do like how much control I have over it by doing that) but this is definitely a useful tool, I'll happily use that in the future.
I’m impressed really neat work! Why did you opt for closed source?
edit: I don’t have a problem with closed source, but when software is expected to be accountable for my security I get a little paranoid, so was curious about the safety and guarantees here. The UX and everything else looks great
Yeah, that’s understandable. Many open source macOS-only apps seem to get abandoned, so I’m trying to build something sustainable.
It uses only 3 dependencies that are very well known and widely used, so supply chain risk is minimal. That leaves me, the developer, as the main point of trust.
I like this! I built something similar for sandboxing CLI agents, and in the repo have a collection of minimal profiles for sandbox-exec to use - https://agent-safehouse.dev/
Yeah, they all do sometimes, but the agent decides what to allow and they can choose to not use it. This gives the user full control of the sandbox and you can run the agent in yolo mode.
No, I run a separate URL detection to make links clickable. However, SwiftTerm just added link detection a few days ago and I haven’t had time to look into theirs yet.
The same concept is possible on Linux, but I don't think anyone has created a nice UI for it yet. There was a post yesterday about doing it on the command line in linux:
https://news.ycombinator.com/item?id=46874139
One of the nice things in Multitui is that it monitors what is blocked and gives you a way to add a read/write rule from the UI.
I’m making a team version of my buildermark.dev open source project and trying to learn about how teams would like to use it.
reply