That tracks with my experience
You have to very carefully scope things for them and have a plan for when they inevitably screw up.
This is a most excellent place for technology news and articles.
That tracks with my experience
You have to very carefully scope things for them and have a plan for when they inevitably screw up.
They’re great for bootstrapping in my experience but then really fall apart when you need it to do something surgical on a larger codebase.
Mine too
I’ve been working on an app and it was fantastic for the basics, then I decided to refactor an API and Claude code would run for hours without really getting there.
Also a good warning: I just had to completely rewrite an mcp server I had Claude build because when I needed to update it, the whole server was one giant if/else statement and utterly unmaintainable.
Yeah I was trying to pull out a nested react component and styles out of a larger component that got to be almost 1500 lines. Claude and GPT both struggled to get down what styles were required and what that subcomponent was actually doing. And generating tests around just made a fuck ton of spaghetti.
Which is fine. LLMs don’t have to be great at everything. But it’d be nice if people stopped saying I’m gonna be out of a job because of em.
Also a good warning: I just had to completely rewrite an mcp server I had Claude build because when I needed to update it, the whole server was one giant if/else statement and utterly unmaintainable.
I’ve noticed that in some of my bootstrapped code (also an MCP server :) ). I think it tends to bias towards single file solutions so it tends to be a lot less maintainable.
Maybe fastmcp is too new for Claude, it’s much less code and still one file
Is that why they like tailwind so much? Philosophically tailwind just seems unsustainable to me, css specifying the intent of an element seems nicer.
I’m having this argument with one of my junior guys who wants to just go with the generated code. We finally got his code functional, months late, and now need to get it maintainable
AI is a useful tool that can help speed up some of the tasks of coding but it’s not magical. It’s never a final result
AI could really help me get more done if we could weed out people following it blindly
At least with AI it's easy to see how shitty it gets as the codebase grows working on even a toy project over a week.
Then again, if you have no frame of reference maybe that doesn't feel as awful as it should.
I use LLMs for one thing only, turn my own ADHD ideas into something others can understand.
I use it to role play historical counter factuals, like how I could win the battle of Cannae through tactics, or how I could invent the telegraph in 13th century France. It's worth every watt <3
Wait a second this is brilliant! You can roleplay like a general in any history battles and see if you can do it differently!
Hah, I do sort of the opposite. My management has drunk the koolaid on ai - now I use it to translate my specs into something obviously ai generated for more acceptance
Yesterday I honestly had the bot flip and flop on an answer while deeply apologising and everytime saying "now you can actually trust, I checked and rechecked and now it's definitely correct".
Like a simple question with two choices.
It didn't know the answer so then it prompted to choose from two answers. One which confidently said one thing and one which confidently said the other. I called it out on making me choose the answer to a question I asked. Then it decided itself. I then questioned it. It changed it mind. And around and around we went for like 20 minutes, everytime it swearing this time it isn't hallucinating.
They suck so bad for most things, but they're useful for some very niche things, but even for those, they still sort of suck at those as well a not non-significant. They're definitely shouldn't be used for official shit, but they very much are.