I have no experience with this ide but I see on the posted log on Reddit that the LLM is talking about a "step 620" - like this is hundreds of queries away from the initial one? The context must have been massive, usually after this many subsequent queries they start to hallucinating hardly
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
without permission
That's what she said. Enjoy your agent thing.
So many things wrong with this.
I am not a programmer by trade, and even though I learned programming in school, it's not a thing I want to spend a lot of time doing, so I do use AI when I need to generate code.
But I have a few HARD rules.
-
I execute all code and commands. Nothing gets to run on my system without me.
-
Anything which can be even remotely destructive, must be flagged and not even shown to me, until I agree to the risk.
-
All information and commands must be verifiable by sourcing documentary links, or providing context links that I can peruse. If documentary evidence is not available, it must provide a rationale why I should execute what it generates.
-
Every command must be accompanied by a description of what the command will do, what each flag means, and what the expected outcome is.
-
I am the final authority on all matters. It is allowed to make suggestions, but never changes without my approval.
Without these constraints, I won't trust it. Even then, I read all of the code it generates and verify it myself, so in the end, if it blows something up, I bear sole responsibility.
i really, really don't understand how this could happen. And how anyone would even want to enable the agent to perform actions without approval. Even in my previous work as a senior software developer, i never pushed any changes, never ran any command on non-disposable hardware, without having someone else double check it. why would you want to disable that?
why the hell aren't people running this shit in isolated containers?
Because people who runs this shit precisely don't know what containers, scope, permissions, etc are. That's exactly the audience.
No one ever claimed, that "artificial intelligence" would indeed be intelligent.
Amazing on so many levels.
Yet another reason to not use any of this AI bullshit