The way you teach an LLM, outside of training your own, is with rules files and MCP tools. Record your architectural constraints, favored dependencies, and style guide information in your rule files and the output you get is going to be vastly improved. Give the agent access to more information with MCP tools and it will make more informed decisions. Update them whenever you run into issues and the vast majority of your repeated problems will be resolved.
If it’s doesn’t work for you, it’s because you’re a failure!
Still not convinced these LLM bros aren’t junior developers (at best) who someone gave a senior title to because everyone else left their shit hole company.
More to the point, that is exactly what the people in this study were doing.
They don’t really do into a lot of detail about what they were doing. But they have a table on limitations of the study that would indicate it is not.
We do not provide evidence that: There are not ways of using existing AI systems more effectively to achieve positive speedup in our exact setting. Cursor does not sample many tokens from LLMs, it may not use optimal prompting/scaffolding, and domain/repository-specific training/finetuning/few-shot learning could yield positive speedup.
Back to this:
even if it did it’s not any easier or cheaper than teaching humans to do it.
In my experience, the kinds of information that an AI needs to do its job effectively has a significant overlap with the info humans need when just starting on a project. The biggest problem for onboarding is typically poor or outdated internal documentation. Fix that for your humans and you have it for your LLMs at no extra cost. Use an LLM to convert your docs into rules files and to keep them up to date.
Your argument depends entirely on the assumption that you know more about using AI to support coding than the experienced devs that participated in this study. You want to support that claim with more than a “trust me, bro”?
Do you think that like nobody has access to AI or something? These guys are the ultimate authorities on AI usage? I won’t claim to be but I am a 15 YOE dev working with AI right now and I’ve found the quality is a lot better with better rules and context.
And, ultimately, I don’t really care if you believe me or not. I’m not here to sell you anything. Don’t use it the tools, doesn’t matter to me. Anybody else who does use them, give my advice a try an see if it helps you.
Again, read and understand the limitations of the study. Just the portion I quoted you alone is enough to show you that you’re leaning way too heavily on conclusions that they don’t even claim to provide evidence for.
That is a moronic take. You would be better off learning to structure your approach to SW development than trying to learn how to use a glorified slop machine to plagiarize other people’s works.
In practice I find the more stuff like this you throw at it the more rope it has to hang itself with. And you spend so much time prompt adjusting so it doesn’t do the wrong things that you were better off just doing half of the tasks yourself.
Unless you are retraining the model locally at your 23 acre data center in your garage after every interaction, it’s still not learning anything. You are just dumping more data in to its temporary context.
And lots fit on personal computers dude, do you even know what different llms there are…?
One for programming doesn’t need all the fluff of books and art, so now it’s a manageable size. Llms are customizable to any degree, use your own data library for the context data even!
“Customizing” is just dumping more data in to it’s context.
Yes, which would fix the incorrect coding issues. It’s not an llm issue, it’s too much data. Or remove the context causing that issue. These require a little legwork and knowledge to make useful. Like anything else.
You do understand that the model weights and the context are not the same thing right? They operate completely differently and have different purposes.
Trying to change the model’s behavior using instructions in the context is going to fail. That’s like trying to change how a word processor works by typing in to the document. Sure, you can kind of get the formatting you want if you manhandle the data, but you haven’t changed how the application works.
Because I work with LLMs daily. I understand how they work. No matter how much I type at an LLM, its behavior will never fundamentally change without regenerating the model. It never learns anything from the content of the context.
The model is the LLM. The context is the document of a word processor.
A Jr developer will actually learn and grow in to a Sr developer and will retain that knowledge as they move from job to job. That is fundamentally different from how an LLM works.
I’m not anti-AI. I’m not “crying” about their issues. I’m just discussing the from a practical standpoint.
Where do you think the errors are coming from? From data bleed over, the word “coding” shows up in books, so yes the context would incorrectly pull book data too.
Or do you not realize coding books exist as well…? And would be in the dataset.
If it’s constantly making an error, fix the context data dude. What about it an llm/ai makes you think this isn’t possible…? Lmfao, you just want to bitch about ai, not comprehend how they work.
Yeah, but LLMs still consistently don’t follow all rules they’re given, they randomly will not follow one or more with no indication they did so, so you can’t really fix these issues consistently, just most of the time.
Edit: to put this a little more clearly after a bit more thought: It’s not even necessarily a problem that it doesn’t always follow rules, it’s more so a problem that when it doesn’t follow the rules, there’s no indication it did so. If it had that, it would actually be fine!
… That keeps making the same mistakes over and over again because it never actually learns from what you try to teach it.
Yep, the junior is capable of learning.
Wait till I get hired as junior
Yeah, not all people who enter the industry should be doing so.
Most of this was boomers being boomers and claiming anyone and everyone should code.
My job believes the solution to this is a 7,000 line agents.md file
Sometimes. And if they’re not, they’ll be replaced or replace themselves.
This is not really true.
The way you teach an LLM, outside of training your own, is with rules files and MCP tools. Record your architectural constraints, favored dependencies, and style guide information in your rule files and the output you get is going to be vastly improved. Give the agent access to more information with MCP tools and it will make more informed decisions. Update them whenever you run into issues and the vast majority of your repeated problems will be resolved.
Well, that’s what they say, but then it doesn’t actually work, and even if it did it’s not any easier or cheaper than teaching humans to do it.
More to the point, that is exactly what the people in this study were doing.
If it’s doesn’t work for you, it’s because you’re a failure!
Still not convinced these LLM bros aren’t junior developers (at best) who someone gave a senior title to because everyone else left their shit hole company.
They don’t really do into a lot of detail about what they were doing. But they have a table on limitations of the study that would indicate it is not.
Back to this:
In my experience, the kinds of information that an AI needs to do its job effectively has a significant overlap with the info humans need when just starting on a project. The biggest problem for onboarding is typically poor or outdated internal documentation. Fix that for your humans and you have it for your LLMs at no extra cost. Use an LLM to convert your docs into rules files and to keep them up to date.
Your argument depends entirely on the assumption that you know more about using AI to support coding than the experienced devs that participated in this study. You want to support that claim with more than a “trust me, bro”?
Do you think that like nobody has access to AI or something? These guys are the ultimate authorities on AI usage? I won’t claim to be but I am a 15 YOE dev working with AI right now and I’ve found the quality is a lot better with better rules and context.
And, ultimately, I don’t really care if you believe me or not. I’m not here to sell you anything. Don’t use it the tools, doesn’t matter to me. Anybody else who does use them, give my advice a try an see if it helps you.
These guys all said the same thing before they participated in a study that proved that they were less efficient than their peers.
Again, read and understand the limitations of the study. Just the portion I quoted you alone is enough to show you that you’re leaning way too heavily on conclusions that they don’t even claim to provide evidence for.
Codex literally lies about being connected to configured MCP servers.
Are you trying to make a point that agents can’t use MCP based off of a picture of a tweet you saw or something?
I’m talking from my personal, daily experience using codex.
That is a moronic take. You would be better off learning to structure your approach to SW development than trying to learn how to use a glorified slop machine to plagiarize other people’s works.
In theory yes.
In practice I find the more stuff like this you throw at it the more rope it has to hang itself with. And you spend so much time prompt adjusting so it doesn’t do the wrong things that you were better off just doing half of the tasks yourself.
This is why you use a downloaded llm and customize it, there’s ways to fix these issues.
Unless you are retraining the model locally at your 23 acre data center in your garage after every interaction, it’s still not learning anything. You are just dumping more data in to its temporary context.
Sounds like you have no clue what an LLM/AI actually is or is capable of.
https://medium.com/sciforce/step-by-step-guide-to-your-own-large-language-model-2b3fed6422d0
It’s not hard to keep a data library updated for context, and some are under a TB in siz.
Where are you getting your information from?
It seems you are still confusing context with training? Did you read that text and understand it?
Did you follow it yourself to build an llm?
I bet they had an LLM read it and summarize it for them
Why do you think it’s solely a training issue?
So, you did not? Ok
Can’t answer the question eh?
What a shocker.
If you can’t explain your or justify your side, I’ve got no time for people like you.
What part of customize did you not understand?
And lots fit on personal computers dude, do you even know what different llms there are…?
One for programming doesn’t need all the fluff of books and art, so now it’s a manageable size. Llms are customizable to any degree, use your own data library for the context data even!
What part about how LLMs actually work do you not understand?
“Customizing” is just dumping more data in to it’s context. You can’t actually change the root behavior of an LLM without rebuilding it’s model.
Yes, which would fix the incorrect coding issues. It’s not an llm issue, it’s too much data. Or remove the context causing that issue. These require a little legwork and knowledge to make useful. Like anything else.
You really don’t know how these work do you?
You do understand that the model weights and the context are not the same thing right? They operate completely differently and have different purposes.
Trying to change the model’s behavior using instructions in the context is going to fail. That’s like trying to change how a word processor works by typing in to the document. Sure, you can kind of get the formatting you want if you manhandle the data, but you haven’t changed how the application works.
Why are you so focused on just the training? The data is ALSO the issue.
Of course if you ignore one fix, that works, of course you can only cry it’s not fixable.
But it is.
Because I work with LLMs daily. I understand how they work. No matter how much I type at an LLM, its behavior will never fundamentally change without regenerating the model. It never learns anything from the content of the context.
The model is the LLM. The context is the document of a word processor.
A Jr developer will actually learn and grow in to a Sr developer and will retain that knowledge as they move from job to job. That is fundamentally different from how an LLM works.
I’m not anti-AI. I’m not “crying” about their issues. I’m just discussing the from a practical standpoint.
LLMs do not learn.
But
Is not inside the context, that comes from training. So you know how an llm works?
Where do you think the errors are coming from? From data bleed over, the word “coding” shows up in books, so yes the context would incorrectly pull book data too.
Or do you not realize coding books exist as well…? And would be in the dataset.
Why would you put whole books into the context?!? Do you even know what an llm is?
If it’s constantly making an error, fix the context data dude. What about it an llm/ai makes you think this isn’t possible…? Lmfao, you just want to bitch about ai, not comprehend how they work.
This is Lemmy, bitching about AI is the norm.
Yeah, but LLMs still consistently don’t follow all rules they’re given, they randomly will not follow one or more with no indication they did so, so you can’t really fix these issues consistently, just most of the time.
Edit: to put this a little more clearly after a bit more thought: It’s not even necessarily a problem that it doesn’t always follow rules, it’s more so a problem that when it doesn’t follow the rules, there’s no indication it did so. If it had that, it would actually be fine!