Service templates are a typical constructing block within the “golden paths” organisations construct for his or her engineering groups, to make it straightforward to do the correct factor. The templates are imagined to be the function fashions for all of the companies within the organisation, all the time representing the hottest coding patterns and requirements.
One of many challenges with service templates although is that when a crew instantiated a service with one, it’s tedious to feed template updates again to these companies. Can GenAI assist with that?
Reference utility as pattern supplier
As half of a bigger experiment that I just lately wrote about right here, I created an MCP server that offers a coding assistant entry to coding samples for typical patterns. In my case, this was for a Spring Boot net utility, the place the patterns had been repository, service and controller courses. It’s a properly established prompting follow at this level that offering LLMs with examples of the outputs that we would like results in higher outcomes. To place “offering examples” into fancier phrases: That is additionally known as “few-shot prompting”, or “in-context studying”.
Once I began working with code samples in prompts, I shortly realised how tedious this was, as a result of I used to be working in a pure language markdown file. It felt a little bit bit like writing my first Java exams at college, in pencil: You have got know concept if the code you’re writing really compiles. And what’s extra, in the event you’re creating prompts for a number of coding patterns, you wish to hold them according to one another. Sustaining code samples in a reference utility challenge you can compile and run (like a service template) makes it rather a lot simpler to offer AI with compilable, constant samples.
Detect drift from the reference utility
Now again to the issue assertion I discussed at the start: As soon as code is generated (be that with AI, or with a service template), after which additional prolonged and maintained, codebases usually drift away from the function mannequin of the reference utility.
So in a second step, I puzzled how we’d use this strategy to do a “code sample drift detection” between the codebase and the reference utility. I examined this with a comparatively easy instance, I added a logger and log.debug statements to the reference utility’s controller courses:
Then I expanded the MCP server to offer entry to the git commits within the reference utility. Asking the agent to first search for the precise modifications within the reference offers me some management over the scope of the drift detection, I can use the commits to speak to AI precisely what sort of drift I’m concerned with. Earlier than I launched this, once I simply requested AI to match the reference controllers with the prevailing controllers, it went a bit overboard with a number of irrelevant comparisons, and I noticed this commit-scoping strategy have affect.
In step one, I simply requested AI to generate a report for me that recognized all of the drift, so I might assessment and edit that report, e.g. take away findings that had been irrelevant. Within the second step, I requested AI to take the report and write code that closes the gaps recognized.
When is AI bringing one thing new to the desk?
A factor so simple as including a logger, or altering a logging framework, can be executed deterministically by codemod instruments like OpenRewrite. So bear that in thoughts earlier than you attain for AI.
The place AI can shine is every time now we have drift that wants coding that’s extra dynamic than is feasible with regular-expression-based codemod recipes. In a complicated type of the logging instance, this could be turning non-standardised, wealthy log statements right into a structured format, the place an LLM could be higher at turning all kinds of current log messages into the respective construction.
The instance MCP server is included in the repository that accompanies the unique article.







