New in v12.0.0 - ✨ Agentic Workflows and much more #877
Replies: 3 comments 3 replies
-
WOW, this looks super cool! I'm also super happy about I wonder if I could use this Agentic Worfklow for typechecking TypeScript projects like this:
Looks very promising. Thank you!! |
Beta Was this translation helpful? Give feedback.
-
This is incredible, and I’m certain this was a massive undertaking. Thank you for all your hard work. I have a question about this:
I find this a bit risky, and it would be fantastic if there was some integration with Git. For instance, if there are any unstated hunks in the buffer, it would be great if the editor tool could automatically add/commit them before it starts overwriting the buffer. Perhaps we should do it manually now, though :) |
Beta Was this translation helpful? Give feedback.
-
This looks great, I'm looking forward to trying it out. Thanks so much for all your hard work on CodeCompanion. I really appreciate what you're doing, not just for adding great new features like this but the frequent incremental bug fixes and updates you put out so often. |
Beta Was this translation helpful? Give feedback.
-
Well this started off as a simple PR a week ago and morphed into something much more...
(Btw, no breaking changes in this release, just felt like bumping to v12.0.0)
Agentic Workflows
I've wanted to be able to benchmark different LLMs in CodeCompanion for a while. In truth, you always could, it was just incredibly manual. Combining that desire with my aim to improve workflows (which I suspect many people have never used or configured), we've ended up here:
AgenticWorkflows.mp4
To summarize what's happening:
Edit<->Test Workflow
The model I'm using in the video is
gpt-4o-2024-08-06
and I'm leveraging a new global variable to auto-approve all edits and not use diff mode. I must have tested this about 100 times now, and can confirm it works brilliantly with:claude 3.5-sonnet
o3-mini
o1-mini
gemini-2.0-flash
gpt-4o-2024-08-06
Unfortunately I've had no success with
gpt-3.5-turbo
,llama3.1
orqwen2.5-coder 7b
. It might be my setup and I should caveat these findings by saying I don't use locally hosted models very often.You can check out how I've implemented Agentic Workflows in the doc site.
In order to get the plugin to the state where Agentic Workflows could function, I had to add a lot of other functionality and fix a few things along the way...
Other changes in this release
Features
#buffer
parameters:#buffer{watch}
- To watch a buffer#buffer{pin}
- To pin a bufferI'd like to run the @cmd_runner tool
, the LLM would be sentI'd like to run the tool
. Now, the LLM will seeI'd like to run the cmd_runner tool
Fixes
:CodeCompanionChat
Refactors
#buffer:10-25
parameters have been removed. It had been broken for a while and no one noticedgemini-2.0-flash
Roadmap
I'm going to focus on improving the inline-assistant next. It's been buggy for too long for a lot of users on locally hosted models and it's too slow.
After that I'll likely streamline workspaces a little bit (move to a new schema version in the process) and then turn my attention to multi-model editing (#735). Regarding the latter...that might just be the answer to getting tools to work with small models.
Closing
As always, hope you like the direction the plugin is going in. Agentic Workflows will not be for everyone but the capabilities I've added to allow that functionality should be.
References
pytest
instead ofunittest
. The latter doesn't write output tostdout
so there would have been no visuals in the video for the user to observeBeta Was this translation helpful? Give feedback.
All reactions