The video shows the Godot code editor with some unfinished code. After the user presses a button offscreen, the code magically completes itself, seemingly due to an AI filling in the blanks. The examples provided include a print_hello_world function and a vector_length function. The user is able to accept and decline the generated code by pressing either tab or backspace
This is an addon I am working on. It can help you write some code and stuff.
It works by hooking into your local LLMs on ollama, which is a FOSS way to run large language models locally.
Here's a chat interface which is also part of the package
video description
The video shows a chat interface in which the user can talk to a large language model. The model can read the users code an answer questions about it.
Do you have any suggestions for what I can improve? (Besides removing the blue particles around the user text field)
Important: This plugin is WIP and not released yet!
I will likely post on here when I release the plugin to GitLab and the AssetLib.
But I also don't want to spam this community, so there won't be many, if any updates until the actual release.
If you want to have something similar right now, there is Fuku for the chat interaction and selfhosted copilot for code completion on the AssetLib! I can't get the code completion one to work, but Fuku works pretty well, but can't read the users code at all.
Currently the completion is implemented via keyboard shortcut.
Would you prefer it, if I made it automatically complete the code? I personally feel, that intentionally asking for it to complete the code is more natural than waiting for it to do so.
Are there some other features you would like to see? I am currently working on a function-refactoring UI.
As far as other features I want; I don't want any. I just want code completion via keyboard shortcut.
I think a hard aspect is figuring out what context to feed the LLM. Iirc GitHub Copilot only feeds what is in the current file, above the cursor, but I think feeding the whole file + other open code tabs would be super useful.
I used the 1.5 B model of the qwen2.5 family for code generation in the example. It works fine, but sometimes it forgets that it's writing code, exits the markdown code block and starts writing an explanation...