Example (More will follow):
This panel consists of a mesh selector, which selects the object the graph generator will be applied to, and a small chatbox for communication. It includes a clear cache button to delete the chat history and a Generate Nodes button that creates the graph using generated Python code.
Create a python env and activate it:
python3 -m venv .venv
source .venv/bin/activate
To instal packages mentioned in .toml file:
poetry install
Generate Blender Shader and Geometry Node graphs directly from natural language.
Blender Node-CoPilot is an open-source Blender add-on designed to dramatically accelerate your 3D modeling and animation workflow. It allows you to describe the node setup you want in plain English, instantly generating the corresponding Shader or Geometry Node graph inside Blender.
Blender’s Shader and Geometry Node Editors are incredibly powerful, but their complexity can create a steep learning curve and slow down the prototyping phase, even for experienced artists. Manually building or translating textual node descriptions into a functional graph is time-consuming and tedious.
Node-CoPilot bridges the gap between creative intent and technical implementation:
- Natural Language Input: Simply type your desired outcome into the add-on's text field (e.g., "Create a procedural brick texture with moss on the crevices").
- AI Compilation: An integrated language model translates your instruction into a structured node graph.
- Instant Generation: The add-on uses Blender's API to automatically build the full node setup directly in your editor.
This process makes prototyping faster and serves as a powerful learning tool for newcomers.
The add-on is currently being extended to offer more functionality:
- Graph Interpretation: Ask the model to explain how an existing node graph works.
- Modification Requests: Request changes to a selected node group using natural language.
To ensure the highest quality and most artist-friendly outputs, the core compiler is a fine-tuned Qwen Coder model. I am applying Reinforcement Learning from Human Feedback (RLHF) with PPO to align the model's outputs with the natural language and problem-solving styles of 3D artists.
Diagram adapted from InstructGPT.
This project combines advanced LLM fine-tuning and Reinforcement Learning with practical Blender API integration, resulting in a tool that lowers the barrier to entry for new artists while providing a significant productivity boost for veterans.
Papers to look into:
@misc{ouyang2022traininglanguagemodelsfollow,
title={Training language models to follow instructions with human feedback},
author={Long Ouyang and Jeff Wu and Xu Jiang and Diogo Almeida and Carroll L. Wainwright and Pamela Mishkin and Chong Zhang and Sandhini Agarwal and Katarina Slama and Alex Ray and John Schulman and Jacob Hilton and Fraser Kelton and Luke Miller and Maddie Simens and Amanda Askell and Peter Welinder and Paul Christiano and Jan Leike and Ryan Lowe},
year={2022},
eprint={2203.02155},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2203.02155},
}@misc{ye202560datapointssufficient,
title={60 Data Points are Sufficient to Fine-Tune LLMs for Question-Answering},
author={Junjie Ye and Yuming Yang and Qi Zhang and Tao Gui and Xuanjing Huang and Peng Wang and Zhongchao Shi and Jianping Fan},
year={2025},
eprint={2409.15825},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2409.15825},
}@misc{vieira2024datadatafinetuninglarge,
title={How Much Data is Enough Data? Fine-Tuning Large Language Models for In-House Translation: Performance Evaluation Across Multiple Dataset Sizes},
author={Inacio Vieira and Will Allred and Séamus Lankford and Sheila Castilho and Andy Way},
year={2024},
eprint={2409.03454},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2409.03454},
}
