UnrealGenAISupport
A plugin that integrates various cutting-edge LLM/GenAI models with Unreal Engine, focusing on game development without the complexities of LLM/GenAI integration.
Unreal Engine Generative AI Support Plugin
Usage Examples:
MCP Example:
Claude spawning scene objects and controlling their transformations and materials, generating blueprints, functions, variables, adding components, running python scripts etc.
API Example:
A project called become human, where NPCs are OpenAI agentic instances. Built using this plugin.
[!WARNING]
This plugin is still under rapid development.
- Do not use it in production environments. โ ๏ธ
- Do not use it without version control. โ ๏ธ
A stable version will be released soon. ๐๐ฅ
Every month, hundreds of new AI models are released by various organizations, making it hard to keep up with the latest advancements.
The Unreal Engine Generative AI Support Plugin allows you to focus on game development without worrying about the LLM/GenAI integration layer.
Currently integrating Model Control Protocol (MCP) with Unreal Engine 5.5.
This project aims to build a long-term support (LTS) plugin for various cutting-edge LLM/GenAI models and foster a community around it. It currently includes OpenAI's GPT-4o, Deepseek R1, Claude Sonnet 3.7 and GPT-4o-mini for Unreal Engine 5.1 or higher, with plans to add , real-time APIs, Gemini, MCP, and Grok 3 APIs soon. The plugin will focus exclusively on APIs useful for game development, evals and interactive experiences. All suggestions and contributions are welcome. The plugin can also be used for setting up new evals and ways to compare models in game battlefields.
Current Progress:
LLM/GenAI API Support:
- OpenAI API Support:
- OpenAI Chat API โ
(models-ref)
gpt-4o
,gpt-4o-mini
Model โgpt-4.5-preview
Model ๐ ๏ธo1-mini
,o1
,o1-pro
Model ๐งo3-mini
Model ๐ ๏ธ
- OpenAI DALL-E API โ (Until new generation models are released)
- OpenAI Vision API ๐ง
- OpenAI Realtime API ๐ ๏ธ
gpt-4o-realtime-preview
gpt-4o-mini-realtime-preview
Model ๐ ๏ธ
- OpenAI Structured Outputs โ
- OpenAI Whisper API ๐ง
- OpenAI Chat API โ
(models-ref)
- Anthropic Claude API Support:
- Claude Chat API โ
claude-3-7-sonnet-latest
Model โclaude-3-5-sonnet
Model โclaude-3-5-haiku-latest
Model โclaude-3-opus-latest
Model โ
- Claude Vision API ๐ง
- Claude Chat API โ
- XAI (Grok 3) API Support:
- XAI Chat Completions API ๐ง
grok-beta
Model ๐งgrok-beta
Streaming API ๐ง
- XAI Image API ๐ง
- XAI Chat Completions API ๐ง
- Google Gemini API Support:
- Gemini Chat API ๐ง๐ค
gemini-2.0-flash-lite
,gemini-2.0-flash
gemini-1.5-flash
Model ๐ง๐ค
- Gemini Imagen API: ๐ง
imagen-3.0-generate-002
Model ๐ง
- Gemini Chat API ๐ง๐ค
- Meta AI API Support:
- Llama Chat API โ (Until new generation models are released)
llama3.3-70b
Model โllama3.1-8b
Model โ
- Local Llama API ๐ง๐ค
- Llama Chat API โ (Until new generation models are released)
- Deepseek API Support:
- Deepseek Chat API โ
deepseek-chat
(DeepSeek-V3) Model โ
- Deepseek Reasoning API, R1 โ
deepseek-reasoning-r1
Model โdeepseek-reasoning-r1
CoT Streaming โ
- Independently Hosted Deepseek Models ๐ง
- Deepseek Chat API โ
- Baidu API Support:
- Baidu Chat API ๐ง
baidu-chat
Model ๐ง
- Baidu Chat API ๐ง
- 3D generative model APIs:
- TripoSR by StabilityAI ๐ง
- Plugin Documentation ๐ ๏ธ๐ค
- Plugin Example Project ๐ ๏ธ here
- Version Control Support
- Perforce Support ๐ง
- Git Submodule Support โ
- LTS Branching ๐ง
- Stable Branch with Bug Fixes ๐ง
- Dedicated Contributor for LTS ๐ง
- Lightweight Plugin (In Builds)
- No External Dependencies โ
- Build Flags to enable/disable APIs ๐ง
- Submodules per API Organization ๐ง
- Exclude MCP from build ๐ง
- Testing
- Automated Testing ๐ง
- Different Platforms ๐ง๐ค
- Different Engine Versions ๐ง๐ค
Unreal MCP (Model Control Protocol):
- Clients Support โ
- Claude Desktop App Support โ
- Cursor IDE Support โ
- OpenAI Operator API Support ๐ง
- Blueprints Auto Generation ๐ ๏ธ
- Creating new blueprint of types โ
- Adding new functions, function/blueprint variables โ
- Adding nodes and connections ๐ ๏ธ (buggy)
- Advanced Blueprints Generation ๐ ๏ธ
- Level/Scene Control for LLMs ๐ ๏ธ
- Spawning Objects and Shapes โ
- Moving, rotating and scaling objects โ
- Changing materials and color โ
- Advanced scene features ๐ ๏ธ
- Generative AI:
- Prompt to 3D model fetch and spawn ๐ ๏ธ
- Control:
- Ability to run Python scripts โ
- Ability to run Console Commands โ
- UI:
- Widgets generation ๐ ๏ธ
- UI Blueprint generation ๐ ๏ธ
- Project Files:
- Create/Edit project files/folders ๏ธโ
- Delete existing project files โ
- Others:
- Project Cleanup ๐ ๏ธ
Where,
- โ - Completed
- ๐ ๏ธ - In Progress
- ๐ง - Planned
- ๐ค - Need Contributors
- โ - Won't Support For Now
Table of Contents
- Setting API Keys
- Setting up MCP
- Adding the plugin to your project
- Fetching the Latest Plugin Changes
- Usage
- Known Issues
- Contribution Guidelines
- References
Setting API Keys:
[!NOTE]
There is no need to set the API key for testing the MCP features in Claude app. Anthropic key only needed for Claude API.
For Editor:
Set the environment variable PS_<ORGNAME>
to your API key.
For Windows:
setx PS_<ORGNAME> "your api key"
For Linux/MacOS:
-
Run the following command in your terminal, replacing yourkey with your API key.
echo "export PS_<ORGNAME>='yourkey'" >> ~/.zshrc
-
Update the shell with the new variable:
source ~/.zshrc
PS: Don't forget to restart the Editor and ALSO the connected IDE after setting the environment variable.
Where <ORGNAME>
can be:
PS_OPENAIAPIKEY
, PS_DEEPSEEKAPIKEY
, PS_ANTHROPICAPIKEY
, PS_METAAPIKEY
, PS_GOOGLEAPIKEY
etc.
For Packaged Builds:
Storing API keys in packaged builds is a security risk. This is what the OpenAI API documentation says about it:
"Exposing your OpenAI API key in client-side environments like browsers or mobile apps allows malicious users to take that key and make requests on your behalf โ which may lead to unexpected charges or compromise of certain account data. Requests should always be routed through your own backend server where you can keep your API key secure."
Read more about it here.
For test builds you can call the GenSecureKey::SetGenAIApiKeyRuntime
either in c++ or blueprints function with your API key in the packaged build.
Setting up MCP:
[!NOTE]
If your project only uses the LLM APIs and not the MCP, you can skip this section.
[!CAUTION]
Discalimer: If you are using the MCP feature of the plugin, it will directly let the Claude Desktop App control your Unreal Engine project. Make sure you are aware of the security risks and only use it in a controlled environment.Please backup your project before using the MCP feature and use version control to track changes.
1. Install any one of the below clients:
2. Setup the mcp config json:
For Claude Desktop App:
claude_desktop_config.json
file in Claude Desktop App's installation directory. (might ask claude where its located for your platform!)
The file will look something like this:
{
"mcpServers": {
"unreal-handshake": {
"command": "python",
"args": ["<your_project_directoy_path>/Plugins/GenerativeAISupport/Content/Python/mcp_server.py"],
"env": {
"UNREAL_HOST": "localhost",
"UNREAL_PORT": "9877"
}
}
}
}
For Cursor IDE:
.cursor/mcp.json
file in your project directory. The file will look something like this:
{
"mcpServers": {
"unreal-handshake": {
"command": "python",
"args": ["<your_project_directoy_path>/Plugins/GenerativeAISupport/Content/Python/mcp_server.py"],
"env": {
"UNREAL_HOST": "localhost",
"UNREAL_PORT": "9877"
}
}
}
}
3. Install MCP[CLI] from with either pip or cv.
pip install mcp[cli]
4. Enable python plugin in Unreal Engine. (Edit -> Plugins -> Python Editor Script Plugin)
Adding the plugin to your project:
With Git:
-
Add the Plugin Repository as a Submodule in your project's repository.
git submodule add https://github.com/prajwalshettydev/UnrealGenAISupport Plugins/GenerativeAISupport
-
Regenerate Project Files: Right-click your .uproject file and select Generate Visual Studio project files.
-
Enable the Plugin in Unreal Editor: Open your project in Unreal Editor. Go to Edit > Plugins. Search for the Plugin in the list and enable it.
-
For Unreal C++ Projects, include the Plugin's module in your project's Build.cs file:
PrivateDependencyModuleNames.AddRange(new string[] { "GenerativeAISupport" });
With Perforce:
Still in development..
With Unreal Marketplace:
Coming soon, for free, in the Unreal Engine Marketplace.
Fetching the Latest Plugin Changes:
With Git:
you can pull the latest changes with:
cd Plugins/GenerativeAISupport
git pull origin main
Or update all submodules in the project:
git submodule update --recursive --remote
With Perforce:
Still in development..
Usage:
There is a example Unreal project that already implements the plugin. You can find it here.
OpenAI:
Currently the plugin supports Chat and Structured Outputs from OpenAI API. Both for C++ and Blueprints.
Tested models are gpt-4o
, gpt-4o-mini
, gpt-4.5
, o1-mini
, o1
, o3-mini-high
.
1. Chat:
C++ Example:
void SomeDebugSubsystem::CallGPT(const FString& Prompt,
const TFunction<void(const FString&, const FString&, bool)>& Callback)
{
FGenChatSettings ChatSettings;
ChatSettings.Model = TEXT("gpt-4o-mini");
ChatSettings.MaxTokens = 500;
ChatSettings.Messages.Add(FGenChatMessage{ TEXT("system"), Prompt });
FOnChatCompletionResponse OnComplete = FOnChatCompletionResponse::CreateLambda(
[Callback](const FString& Response, const FString& ErrorMessage, bool bSuccess)
{
Callback(Response, ErrorMessage, bSuccess);
});
UGenOAIChat::SendChatRequest(ChatSettings, OnComplete);
}
Blueprint Example:
2. Structured Outputs:
C++ Example 1:
Sending a custom schema json directly to function call
FString MySchemaJson = R"({
"type": "object",
"properties": {
"count": {
"type": "integer",
"description": "The total number of users."
},
"users": {
"type": "array",
"items": {
"type": "object",
"properties": {
"name": { "type": "string", "description": "The user's name." },
"heading_to": { "type": "string", "description": "The user's destination." }
},
"required": ["name", "role", "age", "heading_to"]
}
}
},
"required": ["count", "users"]
})";
UGenAISchemaService::RequestStructuredOutput(
TEXT("Generate a list of users and their details"),
MySchemaJson,
[](const FString& Response, const FString& Error, bool Success) {
if (Success)
{
UE_LOG(LogTemp, Log, TEXT("Structured Output: %s"), *Response);
}
else
{
UE_LOG(LogTemp, Error, TEXT("Error: %s"), *Error);
}
}
);
C++ Example 2:
Sending a custom schema json from a file
#include "Misc/FileHelper.h"
#include "Misc/Paths.h"
FString SchemaFilePath = FPaths::Combine(
FPaths::ProjectDir(),
TEXT("Source/:ProjectName/Public/AIPrompts/SomeSchema.json")
);
FString MySchemaJson;
if (FFileHelper::LoadFileToString(MySchemaJson, *SchemaFilePath))
{
UGenAISchemaService::RequestStructuredOutput(
TEXT("Generate a list of users and their details"),
MySchemaJson,
[](const FString& Response, const FString& Error, bool Success) {
if (Success)
{
UE_LOG(LogTemp, Log, TEXT("Structured Output: %s"), *Response);
}
else
{
UE_LOG(LogTemp, Error, TEXT("Error: %s"), *Error);
}
}
);
}
Blueprint Example:
DeepSeek API:
Currently the plugin supports Chat and Reasoning from DeepSeek API. Both for C++ and Blueprints. Points to note:
- System messages are currently mandatory for the reasoning model. API otherwise seems to return null
- Also, from the documentation: "Please note that if the reasoning_content field is included in the sequence of input messages, the API will return a 400 error. Read more about it here"
[!WARNING]
While using the R1 reasoning model, make sure the Unreal's HTTP timeouts are not the default values at 30 seconds. As these API calls can take longer than 30 seconds to respond. Simply setting theHttpRequest->SetTimeout(<N Seconds>);
is not enough So the following lines need to be added to your project'sDefaultEngine.ini
file:[HTTP] HttpConnectionTimeout=180 HttpReceiveTimeout=180
1. Chat and Reasoning:
C++ Example:
FGenDSeekChatSettings ReasoningSettings;
ReasoningSettings.Model = EDeepSeekModels::Reasoner; // or EDeepSeekModels::Chat for Chat API
ReasoningSettings.MaxTokens = 100;
ReasoningSettings.Messages.Add(FGenChatMessage{TEXT("system"), TEXT("You are a helpful assistant.")});
ReasoningSettings.Messages.Add(FGenChatMessage{TEXT("user"), TEXT("9.11 and 9.8, which is greater?")});
ReasoningSettings.bStreamResponse = false;
UGenDSeekChat::SendChatRequest(
ReasoningSettings,
FOnDSeekChatCompletionResponse::CreateLambda(
[this](const FString& Response, const FString& ErrorMessage, bool bSuccess)
{
if (!UTHelper::IsContextStillValid(this))
{
return;
}
// Log response details regardless of success
UE_LOG(LogTemp, Warning, TEXT("DeepSeek Reasoning Response Received - Success: %d"), bSuccess);
UE_LOG(LogTemp, Warning, TEXT("Response: %s"), *Response);
if (!ErrorMessage.IsEmpty())
{
UE_LOG(LogTemp, Error, TEXT("Error Message: %s"), *ErrorMessage);
}
})
);
Blueprint Example:
Anthropic API:
Currently the plugin supports Chat from Anthropic API. Both for C++ and Blueprints.
Tested models are claude-3-7-sonnet-latest
, claude-3-5-sonnet
, claude-3-5-haiku-latest
, claude-3-opus-latest
.
1. Chat:
C++ Example:
// ---- Claude Chat Test ----
FGenClaudeChatSettings ChatSettings;
ChatSettings.Model = EClaudeModels::Claude_3_7_Sonnet; // Use Claude 3.7 Sonnet model
ChatSettings.MaxTokens = 4096;
ChatSettings.Temperature = 0.7f;
ChatSettings.Messages.Add(FGenChatMessage{TEXT("system"), TEXT("You are a helpful assistant.")});
ChatSettings.Messages.Add(FGenChatMessage{TEXT("user"), TEXT("What is the capital of France?")});
UGenClaudeChat::SendChatRequest(
ChatSettings,
FOnClaudeChatCompletionResponse::CreateLambda(
[this](const FString& Response, const FString& ErrorMessage, bool bSuccess)
{
if (!UTHelper::IsContextStillValid(this))
{
return;
}
if (bSuccess)
{
UE_LOG(LogTemp, Warning, TEXT("Claude Chat Response: %s"), *Response);
}
else
{
UE_LOG(LogTemp, Error, TEXT("Claude Chat Error: %s"), *ErrorMessage);
}
})
);
Blueprint Example:
Model Control Protocol (MCP):
This is currently work in progress. The plugin supports various clients like Claude Desktop App, Cursor etc.
Usage:
Running the MCP server:
1. Run the MCP server from the plugin's python directory.
python <your_project_directoy>/Plugins/GenerativeAISupport/Content/Python/mcp_server.py
2. Run the MCP client by opening or restarting the Claude desktop app or Cursor IDE.
3. Open a new Unreal Engine project and run the below python script from the plugin's python directory.
Tools -> Run Python Script -> Select the
Plugins/GenerativeAISupport/Content/Python/unreal_socket_server.py
file.
4. Now you should be able to prompt the Claude Desktop App to use Unreal Engine.
Known Issues:
- Nodes fail to connect properly with MCP
- No undo redo support for MCP
- No streaming support for Deepseek reasoning model
- No complex material generation support for the create material tool
- Issues with running some llm generated valid python scripts
- When LLM compiles a blueprint no proper error handling in its response
- Issues spawning certain nodes, especially with getters and setters
- Doesn't open the right context window during scene and project files edit.
- Doesn't dock the window properly in the editor for blueprints.
Contribution Guidelines:
Setting up for Development:
- Install
unreal
python package and setup the IDE's python interpreter for proper intellisense.
pip install unreal
More details will be added soon.
Project Structure:
More details will be added soon.
References:
- Env Var set logic from: OpenAI-Api-Unreal by KellanM
- MCP Server inspiration from: Blender-MCP by ahujasid