diff --git a/.cursor/mcp.json b/.cursor/mcp.json deleted file mode 100644 index 439c156..0000000 --- a/.cursor/mcp.json +++ /dev/null @@ -1,12 +0,0 @@ -{ - "mcpServers": { - "task-master-ai": { - "command": "/home/geir/Home-lab/result/bin/task-master-ai", - "args": [], - "env": { - "OPENAI_API_KEY": "fake-key-for-local-ollama", - "OPENAI_BASE_URL": "http://grey-area:11434/v1" - } - } - } -} \ No newline at end of file diff --git a/.env.taskmaster b/.env.taskmaster new file mode 100644 index 0000000..4a1b37a --- /dev/null +++ b/.env.taskmaster @@ -0,0 +1,14 @@ +# Taskmaster AI Environment Variables +# Source this file to set up Ollama integration for taskmaster +export OPENAI_API_BASE="http://grey-area:11434/v1" +export OPENAI_API_KEY="ollama" +export OPENAI_BASE_URL="http://grey-area:11434/v1" +export OLLAMA_BASE_URL="http://grey-area:11434/api" +export MODEL="qwen2.5-coder:7b" +export RESEARCH_MODEL="deepseek-r1:7b" +export FALLBACK_MODEL="llama3.1:8b" +export MAX_TOKENS="8192" +export TEMPERATURE="0.3" + +echo "✅ Taskmaster AI environment variables loaded" +echo "🤖 Using Ollama models at grey-area:11434" diff --git a/.taskmaster/config.json b/.taskmaster/config.json index 199491a..ff0b382 100644 --- a/.taskmaster/config.json +++ b/.taskmaster/config.json @@ -1,20 +1,20 @@ { "models": { "main": { - "provider": "anthropic", - "modelId": "claude-3-7-sonnet-20250219", - "maxTokens": 120000, - "temperature": 0.2 + "provider": "ollama", + "modelId": "qwen2.5-coder:7b", + "maxTokens": 8192, + "temperature": 0.3 }, "research": { - "provider": "perplexity", - "modelId": "sonar-pro", - "maxTokens": 8700, + "provider": "ollama", + "modelId": "deepseek-r1:7b", + "maxTokens": 8192, "temperature": 0.1 }, "fallback": { - "provider": "anthropic", - "modelId": "claude-3-5-sonnet-20240620", + "provider": "ollama", + "modelId": "llama3.1:8b", "maxTokens": 8192, "temperature": 0.1 } @@ -25,7 +25,8 @@ "defaultSubtasks": 5, "defaultPriority": "medium", "projectName": "Taskmaster", - "ollamaBaseURL": "http://localhost:11434/api", + "ollamaBaseURL": "http://grey-area:11434/api", + "openaiBaseURL": "http://grey-area:11434/v1", "bedrockBaseURL": "https://bedrock.us-east-1.amazonaws.com", "defaultTag": "master", "azureOpenaiBaseURL": "https://your-endpoint.openai.azure.com/", diff --git a/packages/.cursor/mcp.json b/packages/.cursor/mcp.json deleted file mode 100644 index 7e49eb3..0000000 --- a/packages/.cursor/mcp.json +++ /dev/null @@ -1,23 +0,0 @@ -{ - "mcpServers": { - "task-master-ai": { - "command": "npx", - "args": [ - "-y", - "--package=task-master-ai", - "task-master-ai" - ], - "env": { - "ANTHROPIC_API_KEY": "ANTHROPIC_API_KEY_HERE", - "PERPLEXITY_API_KEY": "PERPLEXITY_API_KEY_HERE", - "OPENAI_API_KEY": "OPENAI_API_KEY_HERE", - "GOOGLE_API_KEY": "GOOGLE_API_KEY_HERE", - "XAI_API_KEY": "XAI_API_KEY_HERE", - "OPENROUTER_API_KEY": "OPENROUTER_API_KEY_HERE", - "MISTRAL_API_KEY": "MISTRAL_API_KEY_HERE", - "AZURE_OPENAI_API_KEY": "AZURE_OPENAI_API_KEY_HERE", - "OLLAMA_API_KEY": "OLLAMA_API_KEY_HERE" - } - } - } -} \ No newline at end of file