vault backup: 2024-11-04 18:52:44
This commit is contained in:
commit
fb6a620724
11 changed files with 177 additions and 45 deletions
4
.obsidian/plugins/companion/data.json
vendored
4
.obsidian/plugins/companion/data.json
vendored
|
@ -13,12 +13,16 @@
|
|||
"min_display_length": 50,
|
||||
"retrigger_threshold": 48
|
||||
},
|
||||
<<<<<<< HEAD
|
||||
"provider_settings": {
|
||||
"ollama": {
|
||||
"settings": "{\"endpoint\":\"http://192.168.188.43:11434\"}",
|
||||
"models": {}
|
||||
}
|
||||
},
|
||||
=======
|
||||
"provider_settings": {},
|
||||
>>>>>>> origin/main
|
||||
"presets": [],
|
||||
"fallback": null
|
||||
}
|
4
.obsidian/plugins/local-gpt/data.json
vendored
4
.obsidian/plugins/local-gpt/data.json
vendored
|
@ -2,7 +2,7 @@
|
|||
"providers": {
|
||||
"ollama": {
|
||||
"ollamaUrl": "http://localhost:11434",
|
||||
"defaultModel": "llama3.1:8b-instruct-q8_0",
|
||||
"defaultModel": "qwen2.5:14b",
|
||||
"embeddingModel": "nomic-embed-text",
|
||||
"type": "ollama"
|
||||
},
|
||||
|
@ -28,7 +28,7 @@
|
|||
"defaults": {
|
||||
"provider": "ollama",
|
||||
"fallbackProvider": "",
|
||||
"creativity": "low"
|
||||
"creativity": ""
|
||||
},
|
||||
"actions": [
|
||||
{
|
||||
|
|
122
.obsidian/plugins/smart-second-brain/data.json
vendored
122
.obsidian/plugins/smart-second-brain/data.json
vendored
|
@ -5,10 +5,10 @@
|
|||
"initialAssistantMessageContent": "Hello, I am your assistant. How can I help you?",
|
||||
"isIncognitoMode": true,
|
||||
"ollamaGenModel": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"model": "qwen2.5:14b",
|
||||
"baseUrl": "http://localhost:11434",
|
||||
"temperature": 0.5,
|
||||
"contextWindow": 1948,
|
||||
"contextWindow": 2048,
|
||||
"lcModel": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
|
@ -22,14 +22,126 @@
|
|||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0.5,
|
||||
"context_window": 1948
|
||||
"context_window": 1248,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0.5,
|
||||
"context_window": 1348,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 1,
|
||||
"context_window": 1448,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0,
|
||||
"context_window": 1548,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0,
|
||||
"context_window": 1648,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0.5,
|
||||
"context_window": 1748,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0.5,
|
||||
"context_window": 1848,
|
||||
"lc_model": {
|
||||
"lc": 1,
|
||||
"type": "constructor",
|
||||
"id": [
|
||||
"langchain",
|
||||
"chat_models",
|
||||
"ollama",
|
||||
"ChatOllama"
|
||||
],
|
||||
"kwargs": {
|
||||
"model": "llama3.1:8b-instruct-q8_0",
|
||||
"base_url": "http://localhost:11434",
|
||||
"temperature": 0.5,
|
||||
"context_window": 1948
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
"ollamaEmbedModel": {
|
||||
"model": "nomic-embed-text",
|
||||
"baseUrl": "http://localhost:11434",
|
||||
"similarityThreshold": 0.75,
|
||||
"similarityThreshold": 0.5,
|
||||
"k": 100
|
||||
},
|
||||
"openAIGenModel": {
|
||||
|
@ -49,7 +161,7 @@
|
|||
"Chats",
|
||||
"*.excalidraw.md"
|
||||
],
|
||||
"isQuickSettingsOpen": true,
|
||||
"isQuickSettingsOpen": false,
|
||||
"isVerbose": false,
|
||||
"isOnboarded": true,
|
||||
"hideIncognitoWarning": false,
|
||||
|
|
Binary file not shown.
Loading…
Add table
Add a link
Reference in a new issue