vault backup: 2024-10-29 16:17:59
This commit is contained in:
23
.obsidian/workspace.json
vendored
23
.obsidian/workspace.json
vendored
@@ -13,7 +13,7 @@
|
|||||||
"state": {
|
"state": {
|
||||||
"type": "markdown",
|
"type": "markdown",
|
||||||
"state": {
|
"state": {
|
||||||
"file": "Paper/CLIP/未命名.md",
|
"file": "Paper/CLIP/Cross-Modal Few-Shot Learning with Multimodal Models.md",
|
||||||
"mode": "source",
|
"mode": "source",
|
||||||
"source": false
|
"source": false
|
||||||
}
|
}
|
||||||
@@ -57,14 +57,6 @@
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "c8c6dfe89e01b54d",
|
|
||||||
"type": "leaf",
|
|
||||||
"state": {
|
|
||||||
"type": "starred",
|
|
||||||
"state": {}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "467ca686d8cb4c95",
|
"id": "467ca686d8cb4c95",
|
||||||
"type": "leaf",
|
"type": "leaf",
|
||||||
@@ -73,8 +65,7 @@
|
|||||||
"state": {}
|
"state": {}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
],
|
]
|
||||||
"currentTab": 2
|
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
"direction": "horizontal",
|
"direction": "horizontal",
|
||||||
@@ -94,7 +85,7 @@
|
|||||||
"state": {
|
"state": {
|
||||||
"type": "backlink",
|
"type": "backlink",
|
||||||
"state": {
|
"state": {
|
||||||
"file": "Paper/CLIP/未命名.md",
|
"file": "Paper/CLIP/Cross-Modal Few-Shot Learning with Multimodal Models.md",
|
||||||
"collapseAll": false,
|
"collapseAll": false,
|
||||||
"extraContext": false,
|
"extraContext": false,
|
||||||
"sortOrder": "alphabetical",
|
"sortOrder": "alphabetical",
|
||||||
@@ -111,7 +102,7 @@
|
|||||||
"state": {
|
"state": {
|
||||||
"type": "outgoing-link",
|
"type": "outgoing-link",
|
||||||
"state": {
|
"state": {
|
||||||
"file": "Paper/CLIP/未命名.md",
|
"file": "Paper/CLIP/Cross-Modal Few-Shot Learning with Multimodal Models.md",
|
||||||
"linksCollapsed": false,
|
"linksCollapsed": false,
|
||||||
"unlinkedCollapsed": true
|
"unlinkedCollapsed": true
|
||||||
}
|
}
|
||||||
@@ -134,7 +125,7 @@
|
|||||||
"state": {
|
"state": {
|
||||||
"type": "outline",
|
"type": "outline",
|
||||||
"state": {
|
"state": {
|
||||||
"file": "Paper/CLIP/未命名.md"
|
"file": "Paper/CLIP/Cross-Modal Few-Shot Learning with Multimodal Models.md"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -166,9 +157,9 @@
|
|||||||
"notion-like-tables:Create loom": false
|
"notion-like-tables:Create loom": false
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"active": "c8c6dfe89e01b54d",
|
"active": "e144afbc26630891",
|
||||||
"lastOpenFiles": [
|
"lastOpenFiles": [
|
||||||
"Paper/CLIP/未命名.md",
|
"Paper/CLIP/Cross-Modal Few-Shot Learning with Multimodal Models.md",
|
||||||
"Paper/CLIP/PromptSRC:Foundational Model Adaptation without Forgetting.md",
|
"Paper/CLIP/PromptSRC:Foundational Model Adaptation without Forgetting.md",
|
||||||
"Paper/CLIP/MaPLe:Multi-modal Prompt Learning.md",
|
"Paper/CLIP/MaPLe:Multi-modal Prompt Learning.md",
|
||||||
"Paper/CLIP/Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models.md",
|
"Paper/CLIP/Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models.md",
|
||||||
|
|||||||
Reference in New Issue
Block a user