Files
gh-neolabhq-context-enginee…/plugin.lock.json
2025-11-30 08:42:59 +08:00

53 lines
1.9 KiB
JSON

{
"$schema": "internal://schemas/plugin.lock.v1.json",
"pluginId": "gh:NeoLabHQ/context-engineering-kit:plugins/reflexion",
"normalized": {
"repo": null,
"ref": "refs/tags/v20251128.0",
"commit": "b9aef3a2782f1926173706f324dc539ed4d5a86d",
"treeHash": "5cca970e49e2491c8b5bd31e7c713376fe713cb863319ff403c5490a3ca3496c",
"generatedAt": "2025-11-28T10:12:09.688815Z",
"toolVersion": "publish_plugins.py@0.2.0"
},
"origin": {
"remote": "git@github.com:zhongweili/42plugin-data.git",
"branch": "master",
"commit": "aa1497ed0949fd50e99e70d6324a29c5b34f9390",
"repoRoot": "/Users/zhongweili/projects/openmind/42plugin-data"
},
"manifest": {
"name": "reflexion",
"description": "Collection of commands that force LLM to reflect on previous response and output. Based on papers like Self-Refine and Reflexion. These techniques improve the output of large language models by introducing feedback and refinement loops.",
"version": "1.0.0"
},
"content": {
"files": [
{
"path": "README.md",
"sha256": "35e53ebb1230cd06b7f28b4169c86d01c01fdd53d34bc375870df13957c41c7f"
},
{
"path": ".claude-plugin/plugin.json",
"sha256": "ace85ada748d3c09734cfc8d848a0fb21521fbdc7212caab006a5cb7a663323b"
},
{
"path": "commands/reflect.md",
"sha256": "e8fced10a97bf6e7c7a203bd9a5e92ca13c306ed07244edb6295ccaa34d54eb6"
},
{
"path": "commands/critique.md",
"sha256": "637e99ae643865b817edf39d4d90bf69b3f033a2838506537629f2e7463b4605"
},
{
"path": "commands/memorize.md",
"sha256": "e35b3005a7bb6cccb0535ad4719a2e6f8d4d5e7754161003425d08e39cba62f9"
}
],
"dirSha256": "5cca970e49e2491c8b5bd31e7c713376fe713cb863319ff403c5490a3ca3496c"
},
"security": {
"scannedAt": null,
"scannerVersion": null,
"flags": []
}
}