{ "$schema": "internal://schemas/plugin.lock.v1.json", "pluginId": "gh:NeoLabHQ/context-engineering-kit:plugins/reflexion", "normalized": { "repo": null, "ref": "refs/tags/v20251128.0", "commit": "b9aef3a2782f1926173706f324dc539ed4d5a86d", "treeHash": "5cca970e49e2491c8b5bd31e7c713376fe713cb863319ff403c5490a3ca3496c", "generatedAt": "2025-11-28T10:12:09.688815Z", "toolVersion": "publish_plugins.py@0.2.0" }, "origin": { "remote": "git@github.com:zhongweili/42plugin-data.git", "branch": "master", "commit": "aa1497ed0949fd50e99e70d6324a29c5b34f9390", "repoRoot": "/Users/zhongweili/projects/openmind/42plugin-data" }, "manifest": { "name": "reflexion", "description": "Collection of commands that force LLM to reflect on previous response and output. Based on papers like Self-Refine and Reflexion. These techniques improve the output of large language models by introducing feedback and refinement loops.", "version": "1.0.0" }, "content": { "files": [ { "path": "README.md", "sha256": "35e53ebb1230cd06b7f28b4169c86d01c01fdd53d34bc375870df13957c41c7f" }, { "path": ".claude-plugin/plugin.json", "sha256": "ace85ada748d3c09734cfc8d848a0fb21521fbdc7212caab006a5cb7a663323b" }, { "path": "commands/reflect.md", "sha256": "e8fced10a97bf6e7c7a203bd9a5e92ca13c306ed07244edb6295ccaa34d54eb6" }, { "path": "commands/critique.md", "sha256": "637e99ae643865b817edf39d4d90bf69b3f033a2838506537629f2e7463b4605" }, { "path": "commands/memorize.md", "sha256": "e35b3005a7bb6cccb0535ad4719a2e6f8d4d5e7754161003425d08e39cba62f9" } ], "dirSha256": "5cca970e49e2491c8b5bd31e7c713376fe713cb863319ff403c5490a3ca3496c" }, "security": { "scannedAt": null, "scannerVersion": null, "flags": [] } }