feat(files): introduce upload_file tool for staging files and enhance save_file documentation
This commit is contained in:
49
README.md
49
README.md
@@ -41,7 +41,8 @@ A Go MCP server for capturing and retrieving thoughts, memory, and project conte
|
|||||||
| `recall_context` | Semantic + recency context block for injection |
|
| `recall_context` | Semantic + recency context block for injection |
|
||||||
| `link_thoughts` | Create a typed relationship between thoughts |
|
| `link_thoughts` | Create a typed relationship between thoughts |
|
||||||
| `related_thoughts` | Explicit links + semantic neighbours |
|
| `related_thoughts` | Explicit links + semantic neighbours |
|
||||||
| `save_file` | Store a file (base64 or by resource URI) and optionally link it to a thought |
|
| `upload_file` | Stage a file from a server-side path or base64 and get an `amcs://files/{id}` resource URI |
|
||||||
|
| `save_file` | Store a file (base64 or resource URI) and optionally link it to a thought |
|
||||||
| `load_file` | Retrieve a stored file by ID; returns metadata, base64 content, and an embedded MCP binary resource |
|
| `load_file` | Retrieve a stored file by ID; returns metadata, base64 content, and an embedded MCP binary resource |
|
||||||
| `list_files` | Browse stored files by thought, project, or kind |
|
| `list_files` | Browse stored files by thought, project, or kind |
|
||||||
| `backfill_embeddings` | Generate missing embeddings for stored thoughts |
|
| `backfill_embeddings` | Generate missing embeddings for stored thoughts |
|
||||||
@@ -185,7 +186,37 @@ Files can optionally be linked to a thought by passing `thought_id`, which also
|
|||||||
|
|
||||||
### MCP tools
|
### MCP tools
|
||||||
|
|
||||||
**Save via base64** (small files or when HTTP is not available):
|
**Stage a file and get a URI** (`upload_file`) — preferred for large or binary files:
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"name": "diagram.png",
|
||||||
|
"content_path": "/absolute/path/to/diagram.png"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Or with base64 for small files (≤10 MB):
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"name": "diagram.png",
|
||||||
|
"content_base64": "<base64-payload>"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Returns `{"file": {...}, "uri": "amcs://files/<id>"}`. Pass `thought_id`/`project` to link immediately, or omit them and use the URI in a later `save_file` call.
|
||||||
|
|
||||||
|
**Link a staged file to a thought** (`save_file` with `content_uri`):
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"name": "meeting-notes.pdf",
|
||||||
|
"thought_id": "optional-thought-uuid",
|
||||||
|
"content_uri": "amcs://files/<id-from-upload_file>"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Save small files inline** (`save_file` with `content_base64`, ≤10 MB):
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
@@ -197,19 +228,7 @@ Files can optionally be linked to a thought by passing `thought_id`, which also
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
**Save via resource URI** (preferred for binary; avoids base64 overhead):
|
`content_base64` and `content_uri` are mutually exclusive in both tools.
|
||||||
|
|
||||||
Upload the file binary via HTTP first (see below), then pass the returned URI to `save_file`:
|
|
||||||
|
|
||||||
```json
|
|
||||||
{
|
|
||||||
"name": "meeting-notes.pdf",
|
|
||||||
"thought_id": "optional-thought-uuid",
|
|
||||||
"content_uri": "amcs://files/<id-from-upload>"
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
`content_base64` and `content_uri` are mutually exclusive.
|
|
||||||
|
|
||||||
**Load a file** — returns metadata, base64 content, and an embedded MCP binary resource (`amcs://files/{id}`):
|
**Load a file** — returns metadata, base64 content, and an embedded MCP binary resource (`amcs://files/{id}`):
|
||||||
|
|
||||||
|
|||||||
@@ -224,7 +224,7 @@ func (c *Client) ExtractMetadata(ctx context.Context, input string) (thoughttype
|
|||||||
attrs := []any{
|
attrs := []any{
|
||||||
slog.String("provider", c.name),
|
slog.String("provider", c.name),
|
||||||
slog.String("model", model),
|
slog.String("model", model),
|
||||||
slog.Duration("duration", time.Since(start)),
|
slog.String("duration", formatLogDuration(time.Since(start))),
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
attrs = append(attrs, slog.String("error", err.Error()))
|
attrs = append(attrs, slog.String("error", err.Error()))
|
||||||
@@ -298,6 +298,18 @@ func (c *Client) ExtractMetadata(ctx context.Context, input string) (thoughttype
|
|||||||
return heuristic, nil
|
return heuristic, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func formatLogDuration(d time.Duration) string {
|
||||||
|
if d < 0 {
|
||||||
|
d = -d
|
||||||
|
}
|
||||||
|
|
||||||
|
totalMilliseconds := d.Milliseconds()
|
||||||
|
minutes := totalMilliseconds / 60000
|
||||||
|
seconds := (totalMilliseconds / 1000) % 60
|
||||||
|
milliseconds := totalMilliseconds % 1000
|
||||||
|
return fmt.Sprintf("%02d:%02d:%03d", minutes, seconds, milliseconds)
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Client) extractMetadataWithModel(ctx context.Context, input, model string) (thoughttypes.ThoughtMetadata, error) {
|
func (c *Client) extractMetadataWithModel(ctx context.Context, input, model string) (thoughttypes.ThoughtMetadata, error) {
|
||||||
if c.shouldBypassModel(model) {
|
if c.shouldBypassModel(model) {
|
||||||
return thoughttypes.ThoughtMetadata{}, fmt.Errorf("%s metadata: model %q temporarily bypassed after repeated empty responses", c.name, model)
|
return thoughttypes.ThoughtMetadata{}, fmt.Errorf("%s metadata: model %q temporarily bypassed after repeated empty responses", c.name, model)
|
||||||
|
|||||||
@@ -47,7 +47,7 @@ func logToolCall[In any, Out any](logger *slog.Logger, toolName string, handler
|
|||||||
result, out, err := handler(ctx, req, in)
|
result, out, err := handler(ctx, req, in)
|
||||||
|
|
||||||
completionAttrs := append([]any{}, attrs...)
|
completionAttrs := append([]any{}, attrs...)
|
||||||
completionAttrs = append(completionAttrs, slog.Duration("duration", time.Since(start)))
|
completionAttrs = append(completionAttrs, slog.String("duration", formatLogDuration(time.Since(start))))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
completionAttrs = append(completionAttrs, slog.String("error", err.Error()))
|
completionAttrs = append(completionAttrs, slog.String("error", err.Error()))
|
||||||
logger.Error("mcp tool completed", completionAttrs...)
|
logger.Error("mcp tool completed", completionAttrs...)
|
||||||
@@ -70,6 +70,18 @@ func truncateArgs(args any) string {
|
|||||||
return string(b[:maxLoggedArgBytes]) + fmt.Sprintf("… (%d bytes total)", len(b))
|
return string(b[:maxLoggedArgBytes]) + fmt.Sprintf("… (%d bytes total)", len(b))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func formatLogDuration(d time.Duration) string {
|
||||||
|
if d < 0 {
|
||||||
|
d = -d
|
||||||
|
}
|
||||||
|
|
||||||
|
totalMilliseconds := d.Milliseconds()
|
||||||
|
minutes := totalMilliseconds / 60000
|
||||||
|
seconds := (totalMilliseconds / 1000) % 60
|
||||||
|
milliseconds := totalMilliseconds % 1000
|
||||||
|
return fmt.Sprintf("%02d:%02d:%03d", minutes, seconds, milliseconds)
|
||||||
|
}
|
||||||
|
|
||||||
func setToolSchemas[In any, Out any](tool *mcp.Tool) error {
|
func setToolSchemas[In any, Out any](tool *mcp.Tool) error {
|
||||||
if tool.InputSchema == nil {
|
if tool.InputSchema == nil {
|
||||||
inputSchema, err := jsonschema.For[In](toolSchemaOptions)
|
inputSchema, err := jsonschema.For[In](toolSchemaOptions)
|
||||||
|
|||||||
@@ -134,9 +134,14 @@ func New(cfg config.MCPConfig, logger *slog.Logger, toolSet ToolSet) http.Handle
|
|||||||
Description: "A stored file. Read a file's raw binary content by its id. Use load_file for metadata.",
|
Description: "A stored file. Read a file's raw binary content by its id. Use load_file for metadata.",
|
||||||
}, toolSet.Files.ReadResource)
|
}, toolSet.Files.ReadResource)
|
||||||
|
|
||||||
|
addTool(server, logger, &mcp.Tool{
|
||||||
|
Name: "upload_file",
|
||||||
|
Description: "Stage a file and get an amcs://files/{id} resource URI. Provide content_path (absolute server-side path, no size limit) or content_base64 (≤10 MB). Optionally link immediately with thought_id/project, or omit them and pass the returned URI to save_file later.",
|
||||||
|
}, toolSet.Files.Upload)
|
||||||
|
|
||||||
addTool(server, logger, &mcp.Tool{
|
addTool(server, logger, &mcp.Tool{
|
||||||
Name: "save_file",
|
Name: "save_file",
|
||||||
Description: "Store a file and optionally link it to a thought. Supply either content_base64 (≤10 MB) or content_uri (amcs://files/{id} from a prior POST /files upload). For files larger than 10 MB, upload via POST /files first and pass the returned URI as content_uri.",
|
Description: "Store a file and optionally link it to a thought. Supply either content_base64 (≤10 MB) or content_uri (amcs://files/{id} from a prior upload_file or POST /files call). For files larger than 10 MB, use upload_file with content_path first.",
|
||||||
}, toolSet.Files.Save)
|
}, toolSet.Files.Save)
|
||||||
|
|
||||||
addTool(server, logger, &mcp.Tool{
|
addTool(server, logger, &mcp.Tool{
|
||||||
|
|||||||
@@ -6,6 +6,8 @@ import (
|
|||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
"encoding/hex"
|
"encoding/hex"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
@@ -65,6 +67,21 @@ type ListFilesInput struct {
|
|||||||
Kind string `json:"kind,omitempty" jsonschema:"optional kind filter such as image, document, audio, or file"`
|
Kind string `json:"kind,omitempty" jsonschema:"optional kind filter such as image, document, audio, or file"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type UploadFileInput struct {
|
||||||
|
Name string `json:"name" jsonschema:"file name including extension, for example photo.png or note.pdf"`
|
||||||
|
ContentPath string `json:"content_path,omitempty" jsonschema:"absolute path to a file on the server; preferred for large files — no base64 overhead"`
|
||||||
|
ContentBase64 string `json:"content_base64,omitempty" jsonschema:"file contents encoded as base64 (≤10 MB); use content_path for larger files"`
|
||||||
|
MediaType string `json:"media_type,omitempty" jsonschema:"optional MIME type such as image/png, application/pdf, or audio/mpeg"`
|
||||||
|
Kind string `json:"kind,omitempty" jsonschema:"optional logical type such as image, document, audio, or file"`
|
||||||
|
ThoughtID string `json:"thought_id,omitempty" jsonschema:"optional thought id to link this file to immediately"`
|
||||||
|
Project string `json:"project,omitempty" jsonschema:"optional project name or id"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type UploadFileOutput struct {
|
||||||
|
File thoughttypes.StoredFile `json:"file"`
|
||||||
|
URI string `json:"uri" jsonschema:"amcs resource URI for this file, e.g. amcs://files/{id}; pass as content_uri in save_file to link without re-uploading"`
|
||||||
|
}
|
||||||
|
|
||||||
type ListFilesOutput struct {
|
type ListFilesOutput struct {
|
||||||
Files []thoughttypes.StoredFile `json:"files"`
|
Files []thoughttypes.StoredFile `json:"files"`
|
||||||
}
|
}
|
||||||
@@ -73,6 +90,60 @@ func NewFilesTool(db *store.DB, sessions *session.ActiveProjects) *FilesTool {
|
|||||||
return &FilesTool{store: db, sessions: sessions}
|
return &FilesTool{store: db, sessions: sessions}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (t *FilesTool) Upload(ctx context.Context, req *mcp.CallToolRequest, in UploadFileInput) (*mcp.CallToolResult, UploadFileOutput, error) {
|
||||||
|
path := strings.TrimSpace(in.ContentPath)
|
||||||
|
b64 := strings.TrimSpace(in.ContentBase64)
|
||||||
|
|
||||||
|
if path != "" && b64 != "" {
|
||||||
|
return nil, UploadFileOutput{}, errInvalidInput("provide content_path or content_base64, not both")
|
||||||
|
}
|
||||||
|
|
||||||
|
var content []byte
|
||||||
|
var mediaTypeFromSource string
|
||||||
|
|
||||||
|
if path != "" {
|
||||||
|
if !filepath.IsAbs(path) {
|
||||||
|
return nil, UploadFileOutput{}, errInvalidInput("content_path must be an absolute path")
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
content, err = os.ReadFile(path)
|
||||||
|
if err != nil {
|
||||||
|
return nil, UploadFileOutput{}, errInvalidInput("cannot read content_path: " + err.Error())
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if b64 == "" {
|
||||||
|
return nil, UploadFileOutput{}, errInvalidInput("content_path or content_base64 is required")
|
||||||
|
}
|
||||||
|
if len(b64) > maxBase64ToolBytes {
|
||||||
|
return nil, UploadFileOutput{}, errInvalidInput(
|
||||||
|
"content_base64 exceeds the 10 MB MCP tool limit; use content_path instead",
|
||||||
|
)
|
||||||
|
}
|
||||||
|
raw, dataURLMediaType := splitDataURL(b64)
|
||||||
|
var err error
|
||||||
|
content, err = decodeBase64(raw)
|
||||||
|
if err != nil {
|
||||||
|
return nil, UploadFileOutput{}, errInvalidInput("content_base64 must be valid base64")
|
||||||
|
}
|
||||||
|
mediaTypeFromSource = dataURLMediaType
|
||||||
|
}
|
||||||
|
|
||||||
|
out, err := t.SaveDecoded(ctx, req, SaveFileDecodedInput{
|
||||||
|
Name: in.Name,
|
||||||
|
Content: content,
|
||||||
|
MediaType: firstNonEmpty(strings.TrimSpace(in.MediaType), mediaTypeFromSource),
|
||||||
|
Kind: in.Kind,
|
||||||
|
ThoughtID: in.ThoughtID,
|
||||||
|
Project: in.Project,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
return nil, UploadFileOutput{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
uri := fileURIPrefix + out.File.ID.String()
|
||||||
|
return nil, UploadFileOutput{File: out.File, URI: uri}, nil
|
||||||
|
}
|
||||||
|
|
||||||
func (t *FilesTool) Save(ctx context.Context, req *mcp.CallToolRequest, in SaveFileInput) (*mcp.CallToolResult, SaveFileOutput, error) {
|
func (t *FilesTool) Save(ctx context.Context, req *mcp.CallToolRequest, in SaveFileInput) (*mcp.CallToolResult, SaveFileOutput, error) {
|
||||||
uri := strings.TrimSpace(in.ContentURI)
|
uri := strings.TrimSpace(in.ContentURI)
|
||||||
b64 := strings.TrimSpace(in.ContentBase64)
|
b64 := strings.TrimSpace(in.ContentBase64)
|
||||||
@@ -338,6 +409,12 @@ func decodeBase64(value string) ([]byte, error) {
|
|||||||
}
|
}
|
||||||
}, value)
|
}, value)
|
||||||
|
|
||||||
|
var candidates []string
|
||||||
|
candidates = append(candidates, cleaned)
|
||||||
|
if trimmed := strings.TrimRight(cleaned, "="); trimmed != cleaned && trimmed != "" {
|
||||||
|
candidates = append(candidates, trimmed)
|
||||||
|
}
|
||||||
|
|
||||||
encodings := []*base64.Encoding{
|
encodings := []*base64.Encoding{
|
||||||
base64.StdEncoding,
|
base64.StdEncoding,
|
||||||
base64.RawStdEncoding,
|
base64.RawStdEncoding,
|
||||||
@@ -346,12 +423,14 @@ func decodeBase64(value string) ([]byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
var lastErr error
|
var lastErr error
|
||||||
for _, encoding := range encodings {
|
for _, candidate := range candidates {
|
||||||
decoded, err := encoding.DecodeString(cleaned)
|
for _, encoding := range encodings {
|
||||||
if err == nil {
|
decoded, err := encoding.DecodeString(candidate)
|
||||||
return decoded, nil
|
if err == nil {
|
||||||
|
return decoded, nil
|
||||||
|
}
|
||||||
|
lastErr = err
|
||||||
}
|
}
|
||||||
lastErr = err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil, lastErr
|
return nil, lastErr
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ func TestDecodeBase64AcceptsWhitespaceAndMultipleVariants(t *testing.T) {
|
|||||||
}{
|
}{
|
||||||
{name: "standard with whitespace", input: "aG V s\nbG8=", want: "hello"},
|
{name: "standard with whitespace", input: "aG V s\nbG8=", want: "hello"},
|
||||||
{name: "raw standard", input: "aGVsbG8", want: "hello"},
|
{name: "raw standard", input: "aGVsbG8", want: "hello"},
|
||||||
|
{name: "standard with extra padding", input: "aGVsbG8==", want: "hello"},
|
||||||
{name: "standard url-safe payload", input: "--8=", want: string([]byte{0xfb, 0xef})},
|
{name: "standard url-safe payload", input: "--8=", want: string([]byte{0xfb, 0xef})},
|
||||||
{name: "raw url-safe payload", input: "--8", want: string([]byte{0xfb, 0xef})},
|
{name: "raw url-safe payload", input: "--8", want: string([]byte{0xfb, 0xef})},
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -32,9 +32,10 @@ At the start of every project session, after setting the active project:
|
|||||||
- Use project memory for code decisions, architecture, TODOs, debugging findings, and context specific to the current repo or workstream.
|
- Use project memory for code decisions, architecture, TODOs, debugging findings, and context specific to the current repo or workstream.
|
||||||
- Before substantial work, always retrieve context with `get_project_context` or `recall_context` so prior decisions inform your approach.
|
- Before substantial work, always retrieve context with `get_project_context` or `recall_context` so prior decisions inform your approach.
|
||||||
- Save durable project facts with `capture_thought` after completing meaningful work.
|
- Save durable project facts with `capture_thought` after completing meaningful work.
|
||||||
- Use `save_file` for project assets the memory should retain, such as screenshots, PDFs, audio notes, and other documents.
|
- Use `save_file` or `upload_file` for project assets the memory should retain, such as screenshots, PDFs, audio notes, and other documents.
|
||||||
- If the goal is to retain the artifact itself, use `save_file` directly instead of first reading, transcribing, or summarizing the file contents.
|
- If the goal is to retain the artifact itself, store the file directly instead of first reading, transcribing, or summarizing its contents.
|
||||||
- When calling `save_file`, prefer `content_uri` over `content_base64` for binary files that were uploaded via HTTP — pass the `amcs://files/{id}` URI returned by the upload instead of re-encoding the bytes as base64.
|
- For binary files or files larger than 10 MB, call `upload_file` with `content_path` (absolute server-side path) first to get an `amcs://files/{id}` URI, then pass that URI to `save_file` as `content_uri` to link it to a thought. This avoids base64 encoding entirely.
|
||||||
|
- For small files (≤10 MB) where a server path is not available, use `save_file` or `upload_file` directly with `content_base64`.
|
||||||
- Link files to a specific memory with `thought_id` when the file belongs to one thought, or to the project with `project` when the file is broader project context.
|
- Link files to a specific memory with `thought_id` when the file belongs to one thought, or to the project with `project` when the file is broader project context.
|
||||||
- Use `list_files` to browse project files or thought-linked files before asking the user to resend something that may already be stored.
|
- Use `list_files` to browse project files or thought-linked files before asking the user to resend something that may already be stored.
|
||||||
- Use `load_file` when you need the actual stored file contents back. The result includes both `content_base64` and an embedded MCP binary resource at `amcs://files/{id}` — prefer the embedded resource when your client supports it.
|
- Use `load_file` when you need the actual stored file contents back. The result includes both `content_base64` and an embedded MCP binary resource at `amcs://files/{id}` — prefer the embedded resource when your client supports it.
|
||||||
@@ -55,7 +56,7 @@ At the start of every project session, after setting the active project:
|
|||||||
- Prefer concise summaries.
|
- Prefer concise summaries.
|
||||||
- Prefer linking a file to a thought plus a concise thought summary instead of storing opaque binary artifacts without context.
|
- Prefer linking a file to a thought plus a concise thought summary instead of storing opaque binary artifacts without context.
|
||||||
- Do not read a file just to make it storable; store the file directly and read it only when the file contents are needed for reasoning.
|
- Do not read a file just to make it storable; store the file directly and read it only when the file contents are needed for reasoning.
|
||||||
- Do not base64-encode a file to pass it to `save_file` if an `amcs://files/{id}` URI is already available from a prior HTTP upload.
|
- Do not base64-encode a file to pass it to `save_file` if an `amcs://files/{id}` URI is already available from a prior `upload_file` or HTTP upload.
|
||||||
- When saving, choose the narrowest correct scope: project if project-specific, global if not.
|
- When saving, choose the narrowest correct scope: project if project-specific, global if not.
|
||||||
|
|
||||||
## Skills and Guardrails
|
## Skills and Guardrails
|
||||||
@@ -68,4 +69,4 @@ At the start of every project session, after setting the active project:
|
|||||||
|
|
||||||
## Short Operational Form
|
## Short Operational Form
|
||||||
|
|
||||||
Use AMCS memory in project scope when the current work matches a known project. If no clear project matches, global notebook memory is allowed for non-project-specific information. At the start of every project session call `list_project_skills` and `list_project_guardrails` and apply what is returned; only create new skills or guardrails if none exist. Store durable notes with `capture_thought`, store supporting binary artifacts with `save_file`, prefer saving a file directly when the artifact itself is what matters, browse stored files with `list_files`, and load them with `load_file` only when their contents are needed. When saving a file uploaded via HTTP, pass `content_uri: "amcs://files/{id}"` instead of re-encoding it as base64. Stored files can also be read as raw binary via MCP resources at `amcs://files/{id}`. Never store project-specific memory globally when a matching project exists, and never store memory in the wrong project. If project matching is ambiguous, ask the user.
|
Use AMCS memory in project scope when the current work matches a known project. If no clear project matches, global notebook memory is allowed for non-project-specific information. At the start of every project session call `list_project_skills` and `list_project_guardrails` and apply what is returned; only create new skills or guardrails if none exist. Store durable notes with `capture_thought`. For binary files or files larger than 10 MB, call `upload_file` with `content_path` to stage the file and get an `amcs://files/{id}` URI, then pass that URI to `save_file` as `content_uri` to link it to a thought. For small files, use `save_file` or `upload_file` with `content_base64` directly. Browse stored files with `list_files`, and load them with `load_file` only when their contents are needed. Stored files can also be read as raw binary via MCP resources at `amcs://files/{id}`. Never store project-specific memory globally when a matching project exists, and never store memory in the wrong project. If project matching is ambiguous, ask the user.
|
||||||
|
|||||||
Reference in New Issue
Block a user