• lmr0x61@lemmy.ml
    link
    fedilink
    English
    arrow-up
    50
    ·
    11 days ago

    Normally, I’d be reading about NPM security breaches and AI security breaches separately, but now I can get them in the same article! Truly amazing how technology has progressed.

  • NocturnalMorning@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    ·
    11 days ago

    By 4:23 am ET, Chaofan Shou (@Fried_rice), an intern at Solayer Labs, broadcasted the discovery on X (formerly Twitter).

    Ha, by an intern

  • spez@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    4
    ·
    edit-2
    9 days ago

    I mean it’s not that big a deal. However, it would another thing if the model itself leaked. Now that would be something.

    edit: Like I thought, it turns out to be a TS wrapper with more internal prompts. The fireship video is really funny, they use regex to detect if the user is angry 😭

  • pelespirit@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    2
    ·
    11 days ago

    Like a healthy brain. And just like a healthy brain, it’ll still hallucinate and make mistakes probably:

    The leaked source reveals a sophisticated, three-layer memory architecture that moves away from traditional “store-everything” retrieval.

    As analyzed by developers like @himanshustwts, the architecture utilizes a “Self-Healing Memory” system.

  • Fmstrat@lemmy.world
    link
    fedilink
    English
    arrow-up
    15
    ·
    10 days ago

    At its core is MEMORY.md, a lightweight index of pointers (~150 characters per line) that is perpetually loaded into the context. This index does not store data; it stores locations.

    Actual project knowledge is distributed across “topic files” fetched on-demand, while raw transcripts are never fully read back into the context, but merely “grep’d” for specific identifiers.

    This “Strict Write Discipline”—where the agent must update its index only after a successful file write—prevents the model from polluting its context with failed attempts.

    For competitors, the “blueprint” is clear: build a skeptical memory. The code confirms that Anthropic’s agents are instructed to treat their own memory as a “hint,” requiring the model to verify facts against the actual codebase before proceeding.

    Interesting to see if continue.dev takes advantage of this methodology. My only complaint has been context with it.

  • WhyJiffie@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    13
    ·
    10 days ago

    In this mode, the agent performs “memory consolidation” while the user is idle. The autoDream logic merges disparate observations, removes logical contradictions, and converts vague insights into absolute facts.

    this blog post reads like a marketing piece

  • Dentzy@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    7
    ·
    10 days ago

    I was like “Ha, ha nice April’s fools”… Then I keep reading the comments and… WTF‽

  • UnderpantsWeevil@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 days ago

    Actual project knowledge is distributed across “topic files” fetched on-demand, while raw transcripts are never fully read back into the context, but merely “grep’d” for specific identifiers.

    Consistent with a lot of bugs and goofs I’ve heard people in long running instance of Claude will encounter.