I’m interested in automatically generating lengthy, coherent stories of 10,000+ words from a single prompt using an open source local large language model (LLM) on low-spec hardware like a laptop without GPU and with i5-8250U, 16GB DDR4-2400MHz. I came across the “Awesome-Story-Generation” repository which lists relevant papers describing promising methods like “Re3: Generating Longer Stories With Recursive Reprompting and Revision”, announced in this Twitter thread from October 2022 and “DOC: Improving Long Story Coherence With Detailed Outline Control”, announced in this Twitter thread from December 2022. However, these papers used GPT-3, and I was hoping to find similar techniques implemented with open source tools that I could run locally. If anyone has experience or knows of resources that could help me achieve long, coherent story generation with an open source LLM on low-spec hardware, I would greatly appreciate any advice or guidance.

  • Possibly linux
    link
    fedilink
    English
    148 months ago

    You aren’t going to get a response that long. That is just the limitations of LLM’s. If you do manage to get something that long it won’t make sense as it can’t hold enough context as it generates.

  • Julian
    link
    fedilink
    English
    108 months ago

    You can get a really cool, coherent story of any length you want by writing one or hiring a writer.

    • I’ve won NaNoWriMo twice and I can confirm that writing your own does not necessarily result in a cool or coherent story. One of the two is likely better than an LLM could come up with, though.

      • The Bard in Green
        link
        fedilink
        1
        edit-2
        8 months ago

        I’ve written more than enough words to win, while failing to finish my story. I’ve also played a lot with local LLMs. Can confirm on all counts.

  • @d416@lemmy.world
    link
    fedilink
    88 months ago

    The limited context lengths for local LLMs will be a barrier to write 10k words in a single prompt. Approaches to this is to have the LLM have a conversation with itself or other LLMs. There are prompts out there that can simulate this, but you will need to intervene every few hundred words or so. Check out ‘AutoGen’ frameworks that can orchestrate this for you. CrewAI is one of the better ones. hope this helps

  • Irdial
    link
    English
    48 months ago

    Ollama provides a Python API which may be useful. You could have it generate the story in chunks, having it generate a list of key points which get passed to subsequent prompts. Maybe…