Llamafile Returns

(blog.mozilla.ai)

89 points | by aittalam 2 days ago ago

14 comments

  • FragenAntworten 12 minutes ago

    The Discord link is broken, in that it links to the server directly rather than to an invitation to join the server, which prevents new members from joining.

  • thangalin 3 hours ago

    Tips:

        # Avoid issues when wine is installed.
        sudo su -c 'echo 0 > /proc/sys/fs/binfmt_misc/status'
    
    And:

        # Capture the entirety of the instructions to obtain the input length.
        readonly INSTRUCT=$(
          join ${PATH_PREFIX_SYSTEM} ${PATH_PROMPT_SYSTEM} ${PATH_PREFIX_SYSTEM}
          join ${PATH_SUFFIX_USER} ${PATH_PROMPT_USER} ${PATH_SUFFIX_USER}
          join ${PATH_SUFFIX_ASSIST} "/dev/null" ${PATH_SUFFIX_ASSIST}
        )
    
        (
          echo ${INSTRUCT}
        ) | ./llamafile \
          -m "${LINK_MODEL}" \
          -e \
          -f /dev/stdin \
          -n 1000 \
          -c ${#INSTRUCT} \
          --repeat-penalty 1.0 \
          --temp 1.5 \
          --silent-prompt > output.txt
  • michaelgiba 2 hours ago

    I’m glad to see llamafile being resurrected. A few things I hope for:

    1. Curate a continuously extended inventory of prebuilt llamafiles for models as they are released 2. Create both flexible builds (with dynamic backend loading for cpu and cuda) and slim minimalist builds 3. Upstreaming as much as they can into llama.cpp and partner with the project

    • michaelgiba 2 hours ago

      Crazier ideas would be: - extend the concept to also have some sort of “agent mode” where the llamafiles can launch with their own minimal file system or isolated context - detailed profiling of main supported models to ensure deterministic outputs

  • swyx 4 hours ago

    justine tunney gave a great intro to Llamafile at AIE last year if it helps anyone: https://www.youtube.com/watch?v=-mRi-B3t6fA

  • jart 2 days ago

    Really exciting to see Mozilla AI starting up and I can't wait to see where the next generation takes the project!

    • bsenftner 4 hours ago

      People are so uninformed, they don't know you are Mozilla AI's star employee.

      • rvz 4 hours ago

        s/are/was

        I don't know if you were informed but you realize that jart is no longer at Mozilla anymore and now at Google Inc?

  • synergy20 3 hours ago

    how is this different from ollama? for me the more/open the merrier.

    • ricardobeat 3 hours ago

      Ollama is a model manager and pretty interface for llama.cpp, llamafile is a cross-platform packaging tool to distribute and run individual models also based on llama.cpp

  • apitman 4 hours ago

    This is great news. Given the proliferation of solid local models, it would be cool if llamafile had a way to build your own custom versions with the model of your choice.

  • behindsight 6 hours ago

    great stuff, working on something around agentic tooling and hope to collab with Mozilla AI as well in the future as they share the same values I have

  • throawayonthe 4 hours ago

    go get that investor money i guess?