WOLFcon 2024 - Understanding and Using AI Workflows with FOLIO

23 September 2024


Meta's Llama

In February 2023, Meta AI (the AI research division of the Meta that also includes Facebook) released an open-source Large Language Model (LLM) called LLaMA with a follow-up release in April 2024. Unlike other commercial LLMs, Meta AI released the weights along with supporting code to allow for training and other uses not possible with similar models released by OpenAI, Anthropic, and Google.

Using Llama

The easiest way to use Llama is go to https://www.meta.ai and start entering chat prompts. This multi-modal web application allows users to generate images and text (although to use the image generator function requires logging in with a Facebook account).

Using Llama Locally

Because of its release as open-source, you are able to download the LLaMA model to run and train locally on your computer or on cloud providers. While there are a number of ways for running LLaMA locally, a convenient method for running this and other open-source models is using a project called gpt4all. gpt4all uses LLaMA.cpp to internally run the models.

While the computing requirements can vary depending on your laptop's hardware and OS, there are desktop versions for Macintosh, Windows, and Ubuntu.

Using gpt4all

If you haven't already prior to this workshop (and if you have administrative rights on your local laptop), please download and install gpt4all.

A nice feature of gpt4all, is that you can use a locally running LLM without an internet connection. This also means that you can restrict access when using RAG on sensitive or private documentation.

Using Llama.cpp

The LLaMA.cpp project allows you to run fine-tuned LLaMA models on your local computer. LLaMA.cpp provides an OpenAI API compatible server that also allows us to integrate with DSPy and edge-ai module.

LLaMA.cpp can also be run with Docker1 on your computer if you don't want or can't compile the C++ source code to run on your computer.

Downloading a LLaMA-based Model

LLaMA.cpp uses the GGUF format for model inference and training. Look for GGUF models on HuggingFace.