commit 7a579223d798a09846387b8f947b10603c69e553 from: Oliver Lowe date: Sat Dec 07 07:33:47 2024 UTC llama: bump to Llama 3.3 70B commit - a3fa2e888f4f864d7d31fbacd4d5b154ec9805c3 commit + 7a579223d798a09846387b8f947b10603c69e553 blob - 0bc9045f55d1cbdd7e2c8e1c0432a3a965b2d586 blob + 9fae067a5a3b05a4f91134370de26f2662166a6d --- bin/llama +++ bin/llama @@ -10,8 +10,8 @@ url = "https://api.groq.com/openai/v1/chat/completions # llama-3.1-8b-instant # llama-3.2-3b-preview # llama-3.2-1b-preview -model = "llama-3.2-3b-preview" -big = "llama-3.1-70b-versatile" +model = "llama-3.1-8b-instant" +big = "llama-3.3-70b-versatile" def read_token(name): with open(name) as f: blob - 9d2f1051da0c09ddd0921b3a3f3dd2751f0c2f5e blob + 354b1af9e311030085c2c02107b8b5e0a80793e1 --- man/llama.1 +++ man/llama.1 @@ -11,7 +11,7 @@ reads a prompt from the standard input and sends it to a large language model hosted by Groq. The reply is written to the standard output. -The default model is Llama 3.2 3B. +The default model is Llama 3.1 8B. .Pp A Groq API token must be written to .Pa $HOME/.config/groq/token . @@ -19,7 +19,7 @@ A Groq API token must be written to The following flags are understood: .Bl -tag -width Ds .It Fl b -Prompt the "bigger" LLama 3.1 70B model. +Prompt the "bigger" Llama 3.3 70B model. .Sh EXAMPLE .Dl echo 'What is LLM slop?' | llama .Sh EXIT STATUS