commit bdb5c09d1ad64dcf8a2db0266ae04f4b065defc2 from: Oliver Lowe date: Sun Sep 29 04:39:17 2024 UTC llama: try out llama 3.2 3B by default commit - d195ecc2648dc74c89e96ed15110ef3dffe19724 commit + bdb5c09d1ad64dcf8a2db0266ae04f4b065defc2 blob - 0154aacfc12328ee609fd5adb4a8e439e8c10bf3 blob + 0bc9045f55d1cbdd7e2c8e1c0432a3a965b2d586 --- bin/llama +++ bin/llama @@ -5,13 +5,13 @@ import os import sys import urllib.request -# these models hallucinate -# model = "llama3.1-8b-instant" -# model = "llama-3.1-70b-versatile" - url = "https://api.groq.com/openai/v1/chat/completions" -model = "llama3-8b-8192" -big = "llama3-70b-8192" +# small models include: +# llama-3.1-8b-instant +# llama-3.2-3b-preview +# llama-3.2-1b-preview +model = "llama-3.2-3b-preview" +big = "llama-3.1-70b-versatile" def read_token(name): with open(name) as f: @@ -28,7 +28,7 @@ message = {"messages": [{"role": "user","content": pro req = urllib.request.Request(url, json.dumps(message).encode()) req.add_header("Content-Type", "application/json") req.add_header("Authorization", "Bearer "+token) -# groq blocks urllib's user agent?! +# groq blocks urllib's user agent req.add_header("User-Agent", "curl/8.9.0") with urllib.request.urlopen(req) as resp: blob - 522c0479bd640805b911a31bca763d9286d4b450 blob + 9d2f1051da0c09ddd0921b3a3f3dd2751f0c2f5e --- man/llama.1 +++ man/llama.1 @@ -11,7 +11,7 @@ reads a prompt from the standard input and sends it to a large language model hosted by Groq. The reply is written to the standard output. -The default model is Llama 3 8B. +The default model is Llama 3.2 3B. .Pp A Groq API token must be written to .Pa $HOME/.config/groq/token . @@ -19,7 +19,7 @@ A Groq API token must be written to The following flags are understood: .Bl -tag -width Ds .It Fl b -Prompt the "bigger" 70B model. +Prompt the "bigger" LLama 3.1 70B model. .Sh EXAMPLE .Dl echo 'What is LLM slop?' | llama .Sh EXIT STATUS