service-llama3.yaml
file.
In the configuration file, replace <REPLACE_WITH_YOUR_HUGGINGFACE_TOKEN>
with your HuggingFace token so that model weights are downloaded.
You’ll have to request access to LLama3 if you haven’t already.
RUNNING
, you can chat with it right from the dashboard!
The endpoint for the service is also provided in the dashboard.