Skip to content

Latest commit

 

History

History
42 lines (25 loc) · 1.54 KB

File metadata and controls

42 lines (25 loc) · 1.54 KB

Starting llamaindex container

jetson-containers run $(autotag llama-index:samples)

This will start the ollama server as well as Jupyter Lab server inside the container.

Running a RAG example with Ollama

This is based on the official tutorial for local models.

Jupyter Notebook Version

When you run start the llama-index container, you should see lines like this on the terminal.

JupyterLab URL:   http://192.168.1.10:8888 (password "nvidia")
JupyterLab logs:  /data/logs/jupyter.log

On your Jetson desktop GUI, or on a PC on the same network as Jetson, open your web browser and access the address. When prompted, type the password nvidia and log in.

Jupyter Lab UI should show up, with LlamaIndex_Local-Models.ipynb listed in the left navigator pane - open it, and follow the guide in the Jupyter notebook.

Python Version

After starting the llamaindex container, you should be on root@<hostname> console. First, download the Llama2 model using ollama

ollama pull llama2

This downloads the default 7-billion parameter Llama2 model - you can optionally specify ollma2:13b and ollma2:70b for other variations, and change the Python script (line 13) accordingly. Then type the following to start the sample Python script:

python3 /opt/llama-index/llamaindex_starter.py