I created an Ollama model from a Modelfile and updated the System prompt to mimic Jarvis from Iron Man (which is also named Jarvis). I'm currently importing Ollama into a python file, but I don't know how to run my code and have it use the Jarvis model I created.
This is all I have right now:
from ollama import chat
stream = chat(
model='llama3.2',
messages=[{'role': 'user', 'content': 'What is my favorite color?'}],
stream=True,
)
for chunk in stream:
print(chunk['message']['content'], end='', flush=True)
I just started fiddling around with llms and Ollama, however I have a feeling that it is something really simple that I'm missing. Any help would be appreciated.
If you created model using Modelfile in console like this
ollama create my_model -f my_model_file
then you can use it as any other model
in Python
ollama.chat(model='my_model', ...)
in console
ollama run my_model
Python also allows to create model directly in code:
import ollama
ollama.create(model="my_model", from_="llama3.2", system="you are jarvis from movie iron man")
and later you can use it as before
in Python
ollama.chat(model='my_model', ...)
in console
ollama run my_model
You can also check if your model is available
in console
ollama list
in Python
import ollama
response = ollama.list()
for item in response.models:
print('Name:', item.model)
You can find more complex script in official examples on GitHub: list.py
In some old articles/tutorials I saw ollama.create(..., modelfile=your_model_file)
but current version (0.5.1
) can't load it directly from file.
Version 0.4.5
still had it (source code) but 0.4.6
doesn't have it (source code)
Example of old article with modelfile
on ollama blog:
Python & JavaScript Libraries, January 23, 2024