Ollama can be install on your computer and changed according with your hardware requirements.
I will show how to duplicate a model and use it, see these commands:
D:\AI_Ollama>ollama list
NAME ID SIZE MODIFIED
orca-mini:3b 2dbd9f439647 2.0 GB 16 hours ago
tinyllama:latest 2644915ede35 637 MB 16 hours ago
gemma2:2b 8ccf136fdd52 1.6 GB 16 hours ago
llama3.1:latest 46e0c10c039e 4.9 GB 5 weeks ago
llava:latest 8dd30f6b0cb1 4.7 GB 2 months ago
D:\AI_Ollama>
D:\AI_Ollama>
D:\AI_Ollama>
D:\AI_Ollama>ollama show tinyllama:latest --modelfile
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
# FROM tinyllama:latest
FROM C:\Users\nicol\.ollama\models\blobs\sha256-2af...
TEMPLATE "<|system|>
{{ .System }}
<|user|>
{{ .Prompt }}
<|assistant|>
"
SYSTEM You are a helpful AI assistant.
PARAMETER stop <|system|>
PARAMETER stop <|user|>
PARAMETER stop <|assistant|>
PARAMETER stop
D:\AI_Ollama>ollama show tinyllama:latest --modelfile > new.modelfile
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
copying file sha256:2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816 100%
parsing GGUF
using existing layer sha256:2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816
using existing layer sha256:af0ddbdaaa26f30d54d727f9dd944b76bdb926fdaf9a58f63f78c532f57c191f
using existing layer sha256:c8472cd9daed5e7c20aa53689e441e10620a002aacd58686aeac2cb188addb5c
using existing layer sha256:fa956ab37b8c21152f975a7fcdd095c4fee8754674b21d9b44d710435697a00d
writing manifest
success
D:\AI_Ollama>ollama list
NAME ID SIZE MODIFIED
new-tinyllama:latest add64faa5d3b 637 MB 5 seconds ago
orca-mini:3b 2dbd9f439647 2.0 GB 16 hours ago
tinyllama:latest 2644915ede35 637 MB 16 hours ago
gemma2:2b 8ccf136fdd52 1.6 GB 16 hours ago
llama3.1:latest 46e0c10c039e 4.9 GB 5 weeks ago
llava:latest 8dd30f6b0cb1 4.7 GB 2 months ago
D:\AI_Ollama>ollama run new-tinyllama:latest
>>> Hello
Yes, of course! I'm happy to hear that you found my message helpful. Let
me know if you have any further questions or concerns in the future. If
you need further assistance, feel free to reach out again at your
convenience. Have a great day!
>>> Send a message (/? for help)