Pages

Se afișează postările cu eticheta ollama. Afișați toate postările
Se afișează postările cu eticheta ollama. Afișați toate postările

marți, 27 mai 2025

News : tested local artificial intelligence with gemma2:2b and void editor.

I had a new full instaled windows 10 with the gemma2:2b in order to work as local artificial intelligence, with the 4Gb RAM!
Ollama gemma2:2b can works with minimal RAM!
The netstat command can show you if ollama server works, see bellow image:
I have done some tests with source code generation and it is a quite functional artificial intelligence, but at a minimal level. If you want something functional with a minimum of 4G RAM, then it works very well !
The settings of local ollama into void editor can be set on first running :

miercuri, 21 mai 2025

News : ... void editor I.D.E with local ollama feature.

Yesterday I tested this I.D.E. named void and works great.
You can see this project on the GitHub repo by voideditor.
You can see more on the official website !!!
NOTE : on web you can find a reference to this feature: you can use ollama as local artificial intelligence with void editor

joi, 17 aprilie 2025

Artificial Intelligence: Ollama duplicate model .

Ollama can be install on your computer and changed according with your hardware requirements.
I will show how to duplicate a model and use it, see these commands:
D:\AI_Ollama>ollama list
NAME                ID              SIZE      MODIFIED
orca-mini:3b        2dbd9f439647    2.0 GB    16 hours ago
tinyllama:latest    2644915ede35    637 MB    16 hours ago
gemma2:2b           8ccf136fdd52    1.6 GB    16 hours ago
llama3.1:latest     46e0c10c039e    4.9 GB    5 weeks ago
llava:latest        8dd30f6b0cb1    4.7 GB    2 months ago

D:\AI_Ollama>
D:\AI_Ollama>
D:\AI_Ollama>
D:\AI_Ollama>ollama show tinyllama:latest --modelfile
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
# FROM tinyllama:latest

FROM C:\Users\nicol\.ollama\models\blobs\sha256-2af...
TEMPLATE "<|system|>
{{ .System }}
<|user|>
{{ .Prompt }}
<|assistant|>
"
SYSTEM You are a helpful AI assistant.
PARAMETER stop <|system|>
PARAMETER stop <|user|>
PARAMETER stop <|assistant|>
PARAMETER stop 


D:\AI_Ollama>ollama show tinyllama:latest --modelfile >  new.modelfile

gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
gathering model components
copying file sha256:2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816 100%
parsing GGUF
using existing layer sha256:2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816
using existing layer sha256:af0ddbdaaa26f30d54d727f9dd944b76bdb926fdaf9a58f63f78c532f57c191f
using existing layer sha256:c8472cd9daed5e7c20aa53689e441e10620a002aacd58686aeac2cb188addb5c
using existing layer sha256:fa956ab37b8c21152f975a7fcdd095c4fee8754674b21d9b44d710435697a00d
writing manifest
success

D:\AI_Ollama>ollama list
NAME                    ID              SIZE      MODIFIED
new-tinyllama:latest    add64faa5d3b    637 MB    5 seconds ago
orca-mini:3b            2dbd9f439647    2.0 GB    16 hours ago
tinyllama:latest        2644915ede35    637 MB    16 hours ago
gemma2:2b               8ccf136fdd52    1.6 GB    16 hours ago
llama3.1:latest         46e0c10c039e    4.9 GB    5 weeks ago
llava:latest            8dd30f6b0cb1    4.7 GB    2 months ago

D:\AI_Ollama>ollama run new-tinyllama:latest
>>> Hello
Yes, of course! I'm happy to hear that you found my message helpful. Let
me know if you have any further questions or concerns in the future. If
you need further assistance, feel free to reach out again at your
convenience. Have a great day!

>>> Send a message (/? for help)