Hi list, Someone asked me privately, so I thought that maybe others would be interested in a simple way to interact between LC and AI models locally on their machine.
The method uses Ollama, which is available for Mac, Win and Linux : https://ollama.com 1- Download and install Ollama (plenty of tutorials on youtube) 2- run Ollama via Terminal (on Mac) : ollama serve 3- load a model via Terminal (from HuggingFace for instance) : ollama run llama3.2:1b 4- in a livecode stack, create a field with the following content : curl -X POST http://localhost:11434/api/generate \ -H "Content-Type: application/json" \ -d '{ "model": "llama3.2:1b", "prompt": "What is the capital of France ?", "stream": false }' 5- create a button with the following script : on mouseup put field 1 into tcurl put shell(tcurl) end mouseup 6- the answer from the model displays in the message box in json format. This works great on machines with a GPU and at least 16 Gb of RAM. The more RAM, the bigger the models that can be used. Enjoy, jbv _______________________________________________ use-livecode mailing list use-livecode@lists.runrev.com Please visit this url to subscribe, unsubscribe and manage your subscription preferences: http://lists.runrev.com/mailman/listinfo/use-livecode