-
-
I have added support for Ollama to my chat-cli tool (formerly named cli-chatgpt, see previous posts). Ollama is a very easy way to run
llama2locally; it runs a local HTTP server and this will be used.Depending on whether an OpenAI key or Ollama are available, the models will be made available and you can see this in
cgt -h. (Usingcgtas a command here is based on my recommendation of using an alias in your shell.)When you’re offline, OpenAI is deemed unavailable, thus–if it is installed and active–Ollama/llama2 will be used automatically. While it currently doesn’t switch between models during a conversation (because of your online state), you can simply exit and re-continue the conversation with another model using
cgt -l.Both models show their responses in a streaming way.
❯ cgt where is gerolstein Model: gpt-3.5-turbo > where is gerolstein Gerolstein is a town in the Bitburg-Prüm district in Rhineland-Palatinate, Germany. It is located in the Eifel mountain range, approximately 15 kilometers southeast of Bitburg. > ^c ### Went offline here. ❯ cgt where is gerolstein Model: llama2:latest > where is gerolstein Gerolstein is a town located in the state of Rhineland-Palatinate, Germany. It is situated in the northern part of the state, approximately 20 kilometers (12 miles) northwest of the city of Mainz. The exact address of Gerolstein is: Gerolstein, Germany If you are planning to visit Gerolstein or need more detailed information, please let me know and I will be happy to help. >While there is no config available, you can change the priority of models in the source.
