How to Set up and Run a Local LLM with Ollama and Llama 2 - The New Stack

Gary Thompson @gwthompson