HomePostsProjects

Using Ollama to run Llama3.1

July 29, 2024

  • download ollama - https://ollama.com/

App:

  • run ollama run llama3.1, this will spin up inference

Python:

  • pip install ollama
import ollama
response = ollama.chat(model='llama3.1', messages=[
  {
    'role': 'user',
    'content': 'Why is the sky blue?',
  },
])
print(response['message']['content'])

Related:

  • https://github.com/continuedev/continue
  • https://marketplace.visualstudio.com/items?itemName=Continue.continue

© 2025 Luke Miloszewski

Email AddressGitHubTwitter