Logo Ricky
  • 首頁
  • 關於
  • 更多的
    技能 經歷 學歷 專案
  • 文章
  • 筆記
  • 活動
  • 跑步
  • Chinese
    English Chinese
  • Dark Theme
    Light Theme Dark Theme System Theme
Logo Inverted Logo
  • 标签
  • AI
  • Ansible
  • APP
  • BASH
  • Browser
  • Byte
  • Chart
  • Checklist
  • Chrome OS
  • CI
  • Cloudflare
  • Command Line
  • Config
  • Container
  • CURL
  • Data
  • Docker
  • Emoji
  • EMQX
  • GIN
  • Git
  • Github
  • Gitlab
  • Go
  • Golang
  • Google
  • GraphQL
  • Infra
  • Infrastructure
  • Introduction
  • ITerm2
  • Kubernetes
  • Leetcode
  • Lightweight
  • Linux
  • Llm
  • Machinelearning
  • MacOS
  • MQTT
  • Network
  • Nginx
  • Ollama
  • Prometheus
  • Push
  • Python
  • Regex
  • Security
  • SFTP
  • SHELL
  • Slides
  • Snow
  • Software
  • Ssh
  • String
  • Switch
  • Synology
  • Terminal
  • Travel
  • Tunnel
  • URL
  • Video
  • Visualization
  • Web
  • Zero Trust
Hero Image
Run llama3

ollama Instructions Docker-compose version: "3.8" services: ollama: image: ollama/ollama:latest container_name: ollama restart: unless-stopped volumes: - ./ollama/ollama:/root/.ollama tty: true ports: - 11434:11434 networks: - ollama-docker # deploy: # resources: # reservations: # devices: # - driver: nvidia # count: 1 # capabilities: [gpu] ollama-webui: image: ghcr.io/ollama-webui/ollama-webui:main container_name: ollama-webui restart: unless-stopped volumes: - ./ollama/ollama-webui:/app/backend/data ports: - 8080:8080 environment: - "/ollama/api=http://ollama:11434/api" extra_hosts: - host.docker.internal:host-gateway networks: - ollama-docker networks: ollama-docker: Setup # Run docker-compose docker-compose up -d # Pull model(https://ollama.

Thursday, April 25, 2024 閱讀
導覽列
  • 關於
  • 技能
  • 經歷
  • 學歷
  • 專案
聯絡方式:
  • zeyanlin@outlook.com
  • linzeyan
  • Ricky
  • Ricky
  • Ricky