0


乌班图服务器24.04安装英伟达显卡4090驱动,并使用Ollama安装qwen2.5:32b大模型

一、安装显卡

1、系统要求和准备

  1. # 查看设备信息
  2. ubuntu-drivers devices

运行后会显示ubuntu系统现在能识别到的显卡:

  1. udevadm hwdb is deprecated. Use systemd-hwdb instead.
  2. udevadm hwdb is deprecated. Use systemd-hwdb instead.
  3. udevadm hwdb is deprecated. Use systemd-hwdb instead.
  4. udevadm hwdb is deprecated. Use systemd-hwdb instead.
  5. udevadm hwdb is deprecated. Use systemd-hwdb instead.
  6. udevadm hwdb is deprecated. Use systemd-hwdb instead.
  7. ERROR:root:aplay command not found
  8. == /sys/devices/pci0000:00/0000:00:01.0/0000:01:00.0 ==
  9. modalias : pci:v000010DEd00002684sv00001043sd00008933bc03sc00i00
  10. vendor : NVIDIA Corporation
  11. model : AD102 [GeForce RTX 4090]
  12. driver : nvidia-driver-535 - distro non-free
  13. driver : nvidia-driver-535-server - distro non-free
  14. driver : nvidia-driver-550-open - distro non-free
  15. driver : nvidia-driver-535-server-open - distro non-free
  16. driver : nvidia-driver-535-open - distro non-free
  17. driver : nvidia-driver-550 - distro non-free recommended
  18. driver : xserver-xorg-video-nouveau - distro free builtin

然后,你要确保你的系统软件包是最新的,避免出现兼容性问题

  1. # 更新系统和软件包
  2. sudo apt update && sudo apt upgrade -y

2、命令行安装方式

如果使用命令行的方式安装显卡驱动(这是最快捷的方式,也是我推荐的方式),你只需要执行下面的命令。

  1. 两种方案,二选一
  1. # 方式1:使用系统工具自动安装
  2. # 使用这个方法会安装带有recommended字段的驱动,即推荐的驱动
  3. ubuntu-drivers devices
  4. sudo ubuntu-drivers autoinstall
  5. # 方式2:
  6. # 或者你想要安装特定版本的驱动,你只需要这样
  7. ubuntu-drivers devices
  8. sudo apt install nvidia-driver-535
  9. # 使用apt命令再加上上面"ubuntu-drivers devices"里列表任意一个驱动

在安装脚本运行完成后你需要重启电脑。

  1. # 重启
  2. sudo reboot

重启完成后使用这条命令查看驱动是否安装成功。

  1. # 查看NVIDIA
  2. nvidia-smi

如果成功输出了显卡信息,代表驱动安装成功了。如下

  1. root@4090:~# nvidia-smi
  2. Thu Sep 19 06:39:10 2024
  3. +-----------------------------------------------------------------------------------------+
  4. | NVIDIA-SMI 550.107.02 Driver Version: 550.107.02 CUDA Version: 12.4 |
  5. |-----------------------------------------+------------------------+----------------------+
  6. | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
  7. | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
  8. | | | MIG M. |
  9. |=========================================+========================+======================|
  10. | 0 NVIDIA GeForce RTX 4090 Off | 00000000:01:00.0 Off | Off |
  11. | 0% 29C P8 12W / 450W | 2MiB / 24564MiB | 0% Default |
  12. | | | N/A |
  13. +-----------------------------------------+------------------------+----------------------+
  14. +-----------------------------------------------------------------------------------------+
  15. | Processes: |
  16. | GPU GI CI PID Type Process name GPU Memory |
  17. | ID ID Usage |
  18. |=========================================================================================|
  19. | No running processes found |
  20. +-----------------------------------------------------------------------------------------+

二、使用Ollama安装qwen2.5:32b模型

官网:https://github.com/ollama/ollama/blob/main/docs/linux.md

1、下载安装包

  1. curl -L https://ollama.com/download/ollama-linux-amd64.tgz -o ollama-linux-amd64.tgz
  2. sudo tar -C /usr -xzf ollama-linux-amd64.tgz

2、将 Ollama 添加为启动服务(推荐)

为 Ollama 创建用户和组:

  1. #创建ollama和用户组,指定家目录且不可登录
  2. sudo useradd -r -s /bin/false -U -m -d /usr/ollama ollama
  3. #把当前用户加入到ollama用户组,获取组的所有权限(让当前用户能进入ollama家目录,读写某些目录和文件的权限)。
  4. sudo usermod -a -G ollama $(whoami)

查看命令:

  1. user@4090:~$ ollama serve -h
  2. Start ollama
  3. Usage:
  4. ollama serve [flags]
  5. Aliases:
  6. serve, start
  7. Flags:
  8. -h, --help help for serve
  9. Environment Variables:
  10. OLLAMA_DEBUG Show additional debug information (e.g. OLLAMA_DEBUG=1)
  11. OLLAMA_HOST IP Address for the ollama server (default 127.0.0.1:11434)
  12. OLLAMA_KEEP_ALIVE The duration that models stay loaded in memory (default "5m")
  13. OLLAMA_MAX_LOADED_MODELS Maximum number of loaded models per GPU
  14. OLLAMA_MAX_QUEUE Maximum number of queued requests
  15. OLLAMA_MODELS The path to the models directory
  16. OLLAMA_NUM_PARALLEL Maximum number of parallel requests
  17. OLLAMA_NOPRUNE Do not prune model blobs on startup
  18. OLLAMA_ORIGINS A comma separated list of allowed origins
  19. OLLAMA_SCHED_SPREAD Always schedule model across all GPUs
  20. OLLAMA_TMPDIR Location for temporary files
  21. OLLAMA_FLASH_ATTENTION Enabled flash attention
  22. OLLAMA_LLM_LIBRARY Set LLM library to bypass autodetection
  23. OLLAMA_GPU_OVERHEAD Reserve a portion of VRAM per GPU (bytes)
  24. OLLAMA_LOAD_TIMEOUT How long to allow model loads to stall before giving up (default "5m")

3、做成systemd服务

在systemd中创建服务文件 :/etc/systemd/system/ollama.service

  1. [Unit]
  2. Description=Ollama Service
  3. After=network-online.target
  4. [Service]
  5. ExecStart=/usr/bin/ollama serve
  6. User=ollama
  7. Group=ollama
  8. Restart=always
  9. RestartSec=3
  10. Environment="PATH=$PATH"
  11. Environment="OLLAMA_HOST=0.0.0.0:11434"#更改监听的IP和端口,默认的只能本机访问
  12. #Environment="OLLAMA_MODELS=/home/user/ollama/ollama_models"#更改模型的存储路径,默认在用户的家目录的.ollama/models下
  13. [Install]
  14. WantedBy=default.target

然后启动服务:

  1. sudo systemctl daemon-reload
  2. sudo systemctl enable ollama
  3. sudo systemctl status ollama

4、安装qwen2.5:32b模型

  1. ollama run qwen2.5:32b

这个模型20G,安装时会有点慢

下载完记得查看一下模型的路径对了没

  1. user@4090:/home/ollama$ du -sh /home/ollama/.ollama/models/
  2. 19G /home/ollama/.ollama/models
标签: linux 机器学习

本文转载自: https://blog.csdn.net/m0_60171711/article/details/142359745
版权归原作者 m0_60171711 所有, 如有侵权,请联系我们删除。

“乌班图服务器24.04安装英伟达显卡4090驱动,并使用Ollama安装qwen2.5:32b大模型”的评论:

还没有评论