Install Proxmox 7 on Debian 12

Install a standard Debian Bookworm, for details consider the Debian installation guide, and configure a static IP.

Note: The Debian installer performs network configuration by IPv6 autoconfiguration and DHCP by default, if available. To force manual network configuration in the UEFI installer, press E and add netcfg/disable_autoconfig=true to the linux command line. For the BIOS installer, select the Help entry and start the installation by typing installgui netcfg/disable_autoconfig=true. For further details, see the Debian installer documentation.

It is recommended to only install the “standard system utilities” and “SSH server” package selection, as Proxmox VE brings its own packages for QEMU and LXC. A desktop environment is not necessary.

Add an /etc/hosts entry for your IP address

The hostname of your machine must be resolvable via /etc/hosts. This means that in /etc/hosts you need one of the following entries for your hostname:

Read more

Share

Fixing Proxmox Delete user failed: cannot update tfa config, following nodes are not up to date:

Please check if all the nodes are on the same package versions: run

pveversion -v

and compare the outputs.

If they’re not matching, you can run

apt update && apt dist-upgrade

To upgrade the nodes. if there’s a kernel upgrade you will also need to reboot the machines. If after the upgrades you still don’t see the same versions of the packages, then maybe you have wrong repositories set up [0]

Share

VNC Client Access on Proxmox KVM guests

Configure VNC Access in the Configuration File

Add a line to the VM’s configuration file /etc/pve/local/qemu-server/<VMID>.conf which specifies the VNC display number as follows (“77” in the example below):

args: -vnc 0.0.0.0:77

If you want to use password protection, add:

args: -vnc 0.0.0.0:77,password=on

The display number can be freely chosen, but each number must occur only once. The VNC service then listens at port 5900+display_number. Note that connections via noVNC use display number 0 and following, therefore it is recommended to use higher numbers in order to avoid conflicts.

You can now connect the VNC client to the host IP address and port as chosen (“5977” in the example above).

Read more

Share

Install pfSense Qemu Guest Agent

Software Versions

The following software versions were used in this post.

  • pfSense Community Edition – 2.6.0
  • QEMU Guest Agent – 5.0.1

Install the qemu-guest-agent package.

pkg install -y qemu-guest-agent

Once installed, there is a note on how to setup the agent.

Modify your `/etc/rc.conf` by adding these settings

qemu_guest_agent_enable="YES"
qemu_guest_agent_flags="-d -v -l /var/log/qemu-ga.log"

and run

Read more

Share

Extreme slow internet speed pfsense over proxmox

For a qemu proxmox guest PFSense acts weriedly with the network speed- it gets extremely slow. So her goes the little tweaks that worked for me-

First, I chose Intel E1000 Interfaces instead VirtIO.

Second, in the PFSense webconsole-

In pfSense GUI, System > Advanced > Networking > Tick on-

  • Disable hardware checksum offload
  • Disable hardware TCP segmentation offload
  • Disable hardware large receive offload
Share

Perfect Proxmox Template with Cloud Image and Cloud Init

Instructions

Choose your Ubuntu Cloud Image. Here in this example will use ubuntu cloud-init image. Download Ubuntu (replace with the url of the one you chose from above)

wget https://cloud-images.ubuntu.com/focal/current/focal-server-cloudimg-amd64.img

Create a new virtual machine

qm create 8000 --memory 2048 --core 2 --name ubuntu-cloud --net0 virtio,bridge=vmbr0

Import the downloaded Ubuntu disk to local-lvm storage

Read more

Share

Proxmox GUI webpage stuck on loading

Connection time out!!!! So boring nah?

I can’t figure out where the problem is. It works and then couple of hours later I’d come back and it keeps on loading forever. SSH works fine, can SSH into containers as well. Just the GUI doesn’t work. Tried full upgrade. Checked for network conflicts, doesn’t seem to be any.

Here’s the tidbit solution to this probelm like similar (BTW works on PVE 6/7 version apparent)-

If browser refresh doesn’t work, then please post the output:

apt update
apt full-upgrade
pveversion -v

Then Run

apt install --reinstall pve-manager proxmox-widget-toolkit
systemctl restart pveproxy pvestatd pvedaemon
systemctl daemon-reload
Share

Convert VirtualBox VM to ProxMox VE

  1. On the VirtualBox host, launch VirtualBox
  2. Right click the VM to convert > Settings
  3. Select Storage from the left navigation
  4. Click the virtual hard disk and copy the Location value for the full path of the disk to the clipboard
  5. Right click on the Start menu > Run > type cmd.exe > Press Enter
  6. Enter the following commands in the command prompt
    # change directory to VirtualBox installation
    cd %programfiles%\Oracle\VirtualBox
    # convert the .vdi to raw .img
    VBoxManage clonehd --format RAW "pasted\full\path\to.vdi" "pasted\full\path\to.raw"
  7. Once the .vdi to .raw conversion completes, open a web browser and navigate to the ProxMox web UI https://ProxMoxDNSorIP:8006/
  8. Click the Create VM button at the top right
  9. On the General tab, enter a VM Name and note the VM ID generated > click Next
  10. On the OS tab select Do not use any media and set the Guest OS > click Next
  11. On the System tab click Next
  12. On the Hard Disk tab set the Disk size to 0.001 > click Next
  13. On the CPU tab set the number of CPUs > click Next
  14. On the Memory tab set the amount of memory to allocate in MiB > click Next
  15. On the Network tab click Next
  16. On the Confirm tab review the settings and click Finish
  17. Select the newly created VM from the left navigation panel > Hardware
  18. Click the Hard Disk to select it
  19. Click the Detach button to detach the hard disk from the VM
  20. Click the Unused disk
  21. Click the Remove button to permanently delete it
  22. Download WinSCP Download
  23. Extract WinSCP and run the executable
  24. Connect to the ProxMox IP server via WinSCP
  25. Copy the VirtualBox created .raw file to a location on the ProxMox server that has enough free disk space, /root for example
  26. Back in the browser, open the ProxMox host Shell
  27. Run the following command to import the raw disk, modify the .raw file name and VM ID noted earlier
    # import the raw disk
    # usage:
    # qm importdisk <VM ID> /root/<source disk file>.raw <destination storage pool name>
    qm importdisk 100 vbox.raw HDD_500GB --format qcow2
  28. Once the disk import completes, select the target VM from the left navigation menu > Hardware
  29. Double click the Unused Disk > Click the Add button
  30. Select Options from the left navigation sub-menu
  31. Double click Boot Order
  32. Check the Enabled box next to the hard disk
  33. Drag the Hard disk up in the boot order as needed, typically below the CD-ROM device
  34. Click OK
  35. Click the Start button in the top right of the screen
  36. Click the Console link to watch the boot process

Src:

https://i12bretro.github.io/tutorials/0295.html

Share

Install Proxmox VE 7 on Debian 11 (Bullseye)

In this guide, we will cover a step-by-step installation of Proxmox VE 7 virtualization software on Debian 11 (Bullseye) Linux system. It’s recommended to deploy Proxmox VE server from a Bare-metal_ISO_Installer, but it’s sometimes inevitable to deploy it on a running instance of Debian 11 (Bullseye) server.

Setup Pre-requisites

For the installation of Proxmox VE 7 on Debian 11 (Bullseye), you need the following requirements to be met;

  • A running instance of Debian Bullseye
  • A 64-bit processor with support for the Intel 64 or AMD64 CPU extensions.
  • Access to Debian server terminal as root or standard user with sudo
  • Server needs internet access
  • Enough hardware resources to be used in Virtualizing other operating systems

Read more

Share

Proxmox VM auto start VM after found in shutdown state

I was having this strange issue, where a running busy VM stopeed all of a sudden due to high CPU or Memory overload issue. So manually had to start everythime. In order to avoid this, created a small script to start the VM in case if it’s down.

#!/bin/bash

# Set environment
PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin

su -

if [[ $(qm status 101) = *"status: stopped"* ]];
then
echo `qm start 101`
fi
Share