First try to unlock the pct (assuming your troubled container is 101):
pct unlock 101
if it works just stop and start again the vm. if it does’nt work (my case) try to stop with this
lxc-stop --name 101
if it’s does’nt work (my case) you can force stop with kill command
ps ax | grep lxc
then kill the process with your id (101 for me) kill pid (replace pid by the process name). After that you can just launch again you’r vm
Copy and paste following command to the terminal
(6.1 and up)
(6.2-11 and up)
(6.2-12 and up) Continue reading “Remove Proxmox Subscription Notice” »
Step 1 : Migrate all VMs to another active node
Migrate all VMs to another active node. You can use the live migration feature if you have a shared storage or offline migration if you only have local storage.
Step 2 : Display all active nodes
Display all active nodes in order identify the name of the node you want to remove Continue reading “Remove Node from Proxmox Cluster” »
When trying to “Stop” or “Shutdown” virtual machine from Proxmox (PVE) web gui, the “Cluster log” shows
end task UPID:pve:xxxxxxxx:xxxxxxxx:xxxxxxx:qmstop:xxx:root@pam: can’t lock file ‘/var/lock/qemu-server/lock-xxx.conf’ -got timeout
end task UPID:pve:xxxxxxxx:xxxxxxxx:xxxxxxx:qmreboot:xxx:root@pam: VM quit/powerdown failed
We can manually delete the lock from following path
# The file will be
Make sure only delete the correct one!
You can also do it using script from this site-
Sparse disk image formats such as qcow2 only consume the physical disk space which they need. For example, if a guest is given a qcow2 image with a size of 100GB but has only written to 10GB then only 10GB of physical disk space will be used. There is some slight overhead associated, so the above example may not be strictly true, but you get the idea.
Sparse disk image files allow you to over allocate virtual disk space – this means that you could allocate 5 virtual machines 100GB of disk space, even if you only have 300GB of physical disk space. If all the guests need 100% of their 100GB disk space then you will have a problem. If you use over allocation of disk space you will need to monitor the physical disk usage very carefully.
There is another problem with sparse disk formats, they don’t automatically shrink. Let’s say you fill 100GB of a sparse disk (we know this will roughly consume 100GB of physical disk space) and then delete some files so that you are only using 50GB. The physical disk space used should be 50GB, right? Wrong. Because the disk image doesn’t shrink, it will always be 100GB on the file system even if the guest is now using less. The below steps will detail how to get round this issue. Continue reading “Reclaim disk space from a sparse image file qcow2/ vmdk” »
In this guide we will go over creating a Proxmox KVM Template from a Cloud Image. This same process will work for any Cloud-Init Openstack based image type you can find online.
Having done a number of these for our Proxmox based VPS service I wanted to post up a guide to help anyone else looking to do the same thing.
My workflow for customizing one of those for use with Proxmox with cloud-init deployment from WHMCS and root login is below. Once you setup one template you can rapidly reinstall new containers and test stuff.
If not installed already installed you will need libguestfs-tools :
apt-get install libguestfs-tools
To edit the image before importing. We will use virt-edit which is a part of libguestfs-tools. Continue reading “Proxmox Cloud-Init OS template creation” »
To update the present cluster host proxmox following files need to be updated:
/etc/pve/corosync.conf (only on one node necessary)
However, corosync.conf needs special way to edit the file!
Editing the corosync.conf file is not always very straightforward. There are two on each cluster node, one in /etc/pve/corosync.conf and the other in /etc/corosync/corosync.conf. Editing the one in our cluster file system will propagate the changes to the local one, but not vice versa. The configuration will get updated automatically as soon as the file changes. This means changes which can be integrated in a running corosync will take effect immediately. So you should always make a copy and edit that instead, to avoid triggering some unwanted changes by an in-between safe.
cp /etc/pve/corosync.conf /etc/pve/corosync.conf.new
Then open the config file with your favorite editor, nano and vim.tiny are preinstalled on any Proxmox VE node for example. Continue reading “Change cluster node IP in Proxmox” »
The qemu-guest-agent is a helper daemon, which is installed in the guest. It is used to exchange information between the host and guest, and to execute command in the guest.
In Proxmox VE, the qemu-guest-agent is used for mainly two things:
- To properly shutdown the guest, instead of relying on ACPI commands or windows policies
- To freeze the guest file system when making a backup (on windows, use the volume shadow copy service VSS).
You have to enable the guest-agent per VM, either set it in the GUI to “Yes” under options (see screenshot):
or via CLI:
qm set VMID --agent 1
On Linux you have to simply install the qemu-guest-agent, please refer to the documentation of your system.
We show here the commands for Debian/Ubuntu and Redhat based systems:
Continue reading “Install Qemu Guest Agent on Proxmox” »
Proxmox Virtual Environment (VE) is an enterprise-grade open-source server virtualization solution based on Debian Linux distribution with a modified Ubuntu LTS kernel. It allows you to deploy and manage both virtual machines and containers.
This setup presumes you have a running Debian 10 Buster Linux server running. If you don’t have one, follow our guide to Install Debian 10 on a dedicated server that will be used as a hypervisor. Please note that you need a 64-bit processor with support for the Intel 64 or AMD64 CPU extensions.
Below are the steps you’ll follow through to install Proxmox VE 6 on Debian 10 (Buster).
Step 1: Update Debian OS
Update apt package index before getting started.
sudo apt -y update
sudo apt -y upgrade
Step 2: Set system hostname
We need to set the hostname and make sure it is resolvable via /etc/hosts.
sudo hostnamectl set-hostname prox6node01.example.com --static
echo "10.1.1.10 prox6node01.example.com prox6node01" | sudo tee -a /etc/hosts
example.com should be replaced with a valid domain name. Continue reading “Install Proxmox VE 6 on Debian 10 (Buster)” »
First remove the existing LVM-Thin:
Then create an normal lvm on existing group. For example:
lvcreate -L 755.96G -n data pve
Mount it in the fstab. The mount target must be empty, so delete everything in there.
/dev/pve/data /var/lib/vz ext4 defaults 0 2
rm -rf /var/lib/vz/* && mount -a
After all create you normal directory storage.