I had 1 GB of swap. It seems Ohava realized they had made a mistake when they provisioned a large group of nodes in October and (1) rebuilt my node and (2) offered to upgrade everyone to 40 GB of disk space in a public announcement on their web site.. My node was rebuilt last night and df now shows Filesystem Size Used Avail Use% Mounted on /dev/mapper/ubuntu--vg-root 19G 1.8G 16G 10% / none 4.0K 0 4.0K 0% /sys/fs/cgroup udev 235M 4.0K 235M 1% /dev tmpfs 50M 368K 49M 1% /run none 5.0M 0 5.0M 0% /run/lock none 246M 0 246M 0% /run/shm none 100M 0 100M 0% /run/user /dev/vda1 236M 68M 156M 31% /boot and 1 GB of swap, according to top: top - 07:36:27 up 8:30, 1 user, load average: 0.00, 0.01, 0.05 Tasks: 72 total, 1 running, 71 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st KiB Mem: 501804 total, 484280 used, 17524 free, 53080 buffers KiB Swap: 1044476 total, 3472 used, 1041004 free. 353988 cached Mem The memory is short (sb 512 MB), but that is a disagreement over 1 KB - 1,000 versus 1 KB = 1,024, which I won't win. So, it seems like an "honest" mistake on their part, unless I am missing something. Thanks! Mark On Thu, Oct 16, 2014 at 10:03 PM, Sesso wrote: > How much swap do you have? > > > On Oct 16, 2014, at 9:41 PM, Mark Phillips > wrote: > > I signed up for a free VPS on Ohava - 20GB is what is advertised. When I > logged into the system, df -h showed this: > > Filesystem Size Used Avail Use% Mounted on > /dev/mapper/ubuntu--vg-root 6.6G 1.8G 4.6G 28% / > none 4.0K 0 4.0K 0% /sys/fs/cgroup > udev 235M 4.0K 235M 1% /dev > tmpfs 50M 368K 49M 1% /run > none 5.0M 0 5.0M 0% /run/lock > none 246M 0 246M 0% /run/shm > none 100M 0 100M 0% /run/user > /dev/vda1 236M 68M 156M 31% /boot > > I queried to the support group, so they sent me instructions to add 10 > more GB. > > df now shows: > > Filesystem Size Used Avail Use% Mounted on > /dev/mapper/ubuntu--vg-root 17G 1.8G 14G 12% / > none 4.0K 0 4.0K 0% /sys/fs/cgroup > udev 235M 4.0K 235M 1% /dev > tmpfs 50M 368K 49M 1% /run > none 5.0M 0 5.0M 0% /run/lock > none 246M 0 246M 0% /run/shm > none 100M 0 100M 0% /run/user > /dev/vda1 236M 68M 156M 31% /boot > > The support groups said: > > > *Our apologies on the confusion. This is a current bug in the machines > being spun up, but we definitely offer (and want to help you get) the full > amount of space. Every instance gets 20GB partitioned to them. There is > some overhead in some of the other partitions of disk space so / won't ever > show the full 20GB, as small parts of the 20GBs are allocated elsewhere. * > > > *The instructions sent to you are for extending the / partition an > additional 10GB. Since there was already some amount of storage there, the > result after completing the instructions (and the correction to the error > that we made when we originally sent you instructions on expanding the lvm > volume..."sudo lvextend -L+10G /dev/ubuntu-vg/root") should get you as > close as possible to 20GB (~18.75GB) on /root while allowing for the > overhead. * > I get the calculation of disk size issue, so a 20 GB drive is really only > 18.63 GB. But shouldn't df show 18.63 GB and not 17 GB? Is the discrepancy > (1.63 GB or 8.75% of the drive) due to formatting the disk and adding > Ubuntu server? > > Thanks, > > Mark > --------------------------------------------------- > PLUG-discuss mailing list - PLUG-discuss@lists.phxlinux.org > To subscribe, unsubscribe, or to change your mail settings: > http://lists.phxlinux.org/mailman/listinfo/plug-discuss > > > > --------------------------------------------------- > PLUG-discuss mailing list - PLUG-discuss@lists.phxlinux.org > To subscribe, unsubscribe, or to change your mail settings: > http://lists.phxlinux.org/mailman/listinfo/plug-discuss >