[Bug 1657646] Comment bridged from LTC Bugzilla
bugproxy
bugproxy at us.ibm.com
Tue Jun 27 13:20:15 UTC 2017
------- Comment From vinuthags at in.ibm.com 2017-06-27 08:44 EDT-------
I tried again after installing thin-provisioning-tools, after the reboot.
The docker service/daemon was up.
I don't see the issue
root at bamlp5:~# apt-get install thin-provisioning-tools
Reading package lists... Done
Building dependency tree
Reading state information... Done
The following NEW packages will be installed:
thin-provisioning-tools
0 upgraded, 1 newly installed, 0 to remove and 0 not upgraded.
Need to get 339 kB of archives.
After this operation, 1,627 kB of additional disk space will be used.
Get:1 http://gb.ports.ubuntu.com/ubuntu-ports xenial/universe ppc64el thin-provisioning-tools ppc64el 0.5.6-1ubuntu1 [339 kB]
Fetched 339 kB in 0s (429 kB/s)
Selecting previously unselected package thin-provisioning-tools.
(Reading database ... 64237 files and directories currently installed.)
Preparing to unpack .../thin-provisioning-tools_0.5.6-1ubuntu1_ppc64el.deb ...
Unpacking thin-provisioning-tools (0.5.6-1ubuntu1) ...
Processing triggers for man-db (2.7.5-1) ...
Setting up thin-provisioning-tools (0.5.6-1ubuntu1) ...
root at bamlp5:~# docker info
Containers: 0
Running: 0
Paused: 0
Stopped: 0
Images: 0
Server Version: 1.12.6
Storage Driver: devicemapper
Pool Name: docker--storage-thinpool
Pool Blocksize: 65.54 kB
Base Device Size: 10.74 GB
Backing Filesystem: xfs
Data file:
Metadata file:
Data Space Used: 11.8 MB
Data Space Total: 30.6 GB
Data Space Available: 30.59 GB
Metadata Space Used: 172 kB
Metadata Space Total: 318.8 MB
Metadata Space Available: 318.6 MB
Thin Pool Minimum Free Space: 3.06 GB
Udev Sync Supported: true
Deferred Removal Enabled: false
Deferred Deletion Enabled: false
Deferred Deleted Device Count: 0
Library Version: 1.02.110 (2015-10-30)
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: null host bridge overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options: apparmor seccomp
Kernel Version: 4.4.0-81-generic
Operating System: Ubuntu 16.04.2 LTS
OSType: linux
Architecture: ppc64le
CPUs: 8
Total Memory: 967.4 MiB
Name: bamlp5
ID: EFRP:SO44:CIMD:TPH5:BBKN:RYBT:QZGE:7U7H:DOYF:OBI6:SKLG:DUCO
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
WARNING: No swap limit support
Insecure Registries:
127.0.0.0/8
root at bamlp5:~# reboot
Connection to bamlp5 closed by remote host.
Connection to bamlp5 closed.
[vinutha at kte ~]$ ssh root at bamlp5
root at bamlp5's password:
Welcome to Ubuntu 16.04.2 LTS (GNU/Linux 4.4.0-81-generic ppc64le)
* Documentation: https://help.ubuntu.com
* Management: https://landscape.canonical.com
* Support: https://ubuntu.com/advantage
********************************************************************************
IBM Business Use Statement
--------------------------
IBM's internal systems must only be used for conducting IBM's business or
for purposes authorized by IBM management.
Use is subject to audit at any time by IBM management.
Distribution: Ubuntu 16.04.2 LTS
Kernel Build: 4.4.0-81-generic
System Name : bamlp5
Model/Type : 8247-22L
Platform : powerpc64le
********************************************************************************
Last login: Tue Jun 27 07:21:48 2017 from 10.33.11.31
root at bamlp5:~#
root at bamlp5:~#
root at bamlp5:~#
root at bamlp5:~# service docker status
* docker.service - Docker Application Container Engine
Loaded: loaded (/lib/systemd/system/docker.service; enabled; vendor preset: enabled)
Active: active (running) since Tue 2017-06-27 07:39:30 CDT; 1min 17s ago
Docs: https://docs.docker.com
Main PID: 1533 (dockerd)
Tasks: 27
Memory: 5.3M
CPU: 691ms
CGroup: /system.slice/docker.service
|-1533 /usr/bin/dockerd -H fd:// -s devicemapper --storage-opt=dm.thinpooldev=/dev/mapper/docker--storage-thinpool
`-1723 containerd -l unix:///var/run/docker/libcontainerd/docker-containerd.sock --shim containerd-shim --metrics-interval=0 --start-timeout 2m --state-dir /var/run/docke
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.186314000-05:00" level=info msg="Graph migration to content-addressability took 0.00 seconds"
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.186787000-05:00" level=warning msg="Your kernel does not support swap memory limit."
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.192019000-05:00" level=info msg="Loading containers: start."
Jun 27 07:39:30 bamlp5 dockerd[1533]: .time="2017-06-27T07:39:30.310077000-05:00" level=info msg="Firewalld running: false"
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.676802000-05:00" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.769394000-05:00" level=info msg="Loading containers: done."
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.776877000-05:00" level=info msg="Daemon has completed initialization"
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.776933000-05:00" level=info msg="Docker daemon" commit=78d1802 graphdriver=devicemapper version=1.12.6
Jun 27 07:39:30 bamlp5 dockerd[1533]: time="2017-06-27T07:39:30.791428000-05:00" level=info msg="API listen on /var/run/docker.sock"
Jun 27 07:39:30 bamlp5 systemd[1]: Started Docker Application Container Engine.
root at bamlp5:~# docker ps -a
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
f1b34b5f7ddd 590e44f15214 "/bin/sh -c ./staf_ba" 5 minutes ago Exited (0) 3 minutes ago bamlp5-base
root at bamlp5:~# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 30G 0 disk
|-sda1 8:1 0 7M 0 part
|-sda2 8:2 0 28.7G 0 part /
`-sda3 8:3 0 1.3G 0 part [SWAP]
sdb 8:16 0 30G 0 disk
|-sdb1 8:17 0 4M 0 part
|-sdb2 8:18 0 1G 0 part
`-sdb3 8:19 0 29G 0 part
|-rhel_bamlp500-swap 252:0 0 2G 0 lvm
`-rhel_bamlp500-root 252:1 0 57G 0 lvm
sdc 8:32 0 30G 0 disk
`-sdc1 8:33 0 30G 0 part
`-rhel_bamlp500-root 252:1 0 57G 0 lvm
sdd 8:48 0 30G 0 disk
|-docker--storage-thinpool_tmeta 252:2 0 304M 0 lvm
| `-docker--storage-thinpool 252:4 0 28.5G 0 lvm
`-docker--storage-thinpool_tdata 252:3 0 28.5G 0 lvm
`-docker--storage-thinpool 252:4 0 28.5G 0 lvm
root at bamlp5:~#
------- Comment From vinuthags at in.ibm.com 2017-06-27 08:51 EDT-------
@SEETEENA THOUFEEK - I'm didn't clearly understand comment 47 nd comment 49.
comment 47 says ,
What's left to check here is if perhaps thin-provisioning-tools should be a strong or weak dependency of docker.io, like a Recommends or Suggests.
@paelzer noted this in his comment #5:
"""
First of all thin provisioning is not mainstream, the dependency is only a suggest, so install the tools
$ sudo apt-get install thin-provisioning-tools
"""
After thin-provisioning tools is installed , the problem is not seen.
And comment 49 -
### External Comment ###
I'm marking this bug as "new" again because of the package change (from docker.io to lvm2), to allow it to be triaged again in this new context.
--
You received this bug notification because you are a member of Ubuntu
Foundations Bugs, which is subscribed to lvm2 in Ubuntu.
https://bugs.launchpad.net/bugs/1657646
Title:
Missing thin-provisioning-tools prevents VG with thin pool LV from
being (de)activated, but not its creation
Status in lvm2 package in Ubuntu:
New
Status in lvm2 package in Debian:
New
Bug description:
Creating a thin pool LV is allowed even when thin-provisioning-tools
is not installed. But deactivating or activating that VG fails. Since
deactivating the VG usually only happens at reboot, the user might
fail to notice this big problem until then.
I think the lvconvert tool, used to combine the two "thin LVs" into a
thin pool LV, should refuse to run if thin-provisioning-tools, or the
needed scripts, aren't installed.
Steps to reproduce:
root at 15-89:~# vgcreate vg /dev/vdb1
Volume group "vg" successfully created
root at 15-89:~# vgs
VG #PV #LV #SN Attr VSize VFree
vg 1 0 0 wz--n- 40.00g 40.00g
root at 15-89:~# lvcreate -n pool0 -l 90%VG vg
Logical volume "pool0" created.
root at 15-89:~# lvcreate -n pool0meta -l 5%VG vg
Logical volume "pool0meta" created.
root at 15-89:~# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
pool0 vg -wi-a----- 36.00g
pool0meta vg -wi-a----- 2.00g
root at 15-89:~# ll /dev/mapper/
total 0
drwxr-xr-x 2 root root 100 Jun 21 14:15 ./
drwxr-xr-x 20 root root 3820 Jun 21 14:15 ../
crw------- 1 root root 10, 236 Jun 21 13:15 control
lrwxrwxrwx 1 root root 7 Jun 21 14:14 vg-pool0 -> ../dm-0
lrwxrwxrwx 1 root root 7 Jun 21 14:15 vg-pool0meta -> ../dm-1
root at 15-89:~# lvconvert --type thin-pool --poolmetadata vg/pool0meta vg/pool0
WARNING: Converting logical volume vg/pool0 and vg/pool0meta to pool's data and metadata volumes.
THIS WILL DESTROY CONTENT OF LOGICAL VOLUME (filesystem etc.)
Do you really want to convert vg/pool0 and vg/pool0meta? [y/n]: y
Converted vg/pool0 to thin pool.
root at 15-89:~# ll /dev/mapper/
total 0
drwxr-xr-x 2 root root 120 Jun 21 14:15 ./
drwxr-xr-x 20 root root 3840 Jun 21 14:15 ../
crw------- 1 root root 10, 236 Jun 21 13:15 control
lrwxrwxrwx 1 root root 7 Jun 21 14:15 vg-pool0 -> ../dm-2
lrwxrwxrwx 1 root root 7 Jun 21 14:15 vg-pool0_tdata -> ../dm-1
lrwxrwxrwx 1 root root 7 Jun 21 14:15 vg-pool0_tmeta -> ../dm-0
root at 15-89:~# lvs -a
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%S
ync Convert
[lvol0_pmspare] vg ewi------- 2.00g
pool0 vg twi-a-tz-- 36.00g 0.00 0.01
[pool0_tdata] vg Twi-ao---- 36.00g
[pool0_tmeta] vg ewi-ao---- 2.00g
If you now reboot the system, all that is gone:
root at 15-89:~# ll /dev/mapper/
total 0
drwxr-xr-x 2 root root 60 Jun 21 14:28 ./
drwxr-xr-x 19 root root 3760 Jun 21 14:28 ../
crw------- 1 root root 10, 236 Jun 21 14:28 control
The same happens if you deactivate the VG (which the reboot
undoubtedly triggers). It fails because of a missing
/usr/sbin/thin_check which is provided by the thin-provisioning-tools
package:
root at 15-89:~# vgchange -a n
/usr/sbin/thin_check: execvp failed: No such file or directory
WARNING: Integrity check of metadata for pool vg/pool0 failed.
0 logical volume(s) in volume group "vg" now active
root at 15-89:~# ll /dev/mapper/
total 0
drwxr-xr-x 2 root root 60 Jun 21 14:29 ./
drwxr-xr-x 19 root root 3760 Jun 21 14:29 ../
crw------- 1 root root 10, 236 Jun 21 14:28 control
To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/lvm2/+bug/1657646/+subscriptions
More information about the foundations-bugs
mailing list