Juju Kubernetes vSphere storage
Micheal B
tictoc at tictoc.us
Thu Sep 7 21:02:13 UTC 2017
YUP THAT’S THE ONE I FOLLOWED .. I will go back through it though and make sure I did not miss anything ..
Thanks for all the support! Much appreciated!
Cheers
Micheal
From: Tim Van Steenburgh <tim.van.steenburgh at canonical.com>
Date: Thursday, September 7, 2017 at 2:02 PM
To: Micheal B <tictoc at tictoc.us>
Cc: juju <Juju at lists.ubuntu.com>
Subject: Re: Juju Kubernetes vSphere storage
On Thu, Sep 7, 2017 at 1:31 PM, Micheal B <tictoc at tictoc.us> wrote:
Thanks!
Stuck on
Step-6 Add flags to controller-manager, API server and Kubelet to enable vSphere Cloud Provider. * Add following flags to kubelet running on every node and to the controller-manager and API server pods manifest files.
--cloud-provider=vsphere
--cloud-config=<Path of the vsphere.conf file>
tried this .. did not make a difference
It's difficult to help because I don't know what steps were performed. You may find it helpful to follow along with the steps here: https://github.com/juju-solutions/bundle-canonical-kubernetes/issues/402. This give more detail about configuring the vsphere cloud provider in the context of CDK.
Configuring Masters
Edit or create the master configuration file on all masters (/etc/origin/master/master-config.yaml by default) and update the contents of the apiServerArguments and controllerArguments sections with the following:
kubernetesMasterConfig:
admissionConfig:
pluginConfig:
{}
apiServerArguments:
cloud-provider:
- "vsphere"
cloud-config:
- "/etc/vsphere/vsphere.conf"
controllerArguments:
cloud-provider:
- "vsphere"
cloud-config:
- "/etc/vsphere/vsphere.conf"
When triggering a containerized installation, only the /etc/origin and /var/lib/origin directories are mounted to the master and node container. Therefore, master-config.yaml must be in /etc/origin/master rather than /etc/.
Configuring Nodes
Edit or create the node configuration file on all nodes (/etc/origin/node/node-config.yaml by default) and update the contents of the kubeletArguments section:
kubeletArguments:
cloud-provider:
- "vsphere"
cloud-config:
- "/etc/vsphere/vsphere.conf"
When triggering a containerized installation, only the /etc/origin and /var/lib/origin directories are mounted to the master and node container. Therefore, node-config.yaml must be in /etc/origin/node rather than /etc/.
From: Tim Van Steenburgh <tim.van.steenburgh at canonical.com>
Date: Thursday, September 7, 2017 at 6:33 AM
To: Micheal B <tictoc at tictoc.us>
Cc: juju <Juju at lists.ubuntu.com>
Subject: Re: Juju Kubernetes vSphere storage
Hi Micheal,
Have you enabled the vsphere cloud provider for kubernetes as documented here: https://kubernetes.io/docs/getting-started-guides/vsphere/ ?
Tim
On Thu, Sep 7, 2017 at 4:06 AM, Micheal B <tictoc at tictoc.us> wrote:
While working through -- https://github.com/kubernetes/examples/tree/master/staging/volumes/vsphere
To test the different storage types on my vSphere lab I seem to either have a bug or am no able to copy and paste some code ☺
None work. All get pretty close to the same error.
MountVolume.SetUp failed for volume "test-volume" : mount failed: exit status 32 Mounting command: mount Mounting arguments: /var/lib/kubelet/plugins/kubernetes.io/vsphere-volume/mounts/[DS_TICTOC01] volumes/myDisk /var/lib/kubelet/pods/aa94ec10-9349-11e7-a663-005056a192ad/volumes/kubernetes.io~vsphere-volume/test-volume [bind] Output: mount: special device /var/lib/kubelet/plugins/kubernetes.io/vsphere-volume/mounts/[DS_TICTOC01] volumes/myDisk does not exist
Unable to mount volumes for pod "test-vmdk_default(aa94ec10-9349-11e7-a663-005056a192ad)": timeout expired waiting for volumes to attach/mount for pod "default"/"test-vmdk". list of unattached/unmounted volumes=[test-volume]
The volume is there /volume/myDisk.vmdk for the first test and the auto create volume also fails. Tested using the paths
>From datastore cluster + datastore to /vmfs/volumes/55b828da-b978a6d4-6619-002655e59984/volumes
datastore: DS_TICTOC01/volumes
datastore: ticsdata/DS_TICTOC01/volumes
My user making the connection is the default Admininistrator and have used it for years to create other assorted vm’s so I know that good. No error in vsphere vcenter either.
I am using vSphere 6.1 / Kubernetes 1.7 / vSphere DRS is enabled using local drives to create the DS Cluster. JUJU had no issues deploying to them.
What am I missing or could try?
Cheers
Micheal
--
Juju mailing list
Juju at lists.ubuntu.com
Modify settings or unsubscribe at: https://lists.ubuntu.com/mailman/listinfo/juju
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.ubuntu.com/archives/juju/attachments/20170907/4ed1a6bc/attachment.html>
More information about the Juju
mailing list