Friday, 2 March 2018

IBM Cloud Private - Helm not happy again

This follows on from yesterday's thread: -


Today, after booting up my ICP VMs ( Master/Boot, Worker and Proxy ), I see this in the GUI: -



and: -


and: -

and: -


all of which makes me think that Helm API isn't happy again.

Digging into the Helm API Pod: -






So I dived onto the command-line: -

docker ps -a|grep -i helm-api

f9ee8a0e5d20        b72c1d4155b8                      "npm start"              2 minutes ago       Exited (0) About a minute ago                       k8s_helmapi_helm-api-5874f9d746-9qcjg_kube-system_7122aa29-1bcb-11e8-ab0b-000c290f4d7f_60
de1a7b2d1bf2        ibmcom/pause:3.0                  "/pause"                 2 hours ago         Up 2 hours                                          k8s_POD_helm-api-5874f9d746-9qcjg_kube-system_7122aa29-1bcb-11e8-ab0b-000c290f4d7f_8

which makes me think that the first container is having a bad day.

I bounced it: -

docker restart f9ee8a0e5d20

and watched the logs: -

docker logs f9ee8a0e5d20 -f

> helmApi@0.0.0 start /usr/src/app
> node ./bin/www

2018-03-02T11:15:31.070Z 'FINE' 'HELM_REPOS  helm_repos'
2018-03-02T11:15:31.074Z 'FINE' 'process.env.DBHOST https://cloudantdb:6984'
2018-03-02T11:15:31.075Z 'FINE' 'db_host https://cloudantdb:6984'
2018-03-02T11:15:31.318Z 'INFO' 'dbutils Authenticating with the database host \'https://cloudantdb:6984\'...'
D0302 11:15:31.459818780      15 env_linux.c:66]             Warning: insecure environment read function 'getenv' used
2018-03-02T11:15:31.503Z 'FINE' 'Tiller url  tiller-deploy.kube-system:44134'
2018-03-02T11:15:31.503Z 'FINE' 'eval ISICP true'
2018-03-02T11:15:31.507Z 'INFO' 'dbutils createMultipleDb(helm_repos)'
2018-03-02T11:15:31.509Z 'FINE' 'ABOUT TO START SYNCH'
2018-03-02T11:15:31.509Z 'FINE' 'startInitialSynch'
2018-03-02T11:15:33.509Z 'INFO' 'dbutils  getDbConnection(\'helm_repos\') Initializing connection to database \'helm_repos\''
2018-03-02T11:15:36.549Z 'WARN' 'dbutils Using db with no auth'
2018-03-02T11:15:36.550Z 'INFO' 'dbutilsPOST Headers using db with no auth'
2018-03-02T11:15:36.551Z 'INFO' 'dbutils Initial DB connections tested.'
2018-03-02T11:15:37.513Z 'INFO' 'dbutils createMultipleDb(dbNameArr, arrIndex)'
2018-03-02T11:15:37.513Z 'INFO' 'dbutils createDb helm_repos'
2018-03-02T11:15:42.536Z 'INFO' 'dbutils addViewsToDb'
2018-03-02T11:15:52.579Z 'INFO' 'dbutils createMultipleDb(dbNameArr, arrIndex)'

> helmApi@0.0.0 start /usr/src/app
> node ./bin/www

2018-03-02T11:18:54.589Z 'FINE' 'HELM_REPOS  helm_repos'
2018-03-02T11:18:54.593Z 'FINE' 'process.env.DBHOST https://cloudantdb:6984'
2018-03-02T11:18:54.595Z 'FINE' 'db_host https://cloudantdb:6984'
2018-03-02T11:18:55.378Z 'INFO' 'dbutils Authenticating with the database host \'https://cloudantdb:6984\'...'
D0302 11:18:55.643869950      15 env_linux.c:66]             Warning: insecure environment read function 'getenv' used
2018-03-02T11:18:55.730Z 'FINE' 'Tiller url  tiller-deploy.kube-system:44134'
2018-03-02T11:18:55.730Z 'FINE' 'eval ISICP true'
2018-03-02T11:18:55.744Z 'INFO' 'dbutils createMultipleDb(helm_repos)'
2018-03-02T11:18:55.745Z 'FINE' 'ABOUT TO START SYNCH'
2018-03-02T11:18:55.745Z 'FINE' 'startInitialSynch'
2018-03-02T11:18:57.745Z 'INFO' 'dbutils  getDbConnection(\'helm_repos\') Initializing connection to database \'helm_repos\''
2018-03-02T11:19:00.809Z 'WARN' 'dbutils Using db with no auth'
2018-03-02T11:19:00.814Z 'INFO' 'dbutilsPOST Headers using db with no auth'
2018-03-02T11:19:00.815Z 'INFO' 'dbutils Initial DB connections tested.'
2018-03-02T11:19:01.745Z 'INFO' 'dbutils createMultipleDb(dbNameArr, arrIndex)'
2018-03-02T11:19:01.746Z 'INFO' 'dbutils createDb helm_repos'
2018-03-02T11:19:06.761Z 'INFO' 'dbutils addViewsToDb'
2018-03-02T11:19:16.781Z 'INFO' 'dbutils createMultipleDb(dbNameArr, arrIndex)'

but things don't appear to be any better: -

docker ps -a|grep -i helm-api

a6616b26719e        b72c1d4155b8                      "npm start"              2 minutes ago       Exited (0) About a minute ago                       k8s_helmapi_helm-api-5874f9d746-9qcjg_kube-system_7122aa29-1bcb-11e8-ab0b-000c290f4d7f_61
de1a7b2d1bf2        ibmcom/pause:3.0                  "/pause"                 2 hours ago         Up 2 hours                                          k8s_POD_helm-api-5874f9d746-9qcjg_kube-system_7122aa29-1bcb-11e8-ab0b-000c290f4d7f_8

( Note that the container ID has changed )

I checked the logs for this new container: -

docker logs a6616b26719e -f

> helmApi@0.0.0 start /usr/src/app
> node ./bin/www

2018-03-02T11:19:24.518Z 'FINE' 'HELM_REPOS  helm_repos'
2018-03-02T11:19:24.529Z 'FINE' 'process.env.DBHOST https://cloudantdb:6984'
2018-03-02T11:19:24.530Z 'FINE' 'db_host https://cloudantdb:6984'
2018-03-02T11:19:24.981Z 'INFO' 'dbutils Authenticating with the database host \'https://cloudantdb:6984\'...'
D0302 11:19:25.206140919      15 env_linux.c:66]             Warning: insecure environment read function 'getenv' used
2018-03-02T11:19:25.287Z 'FINE' 'Tiller url  tiller-deploy.kube-system:44134'
2018-03-02T11:19:25.287Z 'FINE' 'eval ISICP true'
2018-03-02T11:19:25.291Z 'INFO' 'dbutils createMultipleDb(helm_repos)'
2018-03-02T11:19:25.292Z 'FINE' 'ABOUT TO START SYNCH'
2018-03-02T11:19:25.293Z 'FINE' 'startInitialSynch'
2018-03-02T11:19:27.294Z 'INFO' 'dbutils  getDbConnection(\'helm_repos\') Initializing connection to database \'helm_repos\''
2018-03-02T11:19:30.322Z 'WARN' 'dbutils Using db with no auth'
2018-03-02T11:19:30.322Z 'INFO' 'dbutilsPOST Headers using db with no auth'
2018-03-02T11:19:30.323Z 'INFO' 'dbutils Initial DB connections tested.'
2018-03-02T11:19:31.296Z 'INFO' 'dbutils createMultipleDb(dbNameArr, arrIndex)'
2018-03-02T11:19:31.296Z 'INFO' 'dbutils createDb helm_repos'
2018-03-02T11:19:36.314Z 'INFO' 'dbutils addViewsToDb'
2018-03-02T11:19:46.335Z 'INFO' 'dbutils createMultipleDb(dbNameArr, arrIndex)'

so no obvious exception …..

And yet …..

I dug further, looking specifically at the Kubernetes pods in the kube-system namespace ( rather than the default namespace where my "user" workloads reside )

kubectl get pods --namespace kube-system

NAME                                                      READY     STATUS             RESTARTS   AGE
auth-apikeys-fxlxs                                        1/1       Running            5          2d
auth-idp-wwhlw                                            3/3       Running            16         2d
auth-pap-2x8fs                                            1/1       Running            5          2d
auth-pdp-bpk7h                                            1/1       Running            5          2d
calico-node-amd64-ls5pg                                   2/2       Running            18         2d
calico-node-amd64-pk49d                                   2/2       Running            16         2d
calico-node-amd64-qwt42                                   2/2       Running            17         2d
calico-policy-controller-5997c6c956-m9nmd                 1/1       Running            6          2d
catalog-catalog-apiserver-6mgbq                           1/1       Running            17         2d
catalog-catalog-controller-manager-bd9f49c8c-f625b        1/1       Running            23         2d
catalog-ui-dgcq2                                          1/1       Running            6          2d
default-http-backend-8448fbc655-fnqv2                     1/1       Running            1          1d
elasticsearch-client-6c9fc8b5b6-h87wr                     2/2       Running            12         2d
elasticsearch-data-0                                      1/1       Running            0          2h
elasticsearch-master-667485dfc5-bppms                     1/1       Running            6          2d
filebeat-ds-amd64-2h55z                                   1/1       Running            7          2d
filebeat-ds-amd64-dz2dz                                   1/1       Running            7          2d
filebeat-ds-amd64-pnjzf                                   1/1       Running            7          2d
heapster-5fd94775d5-8tjb9                                 2/2       Running            13         2d
helm-api-5874f9d746-9qcjg                                 0/1       CrashLoopBackOff   66         2d
helmrepo-77dccffb66-9xwgd                                 0/1       Running            32         2d
icp-ds-0                                                  1/1       Running            6          2d
icp-router-86p4z                                          1/1       Running            25         2d
image-manager-0                                           2/2       Running            13         2d
k8s-etcd-192.168.1.100                                    1/1       Running            7          2d
k8s-mariadb-192.168.1.100                                 1/1       Running            7          2d
k8s-master-192.168.1.100                                  3/3       Running            22         2d
k8s-proxy-192.168.1.100                                   1/1       Running            7          2d
k8s-proxy-192.168.1.101                                   1/1       Running            8          2d
k8s-proxy-192.168.1.102                                   1/1       Running            7          2d
kube-dns-9494dc977-8tkmv                                  3/3       Running            20         2d
logstash-5ccb9849d6-z9ntw                                 1/1       Running            7          2d
metering-dm-8587b865b4-ng6rc                              1/1       Running            8          2d
metering-reader-amd64-wtk6s                               1/1       Running            9          2d
metering-reader-amd64-xp5h2                               1/1       Running            10         2d
metering-reader-amd64-z2s9j                               1/1       Running            11         2d
metering-server-748d8f8f5b-x57fs                          1/1       Running            7          2d
metering-ui-6c56c5778f-xnnfr                              1/1       Running            11         2d
monitoring-exporter-76b94fdd94-djr96                      1/1       Running            8          2d
monitoring-grafana-5c49f54dd-w498k                        2/2       Running            15         2d
monitoring-prometheus-77d4df9dd6-zqtt5                    3/3       Running            22         2d
monitoring-prometheus-alertmanager-564496655f-np9hn       3/3       Running            22         2d
monitoring-prometheus-kubestatemetrics-776b5dcb86-r6jmg   1/1       Running            8          2d
monitoring-prometheus-nodeexporter-amd64-9vhn6            1/1       Running            9          2d
monitoring-prometheus-nodeexporter-amd64-lr8x5            1/1       Running            7          2d
monitoring-prometheus-nodeexporter-amd64-pbzk8            1/1       Running            8          2d
nginx-ingress-lb-amd64-8jjtz                              1/1       Running            15         2d
platform-api-bkgs8                                        1/1       Running            6          2d
platform-ui-qmtg7                                         1/1       Running            12         2d
rescheduler-xn7jf                                         1/1       Running            6          2d
tiller-deploy-55fb4d8dcc-2b75v                            1/1       Running            7          2d
unified-router-np97k                                      1/1       Running            13         2d


and then looked at the logs for that particular offending pod: -

kubectl logs helm-api-5874f9d746-9qcjg -f --namespace kube-system

2018-03-02T11:37:29.661Z 'FINE' 'GET /healthcheck'
2018-03-02T11:37:29.662Z 'FINE' 'dbHealthcheck \nrepoName: ibm-charts\n'
2018-03-02T11:37:29.908Z 'FINE' 'getMessage ["statusCode",200] en '
2018-03-02T11:37:29.914Z 'FINE' 'loadMessages en'
GET /healthcheck 200 252.498 ms - 16
2018-03-02T11:37:36.088Z 'FINE' 'GET /healthcheck'
2018-03-02T11:37:36.089Z 'FINE' 'getMessage ["statusCode",200] en '
2018-03-02T11:37:36.089Z 'FINE' 'loadMessages en'
GET /healthcheck 200 1.314 ms - 16
2018-03-02T11:37:46.088Z 'FINE' 'GET /healthcheck'
2018-03-02T11:37:46.089Z 'FINE' 'getMessage ["statusCode",200] en '
2018-03-02T11:37:46.089Z 'FINE' 'loadMessages en'
GET /healthcheck 200 0.439 ms - 16
2018-03-02T11:37:56.088Z 'FINE' 'GET /healthcheck'
2018-03-02T11:37:56.089Z 'FINE' 'getMessage ["statusCode",200] en '
2018-03-02T11:37:56.090Z 'FINE' 'loadMessages en'
GET /healthcheck 200 3.081 ms - 16
2018-03-02T11:37:58.656Z 'FINE' 'GET /healthcheck'
2018-03-02T11:37:58.656Z 'FINE' 'dbHealthcheck \nrepoName: ibm-charts\n'
2018-03-02T11:37:58.669Z 'FINE' 'getMessage ["statusCode",200] en '
2018-03-02T11:37:58.670Z 'FINE' 'loadMessages en'
GET /healthcheck 200 14.629 ms - 16
2018-03-02T11:38:06.107Z 'FINE' 'GET /healthcheck'
2018-03-02T11:38:06.107Z 'FINE' 'getMessage ["statusCode",200] en '
2018-03-02T11:38:06.108Z 'FINE' 'loadMessages en'
GET /healthcheck 200 0.393 ms - 16


However, after a bit of time ( and patience ), we have this: -

kubectl get pods --namespace kube-system
 
NAME                                                      READY     STATUS    RESTARTS   AGE
auth-apikeys-fxlxs                                        1/1       Running   5          2d
auth-idp-wwhlw                                            3/3       Running   16         2d
auth-pap-2x8fs                                            1/1       Running   5          2d
auth-pdp-bpk7h                                            1/1       Running   5          2d
calico-node-amd64-ls5pg                                   2/2       Running   18         2d
calico-node-amd64-pk49d                                   2/2       Running   16         2d
calico-node-amd64-qwt42                                   2/2       Running   17         2d
calico-policy-controller-5997c6c956-m9nmd                 1/1       Running   6          2d
catalog-catalog-apiserver-6mgbq                           1/1       Running   17         2d
catalog-catalog-controller-manager-bd9f49c8c-f625b        1/1       Running   23         2d
catalog-ui-dgcq2                                          1/1       Running   6          2d
default-http-backend-8448fbc655-fnqv2                     1/1       Running   1          1d
elasticsearch-client-6c9fc8b5b6-h87wr                     2/2       Running   12         2d
elasticsearch-data-0                                      1/1       Running   0          2h
elasticsearch-master-667485dfc5-bppms                     1/1       Running   6          2d
filebeat-ds-amd64-2h55z                                   1/1       Running   7          2d
filebeat-ds-amd64-dz2dz                                   1/1       Running   7          2d
filebeat-ds-amd64-pnjzf                                   1/1       Running   7          2d
heapster-5fd94775d5-8tjb9                                 2/2       Running   13         2d
helm-api-5874f9d746-9qcjg                                 1/1       Running   67         2d
helmrepo-77dccffb66-9xwgd                                 1/1       Running   33         2d
icp-ds-0                                                  1/1       Running   6          2d
icp-router-86p4z                                          1/1       Running   25         2d
image-manager-0                                           2/2       Running   13         2d
k8s-etcd-192.168.1.100                                    1/1       Running   7          2d
k8s-mariadb-192.168.1.100                                 1/1       Running   7          2d
k8s-master-192.168.1.100                                  3/3       Running   22         2d
k8s-proxy-192.168.1.100                                   1/1       Running   7          2d
k8s-proxy-192.168.1.101                                   1/1       Running   8          2d
k8s-proxy-192.168.1.102                                   1/1       Running   7          2d
kube-dns-9494dc977-8tkmv                                  3/3       Running   20         2d
logstash-5ccb9849d6-z9ntw                                 1/1       Running   7          2d
metering-dm-8587b865b4-ng6rc                              1/1       Running   8          2d
metering-reader-amd64-wtk6s                               1/1       Running   9          2d
metering-reader-amd64-xp5h2                               1/1       Running   10         2d
metering-reader-amd64-z2s9j                               1/1       Running   11         2d
metering-server-748d8f8f5b-x57fs                          1/1       Running   7          2d
metering-ui-6c56c5778f-xnnfr                              1/1       Running   11         2d
monitoring-exporter-76b94fdd94-djr96                      1/1       Running   8          2d
monitoring-grafana-5c49f54dd-w498k                        2/2       Running   15         2d
monitoring-prometheus-77d4df9dd6-zqtt5                    3/3       Running   22         2d
monitoring-prometheus-alertmanager-564496655f-np9hn       3/3       Running   22         2d
monitoring-prometheus-kubestatemetrics-776b5dcb86-r6jmg   1/1       Running   8          2d
monitoring-prometheus-nodeexporter-amd64-9vhn6            1/1       Running   9          2d
monitoring-prometheus-nodeexporter-amd64-lr8x5            1/1       Running   7          2d
monitoring-prometheus-nodeexporter-amd64-pbzk8            1/1       Running   8          2d
nginx-ingress-lb-amd64-8jjtz                              1/1       Running   15         2d
platform-api-bkgs8                                        1/1       Running   6          2d
platform-ui-qmtg7                                         1/1       Running   12         2d
rescheduler-xn7jf                                         1/1       Running   6          2d
tiller-deploy-55fb4d8dcc-2b75v                            1/1       Running   7          2d
unified-router-np97k                                      1/1       Running   13         2d




I still have one unhealthy deployment: -


but that's my DataPower pod: -

kubectl get pods --namespace default

NAME                                                READY     STATUS             RESTARTS   AGE
davesdatapower-ibm-datapower-dev-57f6cf4c95-bl7sk   0/1       CrashLoopBackOff   39         20h
virtuous-joey-ibm-open-l-6db4566d6d-zbgf9           1/1       Running            1          17h

kubectl logs davesdatapower-ibm-datapower-dev-57f6cf4c95-bl7sk -f

20180302T113711.926Z [0x8040006b][system][notice] logging target(default-log): Logging started.
20180302T113712.167Z [0x804000fb][system][error] : Incorrect number of CPUs. Expected minimum is 2, but have 1.
20180302T113712.167Z [0x804000fe][system][notice] : Container instance UUID: 807dada3-37e8-4dea-938d-443e351cd96e, Cores: 1, vCPUs: 1, CPU model: Intel(R) Core(TM) i7-4790 CPU @ 3.60GHz, Memory: 7954.3MB, Platform: docker, OS: dpos, Edition: developers-limited, Up time: 0 minutes
20180302T113712.171Z [0x8040001c][system][notice] : DataPower IDG is on-line.
20180302T113712.172Z [0x8100006f][system][notice] : Executing default startup configuration.
20180302T113712.464Z [0x8100006d][system][notice] : Executing system configuration.
20180302T113712.465Z [0x8100006b][mgmt][notice] domain(default): tid(8175): Domain operational state is up.
davesdatapower-ibm-datapower-dev-57f6cf4c95-bl7sk
Unauthorized access prohibited.
20180302T113715.531Z [0x806000dd][system][notice] cert-monitor(Certificate Monitor): tid(399): Enabling Certificate Monitor to scan once every 1 days for soon to expire certificates
20180302T113716.388Z [0x8100006e][system][notice] : Executing startup configuration.
20180302T113716.409Z [0x8040009f][system][notice] throttle(Throttler): tid(1391): Disabling throttle.
20180302T113716.447Z [0x00350015][mgmt][notice] b2b-persistence(B2BPersistence): tid(111): Operational state down
20180302T113716.663Z [0x0034000d][mgmt][warn] ssh(SSH Service): tid(111): Object is disabled
20180302T113717.447Z [0x00350015][mgmt][notice] smtp-server-connection(default): tid(7055): Operational state down
login: 20180302T113717.447Z [0x00350014][mgmt][notice] smtp-server-connection(default): tid(7055): Operational state up
20180302T113717.546Z [0x0035008f][mgmt][notice] quota-enforcement-server(QuotaEnforcementServer): tid(687): Operational state down pending
20180302T113717.605Z [0x00350014][mgmt][notice] web-mgmt(WebGUI-Settings): tid(303): Operational state up
20180302T113717.671Z [0x8100006b][mgmt][notice] domain(webApplicationProxy): tid(29615): Domain operational state is up.
20180302T113717.683Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'local:'
20180302T113717.683Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'logtemp:'
20180302T113717.683Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'logstore:'
20180302T113717.683Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'temporary:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'export:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'chkpoints:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'policyframework:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'dpnfsstatic:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'dpnfsauto:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'ftp-response:'
20180302T113717.684Z [0x8100001f][mgmt][notice] domain(webApplicationProxy): tid(29615): Created domain folder 'xm70store:'
20180302T113717.712Z [0x8100003b][mgmt][notice] domain(default): Domain configured successfully.
20180302T113718.795Z [webApplicationProxy][0x8040006b][system][notice] logging target(default-log): tid(111): Logging started.
20180302T113721.374Z [webApplicationProxy][0x00330019][mgmt][error] source-https(webApplicationProxy_Web_HTTPS): tid(111): Operation state transition to up failed
20180302T113721.387Z [webApplicationProxy][0x00350015][mgmt][notice] smtp-server-connection(default): tid(35151): Operational state down
20180302T113721.387Z [webApplicationProxy][0x00350014][mgmt][notice] smtp-server-connection(default): tid(35151): Operational state up
20180302T113721.476Z [webApplicationProxy][0x00350016][mgmt][notice] source-https(webApplicationProxy_Web_HTTPS): tid(111): Service installed on port
20180302T113721.476Z [webApplicationProxy][0x00350014][mgmt][notice] source-https(webApplicationProxy_Web_HTTPS): tid(111): Operational state up
20180302T113721.476Z [webApplicationProxy][0x00350014][mgmt][notice] mpgw(webApplicationProxy): tid(111): Operational state up
20180302T113721.531Z [0x8100003b][mgmt][notice] domain(webApplicationProxy): Domain configured successfully.
20180302T113800.626Z [0x80e0047a][system][error] : tid(176): DataPower QuotaEnforcement task is not responding, restart in progress
20180302T113804.626Z [0x00350014][mgmt][notice] quota-enforcement-server(QuotaEnforcementServer): tid(687): Operational state up


This pod was deployed using a Helm chart: -

helm list

NAME           REVISION UPDATED                  STATUS   CHART                   NAMESPACE
davesdatapower 1        Thu Mar  1 14:54:21 2018 DEPLOYED ibm-datapower-dev-1.0.4 default  
virtuous-joey  1        Thu Mar  1 18:03:30 2018 DEPLOYED ibm-open-liberty-1.0.0  default  

so I removed the release: -

helm delete --purge davesdatapower

release "davesdatapower" deleted

helm list

NAME          REVISION UPDATED                  STATUS   CHART                  NAMESPACE
virtuous-joey 1        Thu Mar  1 18:03:30 2018 DEPLOYED ibm-open-liberty-1.0.0 default 

and now all looks good: -


kubectl get pods --namespace default

NAME                                        READY     STATUS    RESTARTS   AGE
virtuous-joey-ibm-open-l-6db4566d6d-zbgf9   1/1       Running   1          17h


I've since redeployed my DataPower pod: -

helm install --name davesdatapower -f dp.yaml ibm-charts/ibm-datapower-dev

NAME:   davesdatapower
LAST DEPLOYED: Fri Mar  2 12:07:41 2018
NAMESPACE: default
STATUS: DEPLOYED

RESOURCES:
==> v1/Service
NAME                              CLUSTER-IP  EXTERNAL-IP  PORT(S)         AGE
davesdatapower-ibm-datapower-dev  10.0.0.96   <nodes>      8443:31954/TCP  1s

==> v1beta1/Deployment
NAME                              DESIRED  CURRENT  UP-TO-DATE  AVAILABLE  AGE
davesdatapower-ibm-datapower-dev  1        1        1           0          1s

==> v1/Secret
NAME                                     TYPE    DATA  AGE
davesdatapower-ibm-datapower-dev-secret  Opaque  2     1s

==> v1/ConfigMap
NAME                                     DATA  AGE
davesdatapower-ibm-datapower-dev-config  3     1s


NOTES:

1. Get the application URL by running these commands:
      export NODE_PORT=$(kubectl get --namespace default -o jsonpath="{.spec.ports[0].nodePort}" services davesdatapower-ibm-datapower-dev)
      export NODE_IP=$(kubectl get nodes --namespace default -o jsonpath="{.items[0].status.addresses[0].address}")
      echo https://$NODE_IP:$NODE_PORT

and, as suggested, grabbed the endpoint URL: -

export NODE_PORT=$(kubectl get --namespace default -o jsonpath="{.spec.ports[0].nodePort}" services davesdatapower-ibm-datapower-dev)

export NODE_IP=$(kubectl get nodes --namespace default -o jsonpath="{.items[0].status.addresses[0].address}")

and checked that the pod containing the DataPower instance was up and running: -

kubectl get pods

NAME                                                READY     STATUS    RESTARTS   AGE
davesdatapower-ibm-datapower-dev-57f6cf4c95-9dqwp   1/1       Running   0          2m
virtuous-joey-ibm-open-l-6db4566d6d-zbgf9           1/1       Running   1          18h

and, finally, hit the DP endpoint: -



For the record, DP is acting as a Web Application Proxy (WAP) against IBM.COM :-)

I'm following this recipe: -


on IBM developerWorks.


2 comments:

Claudio Retico said...

Hello Dave, same problem and same "by patience" resolution.

It seems that helm-api receive many timeout error like this:

2018-04-08T09:56:52.779Z 'ERROR' 'downloadAndStore Problem getting response from URL https://kubernetes-charts.storage.googleapis.com/kubernetes-dashboard-0.4.0.tgz {"message":"read ETIMEDOUT","stack":"Error: read ETIMEDOUT\\n at exports._errnoException (util.js:1020:11)\\n at TLSWrap.onread (net.js:568:26)","code":"ETIMEDOUT","errno":"ETIMEDOUT","syscall":"read"}'

I put some other non IBM repository si i think there is some "latency" when a repository refresh occur.

Dave Hay said...

Claudio, thanks for the comments and the feedback. Cheers, Dave

LinuxONE for Dummies

As more companies transform their infrastructures with hybrid cloud services, they require environments that protect the safety of their ...