-
Notifications
You must be signed in to change notification settings - Fork 0
/
Makefile
689 lines (631 loc) · 32.6 KB
/
Makefile
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
# SHELL = bash -xv
TMPDIR ?= /tmp
UNAME = $(shell uname)
UNAME_SHORT = $(shell if [ "$(UNAME)" = "Darwin" ]; then echo 'osx'; else echo 'linux'; fi)
SERVICES = metrics lcm trainer restapi jobmonitor
IMAGES = metrics lcm trainer restapi jobmonitor controller databroker_objectstorage databroker_s3
THIS_DIR := $(dir $(abspath $(lastword $(MAKEFILE_LIST))))
HELM_DEPLOY_DIR := $(dir $(abspath $(lastword $(MAKEFILE_LIST))))helmdeploy
DOCKER_REPO ?= docker.io
DOCKER_REPO_USER ?= user-test
DOCKER_REPO_PASS ?= test
DOCKER_REPO_DIR ?= ~/docker-registry/
DOCKER_NAMESPACE ?= ffdl
DOCKER_PULL_POLICY ?= IfNotPresent
DLAAS_LEARNER_REGISTRY ?= ${DOCKER_REPO}/${DOCKER_NAMESPACE}
INVENTORY ?= ansible/envs/local/minikube.ini
IMAGE_NAME_PREFIX = ffdl-
WHOAMI ?= $(shell whoami)
IMAGE_TAG ?= user-$(WHOAMI)
TEST_SAMPLE ?= tf-model
# VM_TYPE is "vagrant", "minikube" or "none"
VM_TYPE ?= minikube
HAS_STATIC_VOLUMES?=false
TEST_USER = test-user
SET_LOCAL_ROUTES ?= 0
MINIKUBE_RAM ?= 4096
MINIKUBE_CPUS ?= 3
MINIKUBE_DRIVER ?= hyperkit
MINIKUBE_BRIDGE ?= $(shell (ifconfig | grep -e "^bridge100:" || ifconfig | grep -e "^bridge0:") | sed 's/\(.*\):.*/\1/')
UI_REPO = git@github.com:IBM/FfDL-dashboard.git
CLI_CMD = $(shell pwd)/cli/bin/ffdl-$(UNAME_SHORT)
CLUSTER_NAME ?= mycluster
PUBLIC_IP ?= 127.0.0.1
CI_MINIKUBE_VERSION ?= v0.25.1
CI_KUBECTL_VERSION ?= v1.9.4
NAMESPACE ?= default
AWS_ACCESS_KEY_ID ?= test
AWS_SECRET_ACCESS_KEY ?= test
AWS_URL ?= http:\/\/s3\.default\.svc\.cluster\.local
# This will used for "volume.beta.kubernetes.io/storage-class" for the shared volume
# Most likely "standard" in Minikube and "" in DIND, (other value is "default" or "ibmc-s3fs-standard")
# Use export SHARED_VOLUME_STORAGE_CLASS="ibmc-file-gold" for IBM Cloud deployment
ifeq ($(VM_TYPE),minikube)
SHARED_VOLUME_STORAGE_CLASS ?= "standard"
else
SHARED_VOLUME_STORAGE_CLASS ?= ""
endif
IMAGE_DIR := $(IMAGE_NAME)
ifneq ($(filter $(IMAGE_NAME),controller ),)
IMAGE_DIR := lcm/$(IMAGE_NAME)
endif
ifeq ($(IMAGE_NAME),databroker_objectstorage)
IMAGE_DIR := databroker/objectstorage
endif
ifeq ($(IMAGE_NAME),databroker_s3)
IMAGE_DIR := databroker/s3
endif
# Main targets
showvars:
@echo THIS_DIR=$(THIS_DIR); \
echo DEPLOY_DIR=$(DEPLOY_DIR)
usage: ## Show this help
@fgrep -h " ## " $(MAKEFILE_LIST) | fgrep -v fgrep | sed -e 's/\\$$//' | sed -e 's/##//'
build: ## Build the code for all services
build: $(addprefix build-, $(SERVICES))
docker-build-images: $(addprefix docker-build-, $(IMAGES))
docker-build: ## Build the Docker images for all services
docker-build: docker-build-images docker-build-ui docker-build-logcollectors
docker-push:
@if [[ -z "${DOCKER_REPO_USER}" ]] || [[ -z "${DOCKER_REPO_PASS}" ]] ; then \
echo "Please define DOCKER_REPO_USER and DOCKER_REPO_PASS."; \
exit 1; \
else \
if [ "${DOCKER_REPO}" = "docker.io" ]; then \
docker login --username="${DOCKER_REPO_USER}" --password="${DOCKER_REPO_PASS}"; \
for i in $$(docker images --format '{{.Repository}}:{{.Tag}}' | grep ${DOCKER_NAMESPACE} | grep :${IMAGE_TAG} | grep -v '<none>'); do \
echo "docker push $$i"; \
docker push $$i; \
done; \
else \
docker login --username="${DOCKER_REPO_USER}" --password="${DOCKER_REPO_PASS}" https://${DOCKER_REPO}; \
for i in $$(docker images --format '{{.Repository}}:{{.Tag}}' | grep ${DOCKER_REPO}/${DOCKER_NAMESPACE} | grep :${IMAGE_TAG} | grep -v '<none>'); do \
echo "docker push $$i"; \
docker push $$i; \
done; \
fi; \
fi;
# TODO: setup-registry
create-registry:
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@kubectl create secret docker-registry regcred --docker-server=${DOCKER_REPO} --docker-username="${DOCKER_REPO_USER}" --docker-password="${DOCKER_REPO_PASS}" --docker-email=unknown@docker.io ; \
cd ${DOCKER_REPO_DIR} ; \
docker-compose up -d
# TODO Need to iterate over PVCs - currently only deletes static-volume-1
create-volumes:
@kubectl delete --ignore-not-found=true pv/nfs pvc/static-volume-1 cm/static-volumes
@cd bin; \
./create_static_pv.sh; \
./create_static_volumes.sh; \
./create_static_volumes_config.sh
deploy-plugin:
@# deploy the stack via helm
@echo Deploying services to Kubernetes. This may take a while.
@if ! helm list > /dev/null 2>&1; then \
echo 'Installing helm/tiller'; \
helm init; \
sleep 5; \
echo "Waiting tiller to be ready"; \
while ! (kubectl get pods --all-namespaces | grep tiller-deploy | grep '1/1' > /dev/null); \
do \
sleep 1; \
done; \
fi;
@existingPlugin=$$(helm list | grep ibmcloud-object-storage-plugin | awk '{print $$1}' | head -n 1);
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@if [ "$(VM_TYPE)" = "dind" ]; then \
export FFDL_PATH=$$(pwd); \
./bin/s3_driver.sh; \
sleep 10; \
(if [ -z "$$existingPlugin" ]; then \
helm install --set dind=true,cloud=false,namespace=$$NAMESPACE storage-plugin; \
else \
helm upgrade --set dind=true,cloud=false,namespace=$$NAMESPACE $$existingPlugin storage-plugin; \
fi) & pid=$$!; \
else \
(if [ -z "$$existingPlugin" ]; then \
helm install --set namespace=$$NAMESPACE storage-plugin; \
else \
helm upgrade --set namespace=$$NAMESPACE $$existingPlugin storage-plugin; \
fi) & pid=$$!; \
fi;
@echo "Wait while kubectl get pvc shows static-volume-1 in state Pending"
@./bin/create_static_volumes.sh
@./bin/create_static_volumes_config.sh
@sleep 3
quickstart-deploy:
@echo "collecting existing pods"
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@while kubectl get pods --all-namespaces | \
grep -v RESTARTS | \
grep -v Running | \
grep 'alertmanager\|etcd0\|lcm\|restapi\|trainer\|trainingdata\|ui\|mongo\|prometheus\|pushgateway\|storage' > /dev/null; \
do \
sleep 1; \
done
@echo "calling big command from quickstart-deploy "
@sleep 5;
@set -o verbose; \
existing=$$(helm list | grep ffdl | awk '{print $$1}' | head -n 1); \
(if [ -z "$$existing" ]; then \
echo "Deploying the stack via Helm. This will take a while."; \
helm install --set lcm.shared_volume_storage_class=$$SHARED_VOLUME_STORAGE_CLASS,namespace=$$NAMESPACE . ; \
sleep 10; \
else \
echo "Upgrading existing Helm deployment ($$existing). This will take a while."; \
helm upgrade --set lcm.shared_volume_storage_class=$$SHARED_VOLUME_STORAGE_CLASS,namespace=$$NAMESPACE $$existing . ; \
fi) & pid=$$!; \
sleep 5; \
while kubectl get pods --all-namespaces | \
grep -v RESTARTS | \
grep -v Running | \
grep 'alertmanager\|etcd0\|lcm\|restapi\|trainer\|trainingdata\|ui\|mongo\|prometheus\|pushgateway\|storage'; \
do \
sleep 5; \
done; \
existing=$$(helm list | grep ffdl | awk '{print $$1}' | head -n 1); \
for i in $$(seq 1 10); do \
status=`helm status $$existing | grep STATUS:`; \
echo $$status; \
if echo "$$status" | grep "DEPLOYED" > /dev/null; then \
kill $$pid > /dev/null 2>&1; \
exit 0; \
fi; \
sleep 3; \
done; \
exit 0
@echo done with big command
@echo Initializing...
@# wait for pods to be ready
@while kubectl get pods --all-namespaces | \
grep -v RESTARTS | \
grep -v Running | \
grep 'alertmanager\|etcd0\|lcm\|restapi\|trainer\|trainingdata\|ui\|mongo\|prometheus\|pushgateway\|storage' > /dev/null; \
do \
sleep 5; \
done
@if [ "$(VM_TYPE)" = "dind" ]; then \
./bin/dind-port-forward.sh; \
fi;
@echo initialize monitoring dashboards
@if [ "$$CI" != "true" ]; then bin/grafana.init.sh; fi
@echo
@echo System status:
@make status
test-job-submit: ## Submit test training job
@# make sure the buckets with training data exist
@echo Downloading Docker images and test training data. This may take a while.
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@if [ "$(VM_TYPE)" = "minikube" ]; then \
eval $(minikube docker-env); docker images | grep tensorflow | grep latest > /dev/null || docker pull tensorflow/tensorflow > /dev/null; \
fi
@node_ip=$$(make --no-print-directory kubernetes-ip); \
s3_ip=$$(kubectl get pods | grep restapi- | awk '{print $1}' | xargs -I '{}' kubectl get pod '{}' -o jsonpath='{.status.hostIP}'); \
s3_port=$$(kubectl get service s3 -o jsonpath='{.spec.ports[0].nodePort}'); \
s3_url=http://$$s3_ip:$$s3_port; \
s3_raw_url=$$s3_ip:$$s3_port; \
echo "Submitting example training job ($(TEST_SAMPLE))"; \
restapi_port=$$(kubectl get service ffdl-restapi -o jsonpath='{.spec.ports[0].nodePort}'); \
restapi_url=http://$$node_ip:$$restapi_port; \
echo S3 URL: $$s3_url REST URL: $$restapi_url; \
export DLAAS_URL=http://$$node_ip:$$restapi_port; export DLAAS_USERNAME=$(TEST_USER); export DLAAS_PASSWORD=test; \
echo Executing in etc/examples/$(TEST_SAMPLE): DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$$DLAAS_USERNAME DLAAS_PASSWORD=test $(CLI_CMD) train manifest.yml . ; \
cp etc/examples/tf-model/manifest.yml etc/examples/tf-model/manifest_testrun.yml ; \
sed -i '' -e "s/s3.default.svc.cluster.local/$$(kubectl get pods | grep restapi- | awk '{print $$1}' | xargs -I '{}' kubectl get pod '{}' -o jsonpath='{.status.hostIP}'):$$s3_port/g" etc/examples/tf-model/manifest_testrun.yml ; \
cat etc/examples/tf-model/manifest_testrun.yml ; \
(cd etc/examples/$(TEST_SAMPLE); pwd; $(CLI_CMD) train manifest_testrun.yml .); \
rm -f etc/examples/tf-model/manifest_testrun.yml ; \
echo Test job submitted. Track the status via '"'DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$(TEST_USER) DLAAS_PASSWORD=test $(CLI_CMD) list'"'. ; \
sleep 10; \
(for i in $$(seq 1 50); do output=$$($(CLI_CMD) list 2>&1 | grep training-); \
if echo $$output | grep 'FAILED'; then echo 'Job failed'; exit 1; fi; \
if echo $$output | grep 'COMPLETED'; then echo 'Job completed'; exit 0; fi; \
echo $$output; \
sleep 20; \
done; exit 1) || \
($(CLI_CMD) list; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl describe pod '{}'; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}' -c learner; \
exit 1);
deploy: ## Deploy the services to Kubernetes
@docker images
@echo $(DOCKER_REPO)
@echo $(DOCKER_PULL_POLICY)
@# deploy the stack via helm
@echo Deploying services to Kubernetes. This may take a while.
@if ! helm list > /dev/null 2>&1; then \
echo 'Installing helm/tiller'; \
helm init > /dev/null 2>&1; \
sleep 3; \
fi;
@echo collecting existing pods
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@while kubectl get pods --all-namespaces | \
grep -v RESTARTS | \
grep -v Running | \
grep 'alertmanager\|etcd0\|lcm\|restapi\|trainer\|trainingdata\|ui\|mongo\|prometheus\|pushgateway\|storage' > /dev/null; \
do \
sleep 1; \
done
@echo finding out which pods are running
@while ! (kubectl get pods --all-namespaces | grep tiller-deploy | grep '1/1' > /dev/null); \
do \
sleep 1; \
done
@echo calling big command
@set -o verbose; \
echo HELM_DEPLOY_DIR: ${HELM_DEPLOY_DIR}; \
mkdir -p ${HELM_DEPLOY_DIR}; \
cp -rf Chart.yaml values.yaml templates ${HELM_DEPLOY_DIR}; \
existing=$$(helm list | grep ffdl | awk '{print $$1}' | head -n 1); \
if [ "$$CI" = "true" ]; then \
export helm_params='--set lcm.shared_volume_storage_class=${SHARED_VOLUME_STORAGE_CLASS},has_static_volumes=${HAS_STATIC_VOLUMES},namespace=${NAMESPACE},prometheus.deploy=false,learner.docker_namespace=${DOCKER_NAMESPACE},docker.namespace=${DOCKER_NAMESPACE},learner.tag=${IMAGE_TAG},docker.pullPolicy=${DOCKER_PULL_POLICY},docker.registry=${DOCKER_REPO},trainer.version=${IMAGE_TAG},restapi.version=${IMAGE_TAG},lcm.version=${IMAGE_TAG},trainingdata.version=${IMAGE_TAG},databroker.tag=${IMAGE_TAG},databroker.version=${IMAGE_TAG},webui.version=${IMAGE_TAG}'; \
else \
export helm_params='--set lcm.shared_volume_storage_class=${SHARED_VOLUME_STORAGE_CLASS},has_static_volumes=${HAS_STATIC_VOLUMES},namespace=${NAMESPACE},learner.docker_namespace=${DOCKER_NAMESPACE},docker.namespace=${DOCKER_NAMESPACE},learner.tag=${IMAGE_TAG},docker.pullPolicy=${DOCKER_PULL_POLICY},docker.registry=${DOCKER_REPO},trainer.version=${IMAGE_TAG},restapi.version=${IMAGE_TAG},lcm.version=${IMAGE_TAG},trainingdata.version=${IMAGE_TAG},databroker.tag=${IMAGE_TAG},databroker.version=${IMAGE_TAG},webui.version=${IMAGE_TAG}'; \
fi; \
(if [ -z "$$existing" ]; then \
echo "Deploying the stack via Helm. This will take a while."; \
helm --debug install $$helm_params ${HELM_DEPLOY_DIR} ; \
sleep 10; \
else \
echo "Upgrading existing Helm deployment ($$existing). This will take a while."; \
helm --debug upgrade $$helm_params $$existing ${HELM_DEPLOY_DIR} ; \
fi) & pid=$$!; \
sleep 5; \
while kubectl get pods --all-namespaces | \
grep -v RESTARTS | \
grep -v Running | \
grep 'alertmanager\|etcd0\|lcm\|restapi\|trainer\|trainingdata\|ui\|mongo\|prometheus\|pushgateway\|storage'; \
do \
sleep 5; \
done; \
existing=$$(helm list | grep ffdl | awk '{print $$1}' | head -n 1); \
for i in $$(seq 1 10); do \
status=`helm status $$existing | grep STATUS:`; \
echo $$status; \
if echo "$$status" | grep "DEPLOYED" > /dev/null; then \
kill $$pid > /dev/null 2>&1; \
exit 0; \
fi; \
sleep 3; \
done; \
exit 0
@echo done with big command
@echo Initializing...
@# wait for pods to be ready
@while kubectl get pods --all-namespaces | \
grep -v RESTARTS | \
grep -v Running | \
grep 'alertmanager\|etcd0\|lcm\|restapi\|trainer\|trainingdata\|ui\|mongo\|prometheus\|pushgateway\|storage' > /dev/null; \
do \
sleep 5; \
done
@echo initialize monitoring dashboards
@if [ "$$CI" != "true" ]; then bin/grafana.init.sh; fi
@echo
@echo System status:
@make status
undeploy: ## Undeploy the services from Kubernetes
@# undeploy the stack
@existing=$$(helm list | grep ffdl | awk '{print $$1}' | head -n 1); \
(if [ ! -z "$$existing" ]; then echo "Undeploying the stack via helm. This may take a while."; helm delete "$$existing"; echo "The stack has been undeployed."; fi) > /dev/null;
$(addprefix undeploy-, $(SERVICES)): undeploy-%: %
@SERVICE_NAME=$< make .undeploy-service
.undeploy-service:
@echo deleting $(SERVICE_NAME)
(kubectl delete deploy,svc,statefulset --selector=service="ffdl-$(SERVICE_NAME)")
rebuild-and-deploy-lcm: build-lcm docker-build-lcm undeploy-lcm deploy
rebuild-and-deploy-trainer: build-trainer docker-build-trainer undeploy-trainer deploy
status: ## Print the current system status and service endpoints
@tiller=s; \
status_kube=$$(kubectl config current-context) && status_kube="Running (context '$$status_kube')" || status_kube="n/a"; \
echo "Kubernetes:\t$$status_kube"; \
node_ip=$$(make --no-print-directory kubernetes-ip); \
status_tiller=$$(helm list 2> /dev/null) && status_tiller="Running" || status_tiller="n/a"; \
echo "Helm/Tiller:\t$$status_tiller"; \
status_ffdl=$$(helm list 2> /dev/null | grep ffdl | awk '{print $$1}' | head -n 1) && status_ffdl="Running ($$(helm status "$$status_ffdl" 2> /dev/null | grep STATUS:))" || status_ffdl="n/a"; \
echo "FfDL Services:\t$$status_ffdl"; \
port_api=$$(kubectl get service ffdl-restapi -o jsonpath='{.spec.ports[0].nodePort}' 2> /dev/null) && status_api="Running (http://$$node_ip:$$port_api)" || status_api="n/a"; \
echo "REST API:\t$$status_api"; \
port_ui=$$(kubectl get service ffdl-ui -o jsonpath='{.spec.ports[0].nodePort}' 2> /dev/null) && status_ui="Running (http://$$node_ip:$$port_ui/#/login?endpoint=$$node_ip:$$port_api&username=test-user)" || status_ui="n/a"; \
echo "Web UI:\t\t$$status_ui"; \
status_grafana=$$(kubectl get service grafana -o jsonpath='{.spec.ports[0].nodePort}' 2> /dev/null) && status_grafana="Running (http://$$node_ip:$$status_grafana) (login: admin/admin)" || status_grafana="n/a"; \
echo "Grafana:\t$$status_grafana"
vagrant: ## Configure local Kubernetes cluster in Vagrant
vagrant up
minikube: ## Configure Minikube (local Kubernetes)
ifeq ($(filter $(UNAME),Linux Darwin),)
@echo "Sorry, currently only supported for Mac OS and Linux"; exit 1
endif
@which minikube || (echo Please install Minikube; exit 1)
@sudo minikube ip 2>&1 || ( \
echo "Starting up Minikube"; \
sudo minikube start --insecure-registry 9.0.0.0/8 --insecure-registry 10.0.0.0/8 \
--cpus $(MINIKUBE_CPUS) \
--memory $(MINIKUBE_RAM) \
--vm-driver=$(MINIKUBE_DRIVER) --apiserver-ips 127.0.0.1 --apiserver-name localhost ; \
sleep 5; \
)
@set -o verbose; \
sudo minikube ip 2>&1 && ( \
if [ "$(SET_LOCAL_ROUTES)" == "1" ]; then \
echo "Update local network routes, using '$(MINIKUBE_BRIDGE)' as network bridge (you may be prompted for your sudo password)"; \
if [ "$(UNAME)" == "Linux" ]; then \
sudo route -n add -net 10.0.0.0/24 gw $(minikube ip); \
elif [ "$(UNAME)" == "Darwin" ]; then \
sudo route -n delete 10.0.0.0/24 2>&1; \
sudo route -n add 10.0.0.0/24 $$(minikube ip) 2>&1; \
member_interface=$$(ifconfig $(MINIKUBE_BRIDGE) | grep member | awk '{print $$2}' | head -1); \
if [ "$$member_interface" != "" ]; then sudo ifconfig $(MINIKUBE_BRIDGE) -hostfilter $$member_interface; fi; \
fi; \
kubectl get configmap/kube-dns --namespace kube-system -o yaml > /tmp/kube-dns.cfg; \
grep upstreamNameservers /tmp/kube-dns.cfg || (echo 'data:' >> /tmp/kube-dns.cfg; \
echo ' upstreamNameservers: '"'"'["8.8.8.8"]'"'" >> /tmp/kube-dns.cfg); \
kubectl replace -f /tmp/kube-dns.cfg; \
fi; \
)
minikube-undo: ## Revert changes applied by the "minikube" target (e.g., local network routes)
@sudo minikube ip > /dev/null 2>&1 && ( \
sudo route -n delete 10.0.0.0/24 > /dev/null 2>&1; \
) || true
# Build targets
gen-certs:
cd commons/certs ; \
./generate.sh ; \
./install.sh
del-certs:
rm -rf ./metrics/docker/certs; \
rm -rf ./metrics/log_collectors/training_data_service_client/certs; \
rm -rf ./restapi/certs; \
rm -rf ./lcm/certs; \
rm -rf ./jobmonitor/certs; \
rm -rf ./trainer/docker/certs; \
rm -f ./commons/certs/ca.*; \
rm -f ./commons/certs/client.*; \
rm -f ./commons/certs/server.*
$(addprefix build-, $(SERVICES)): build-%: %
@SERVICE_NAME=$< BINARY_NAME=main make .build-service
build-cli:
cd ./cli/ && (CGO_ENABLED=0 go build -ldflags "-s -w" -a -installsuffix cgo -o bin/ffdl-osx; \
CGO_ENABLED=0 GOOS=linux go build -ldflags "-s -w" -a -installsuffix cgo -o bin/ffdl-linux)
build-cli-opt:
make build-cli
cd ./cli/; upx --brute bin/ffdl-osx; \
docker run -v `pwd`:/ffdl-cli ubuntu bash -c 'cd /ffdl-cli; apt-get update; apt-get install -y upx-ucl; upx --brute bin/ffdl-linux'
# Docker build targets
$(addprefix docker-build-, $(IMAGES)): docker-build-%: %
@IMAGE_NAME=$< make .docker-build
docker-build-ui:
mkdir -p build; test -e dashboard || (git clone $(UI_REPO) build/ffdl-ui; ln -s build/ffdl-ui dashboard)
(cd dashboard && (if [ "$(VM_TYPE)" = "minikube" ]; then eval $$(minikube docker-env); fi; \
docker build -q -t $(DOCKER_REPO)/$(DOCKER_NAMESPACE)/$(IMAGE_NAME_PREFIX)ui:$(IMAGE_TAG) .; \
(test ! `which docker-squash` || docker-squash -t $(DOCKER_REPO)/$(DOCKER_NAMESPACE)/$(IMAGE_NAME_PREFIX)ui $(DOCKER_REPO)/$(DOCKER_NAMESPACE)/$(IMAGE_NAME_PREFIX)ui)))
docker-build-base:
if [ "$(VM_TYPE)" = "minikube" ]; then \
eval $$(minikube docker-env); \
fi; \
(cd etc/dlaas-service-base; make build)
docker-build-logcollectors:
if [ "$(VM_TYPE)" = "minikube" ]; then \
eval $$(minikube docker-env); \
fi; \
cd metrics/log_collectors && IMAGE_TAG=$(IMAGE_TAG) && make build-simple_log_collector build-emetrics_file build-regex_extractor build-tensorboard-all
test-push-data-s3: ## Test
@# Pushes test data to S3 buckets.
@echo Pushing test data.
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@s3_ip=$$(make --no-print-directory kubernetes-ip); \
s3_port=$$(kubectl get service s3 -o jsonpath='{.spec.ports[0].nodePort}'); \
s3_url=http://$$s3_ip:$$s3_port; \
export AWS_ACCESS_KEY_ID=test; export AWS_SECRET_ACCESS_KEY=test; export AWS_DEFAULT_REGION=us-east-1; \
s3cmd="aws --endpoint-url=$$s3_url s3"; \
$$s3cmd mb s3://tf_training_data > /dev/null; \
$$s3cmd mb s3://tf_trained_model > /dev/null; \
$$s3cmd mb s3://mnist_lmdb_data > /dev/null; \
$$s3cmd mb s3://dlaas-trained-models > /dev/null; \
for file in t10k-images-idx3-ubyte.gz t10k-labels-idx1-ubyte.gz train-images-idx3-ubyte.gz train-labels-idx1-ubyte.gz; do \
test -e $(TMPDIR)/$$file || wget -q -O $(TMPDIR)/$$file http://yann.lecun.com/exdb/mnist/$$file; \
$$s3cmd cp $(TMPDIR)/$$file s3://tf_training_data/$$file > /dev/null; \
done; \
for phase in train test; do \
for file in data.mdb lock.mdb; do \
tmpfile=$(TMPDIR)/$$phase.$$file; \
test -e $$tmpfile || wget -q -O $$tmpfile https://github.com/albarji/caffe-demos/raw/master/mnist/mnist_"$$phase"_lmdb/$$file; \
$$s3cmd cp $$tmpfile s3://mnist_lmdb_data/$$phase/$$file > /dev/null; \
done; \
done;
@echo "Done uploading data to S3"
test-push-data-hostmount: ## Test
@# Pushes test data to S3 buckets.
@echo Copying test data to host mount area
@s3_ip=$$(kubectl get po/storage-0 -o=jsonpath='{.status.hostIP}'); \
id ; \
echo ${HOME} ; \
sudo mkdir /cosdata; \
sudo mkdir /cosdata/local-dlaas-ci-tf-training-data; \
sudo mkdir /cosdata/mnist_lmdb_data; \
sudo mkdir /cosdata/local-dlaas-ci-trained-results-tf-training-data; \
sudo mkdir /cosdata/local-dlaas-ci-trained-results-tf-training-data/_submitted_code; \
sudo chmod 777 -R /cosdata ; \
sudo apt-get install zip > /dev/null; \
rm /cosdata/local-dlaas-ci-trained-results-tf-training-data/_submitted_code/model.zip > /dev/null; \
(cd etc/examples/$(TEST_SAMPLE); zip -r /cosdata/local-dlaas-ci-trained-results-tf-training-data/_submitted_code/model.zip .); \
for file in t10k-images-idx3-ubyte.gz t10k-labels-idx1-ubyte.gz train-images-idx3-ubyte.gz train-labels-idx1-ubyte.gz; do \
test -e $(TMPDIR)/$$file || wget -q -O $(TMPDIR)/$$file http://yann.lecun.com/exdb/mnist/$$file; \
cp $(TMPDIR)/$$file /cosdata/local-dlaas-ci-tf-training-data; \
done; \
for phase in train test; do \
mkdir /cosdata/mnist_lmdb_data/$$phase; \
for file in data.mdb lock.mdb; do \
tmpfile=$(TMPDIR)/$$phase.$$file; \
test -e $$tmpfile || wget -q -O $$tmpfile https://github.com/albarji/caffe-demos/raw/master/mnist/mnist_"$$phase"_lmdb/$$file; \
cp $$tmpfile /cosdata/mnist_lmdb_data/$$phase; \
done; \
done;
test-submit-minikube-run-test: ## Submit test training job
@echo Downloading Docker images
@if [ "$(VM_TYPE)" = "minikube" ]; then \
eval $(minikube docker-env); docker images | grep tensorflow | grep latest > /dev/null || docker pull tensorflow/tensorflow > /dev/null; \
fi
@node_ip=$$(make --no-print-directory kubernetes-ip); \
echo "Submitting example training job ($(TEST_SAMPLE))"; \
restapi_port=$$(kubectl get service ffdl-restapi -o jsonpath='{.spec.ports[0].nodePort}'); \
restapi_url=http://$$node_ip:$$restapi_port; \
echo REST URL: $$restapi_url; \
export DLAAS_URL=http://$$node_ip:$$restapi_port; export DLAAS_USERNAME=$(TEST_USER); export DLAAS_PASSWORD=test; \
echo Executing in etc/examples/$(TEST_SAMPLE): DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$$DLAAS_USERNAME DLAAS_PASSWORD=test $(CLI_CMD) train manifest.yml . ; \
sudo rm -rf /cosdata/local-dlaas-ci-trained-results-tf-training-data/model; \
sudo rm -rf /cosdata/local-dlaas-ci-trained-results-tf-training-data/learner-1; \
(cd etc/examples/$(TEST_SAMPLE); pwd; $(CLI_CMD) train manifest-hostmount.yml . 2>&1 | tee train.log); \
training_id=$$(cat etc/examples/$(TEST_SAMPLE)/train.log | grep "Model ID" | cut -d " " -f 3); \
echo Training id: $${training_id} ; \
rm -f etc/examples/$(TEST_SAMPLE)/train.log ; \
echo Test job submitted. Track the status via '"'DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$(TEST_USER) DLAAS_PASSWORD=test $(CLI_CMD) list'"'. ; \
(for i in $$(seq 1 500); do output=$$($(CLI_CMD) list 2>&1 | grep $${training_id}); \
if echo $$output | grep 'FAILED'; then echo 'Job failed'; exit 1; fi; \
if echo $$output | grep 'COMPLETED'; then echo 'Job completed'; exit 0; fi; \
echo "kubectl dump objects:"; \
kubectl get pod,pvc,pv,sc,deploy,svc,statefulset,secrets --show-all -o wide ; \
echo "kubectl dump secrets:"; \
kubectl get secrets -o yaml ; \
echo "Dumping Statefulsets" ; \
kubectl get statefulsets | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl get statefulsets '{}' -o yaml; \
echo "================:"; \
echo "LCM:"; \
kubectl logs --selector=service=ffdl-lcm ; \
echo "---------------- previous:"; \
kubectl logs -p --selector=service=ffdl-lcm ; \
echo "---------------- Describe:"; \
kubectl describe pods --selector=service=ffdl-lcm ; \
echo "================:"; \
echo "Trainer:"; \
kubectl get pods | grep trainer- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}'; \
echo "Jobmonitor:"; \
kubectl get pods | grep jobmonitor- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}'; \
echo "Learner:"; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl describe pod '{}'; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}' -c learner; \
echo $$output; \
sleep 60; \
done; exit 1) || \
($(CLI_CMD) list; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl describe pod '{}'; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}' -c learner; \
exit 1);
test-submit-minikube-ci: test-push-data-hostmount test-submit-minikube-run-test
test-submit: ## Submit test training job
@# make sure the buckets with training data exist
@echo Downloading Docker images and test training data. This may take a while.
@if [ "$(VM_TYPE)" = "minikube" ]; then \
eval $(minikube docker-env); docker images | grep tensorflow | grep latest > /dev/null || docker pull tensorflow/tensorflow > /dev/null; \
fi
@kubectl config set-context $$(kubectl config current-context) --namespace=$$NAMESPACE
@node_ip=$$(make --no-print-directory kubernetes-ip); \
s3_ip=$$(kubectl get po/storage-0 -o=jsonpath='{.status.hostIP}'); \
s3_port=$$(kubectl get service s3 -o jsonpath='{.spec.ports[0].nodePort}'); \
./bin/escape_for_sed.sh ${AWS_URL}; \
s3_url=$$(./bin/escape_for_sed.sh ${AWS_URL}); \
echo "s3_url=$${s3_url}"; \
echo "Submitting example training job ($(TEST_SAMPLE))"; \
restapi_port=$$(kubectl get service ffdl-restapi -o jsonpath='{.spec.ports[0].nodePort}'); \
restapi_url=http://$$node_ip:$$restapi_port; \
echo S3 URL: $$s3_url REST URL: $$restapi_url; \
export DLAAS_URL=http://$$node_ip:$$restapi_port; export DLAAS_USERNAME=$(TEST_USER); export DLAAS_PASSWORD=test; \
echo Executing in etc/examples/$(TEST_SAMPLE): DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$$DLAAS_USERNAME DLAAS_PASSWORD=test $(CLI_CMD) train manifest-for-test.yml . ; \
cp etc/examples/tf-model/manifest-for-test.yml etc/examples/tf-model/manifest_testrun.yml ; \
sed -i -e "s/user_name: test/user_name: ${AWS_ACCESS_KEY_ID}/g" etc/examples/tf-model/manifest_testrun.yml ; \
sed -i -e "s/password: test/password: ${AWS_SECRET_ACCESS_KEY}/g" etc/examples/tf-model/manifest_testrun.yml ; \
sed -i "s/auth_url:\stest/auth_url: $${s3_url}/g" etc/examples/tf-model/manifest_testrun.yml ; \
# cat etc/examples/tf-model/manifest_testrun.yml ; \
(cd etc/examples/$(TEST_SAMPLE); pwd; $(CLI_CMD) train manifest_testrun.yml . 2>&1 | tee train.log); \
training_id=$$(cat etc/examples/$(TEST_SAMPLE)/train.log | grep "Model ID" | cut -d " " -f 3); \
echo Training id: $${training_id} ; \
rm -f etc/examples/$(TEST_SAMPLE)/train.log ; \
rm -f etc/examples/$(TEST_SAMPLE)/manifest_testrun.yml ; \
echo Test job submitted. Track the status via '"'DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$(TEST_USER) DLAAS_PASSWORD=test $(CLI_CMD) list'"'. ; \
sleep 10; \
(for i in $$(seq 1 500); do output=$$($(CLI_CMD) list 2>&1 | grep $${training_id}); \
if echo $$output | grep 'FAILED'; then echo 'Job failed'; exit 1; fi; \
if echo $$output | grep 'COMPLETED'; then echo 'Job completed'; exit 0; fi; \
echo $$output; \
sleep 20; \
done; exit 1) || \
($(CLI_CMD) list; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl describe pod '{}'; \
kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}' -c learner; \
exit 1);
test-localmount-submit: ## Submit test training job
@# make sure the buckets with training data exist
@echo Downloading Docker images and test training data. This may take a while.
@if [ "$(VM_TYPE)" = "minikube" ]; \
then \
echo "Checking if we have a tensorflow image, and maybe pulling..."; \
eval $(minikube docker-env); \
docker images | grep tensorflow | grep latest > /dev/null || docker pull tensorflow/tensorflow > /dev/null; \
echo "Done checking if we have a tensorflow image"; \
fi
@node_ip=$$(make --no-print-directory kubernetes-ip); \
echo "Submitting example training job ($(TEST_SAMPLE))"; \
restapi_port=$$(kubectl get service ffdl-restapi -o jsonpath='{.spec.ports[0].nodePort}'); \
export DLAAS_URL=http://$$node_ip:$$restapi_port; export DLAAS_USERNAME=$(TEST_USER); export DLAAS_PASSWORD=test; \
(cd etc/examples/$(TEST_SAMPLE); $(CLI_CMD) train manifest-hostmount.yml .); \
echo Test job submitted. Track the status via '"'DLAAS_URL=$$DLAAS_URL DLAAS_USERNAME=$(TEST_USER) DLAAS_PASSWORD=test $(CLI_CMD) list'"'. ; \
sleep 10; \
for i in $$(seq 1 50); do output=$$($(CLI_CMD) list 2>&1 | grep training-); \
if echo $$output | grep 'FAILED'; then echo 'Job failed'; exit 1; fi; \
if echo $$output | grep 'COMPLETED'; then echo 'Job completed'; exit 0; fi; \
echo $$output; \
sleep 20; \
done;
# || \
# ($(CLI_CMD) list; \
# kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl describe pod '{}'; \
# kubectl get pods | grep learner- | awk '{print $$1}' | xargs -I '{}' kubectl logs '{}' -c learner; \
# exit 0);
# Helper targets
test-s3:
node_ip=$$(make --no-print-directory kubernetes-ip); \
s3_port=$$(kubectl get service s3 -o jsonpath='{.spec.ports[0].nodePort}'); \
s3_url=http://$$node_ip:$$s3_port; \
s3_url=${AWS_URL}; \
export AWS_DEFAULT_REGION=us-west-1; \
s3cmd="aws --endpoint-url=${AWS_URL} s3" ; \
$$s3cmd ls
.build-service:
(cd ./$(SERVICE_NAME)/ && (test ! -e main.go || CGO_ENABLED=0 GOOS=linux go build -ldflags "-s -w" -a -installsuffix cgo -o bin/$(BINARY_NAME)))
.docker-build:
(full_img_name=$(IMAGE_NAME_PREFIX)$(IMAGE_NAME); \
cd ./$(IMAGE_DIR)/ && ( \
if [ "$(VM_TYPE)" = "minikube" ]; then \
eval $$(minikube docker-env); \
fi; \
docker build -q -t $(DOCKER_REPO)/$(DOCKER_NAMESPACE)/$$full_img_name:$(IMAGE_TAG) .))
kubernetes-ip:
@if [ "$$CI" = "true" ]; then kubectl get nodes -o jsonpath='{ .items[0].status.addresses[?(@.type=="InternalIP")].address }'; \
elif [ "$(VM_TYPE)" = "vagrant" ]; then \
node_ip_line=$$(vagrant ssh master -c 'ifconfig eth1 | grep "inet "' 2> /dev/null); \
node_ip=$$(echo $$node_ip_line | sed "s/.*inet \([^ ]*\) .*/\1/"); \
echo $$node_ip; \
elif [ "$(VM_TYPE)" = "minikube" ]; then \
echo $$(minikube ip); \
elif [ "$(VM_TYPE)" = "ibmcloud" ]; then \
echo $$(bx cs workers $(CLUSTER_NAME) | grep Ready | awk '{ print $$2;exit }'); \
elif [ "$(VM_TYPE)" = "none" ]; then \
echo "$(PUBLIC_IP)"; \
else \
echo "$(PUBLIC_IP)"; \
fi
install-minikube-in-ci:
@echo "Starting local kubernetes cluster (minikube)"
@curl -s -Lo kubectl https://storage.googleapis.com/kubernetes-release/release/$(CI_KUBECTL_VERSION)/bin/linux/amd64/kubectl && \
chmod +x kubectl && sudo mv kubectl /usr/local/bin/
@curl -s -Lo minikube https://storage.googleapis.com/minikube/releases/$(CI_MINIKUBE_VERSION)/minikube-linux-amd64 && \
chmod +x minikube && sudo mv minikube /usr/local/bin/
@sudo minikube start --vm-driver=none
@minikube update-context
@JSONPATH='{range .items[*]}{@.metadata.name}:{range @.status.conditions[*]}{@.type}={@.status};{end}{end}'; \
until kubectl get nodes -o jsonpath="$$JSONPATH" 2>&1 | grep -q "Ready=True"; do sleep 1; done
@while ! kubectl get svc/kubernetes > /dev/null; do sleep 1; done
@while ! kubectl get rc/kubernetes-dashboard --namespace=kube-system > /dev/null 2>&1; do sleep 1; done
@kubectl delete --namespace=kube-system rc/kubernetes-dashboard svc/kubernetes-dashboard || true
.PHONY: build deploy controller databroker_objectstorage databroker_s3