Skip to content

Commit 639fc00

Browse files
[PPML] Update Version in doc to 2.0.0 (#4206)
* [PPML] Update PPML Image Version to 2.0.0 * Update PPML Image and Version to 2.0.0
1 parent 39beb41 commit 639fc00

21 files changed

+30
-30
lines changed

ppml/kms-client/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ Pay attention to the variables and set them to correct values according to your
6969
```bash
7070
export INPUT_DIR_PATH=YOUR_DATA_FILE_DIRECTORY_PATH # For example, multiple CSV files are in this directory
7171
export KMS_SERVER_IP=YOUR_KMS_SERVER_IP # IP address of node where the previous KMS server is deployed
72-
export DOCKER_IMAGE=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:0.14.0-SNAPSHOT
72+
export DOCKER_IMAGE=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:2.0.0
7373

7474
sudo docker pull $DOCKER_IMAGE
7575

ppml/trusted-big-data-ml/python/docker-graphene/README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -229,7 +229,7 @@ Run the example with SGX spark local mode with the following command in the term
229229

230230
```bash
231231
SGX=1 ./pal_loader bash -c "/opt/jdk8/bin/java -cp \
232-
'/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/jars/*:/ppml/trusted-big-data-ml/work/spark-3.1.2/conf/:/ppml/trusted-big-data-ml/work/spark-3.1.2/jars/*' \
232+
'/ppml/trusted-big-data-ml/work/bigdl-2.0.0/jars/*:/ppml/trusted-big-data-ml/work/spark-3.1.2/conf/:/ppml/trusted-big-data-ml/work/spark-3.1.2/jars/*' \
233233
-Xmx2g \
234234
org.apache.spark.deploy.SparkSubmit \
235235
--master 'local[4]' \
@@ -239,13 +239,13 @@ SGX=1 ./pal_loader bash -c "/opt/jdk8/bin/java -cp \
239239
--conf spark.rpc.message.maxSize=190 \
240240
--conf spark.network.timeout=10000000 \
241241
--conf spark.executor.heartbeatInterval=10000000 \
242-
--properties-file /ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/conf/spark-bigdl.conf \
243-
--py-files /ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/python/bigdl-orca-spark_3.1.2-0.14.0-SNAPSHOT-python-api.zip,/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/python/bigdl-dllib-spark_3.1.2-0.14.0-SNAPSHOT-python-api.zip,/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/examples/dllib/lenet/lenet.py \
242+
--properties-file /ppml/trusted-big-data-ml/work/bigdl-2.0.0/conf/spark-bigdl.conf \
243+
--py-files /ppml/trusted-big-data-ml/work/bigdl-2.0.0/python/bigdl-orca-spark_3.1.2-2.0.0-python-api.zip,/ppml/trusted-big-data-ml/work/bigdl-2.0.0/python/bigdl-dllib-spark_3.1.2-2.0.0-python-api.zip,/ppml/trusted-big-data-ml/work/bigdl-2.0.0/examples/dllib/lenet/lenet.py \
244244
--driver-cores 2 \
245245
--total-executor-cores 2 \
246246
--executor-cores 2 \
247247
--executor-memory 8g \
248-
/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/examples/dllib/lenet/lenet.py \
248+
/ppml/trusted-big-data-ml/work/bigdl-2.0.0/examples/dllib/lenet/lenet.py \
249249
--dataPath /ppml/trusted-big-data-ml/work/data/mnist \
250250
--maxEpoch 2" 2>&1 | tee test-bigdl-lenet-sgx.log
251251
```
@@ -581,7 +581,7 @@ export KEYS_PATH=/YOUR_DIR/keys
581581
export SECURE_PASSWORD_PATH=/YOUR_DIR/password
582582
export KUBECONFIG_PATH=/YOUR_DIR/kuberconfig
583583
export LOCAL_IP=$LOCAL_IP
584-
export DOCKER_IMAGE=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:0.14.0-SNAPSHOT
584+
export DOCKER_IMAGE=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:2.0.0
585585
sudo docker run -itd \
586586
--privileged \
587587
--net=host \

ppml/trusted-big-data-ml/python/docker-graphene/deploy-local-spark-sgx.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ export ENCLAVE_KEY_PATH=YOUR_LOCAL_ENCLAVE_KEY_PATH
44
export DATA_PATH=YOUR_LOCAL_DATA_PATH
55
export KEYS_PATH=YOUR_LOCAL_KEYS_PATH
66
export LOCAL_IP=YOUR_LOCAL_IP
7-
export DOCKER_IMAGE=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:0.14.0-SNAPSHOT
7+
export DOCKER_IMAGE=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:2.0.0
88

99
sudo docker pull $DOCKER_IMAGE
1010

ppml/trusted-big-data-ml/python/docker-graphene/environment.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
export MASTER=YOUR_MASTER_IP
33
export WORKERS=(YOUR_WORKER_IP_1 YOUR_WORKER_IP_2 YOUR_WORKER_IP_3)
44

5-
export TRUSTED_BIGDATA_ML_DOCKER=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:0.14.0-SNAPSHOT
5+
export TRUSTED_BIGDATA_ML_DOCKER=intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:2.0.0
66

77
export SOURCE_ENCLAVE_KEY_PATH=YOUR_LOCAL_ENCLAVE_KEY_PATH
88
export SOURCE_KEYS_PATH=YOUR_LOCAL_KEYS_PATH

ppml/trusted-big-data-ml/python/docker-graphene/start-scripts/start-spark-local-bigdl-sgx.sh

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
cd /ppml/trusted-big-data-ml
33

44
SGX=1 ./pal_loader bash -c "/opt/jdk8/bin/java -cp \
5-
'/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/jars/*:/ppml/trusted-big-data-ml/work/spark-3.1.2/conf/:/ppml/trusted-big-data-ml/work/spark-3.1.2/jars/*' \
5+
'/ppml/trusted-big-data-ml/work/bigdl-2.0.0/jars/*:/ppml/trusted-big-data-ml/work/spark-3.1.2/conf/:/ppml/trusted-big-data-ml/work/spark-3.1.2/jars/*' \
66
-Xmx2g \
77
org.apache.spark.deploy.SparkSubmit \
88
--master 'local[4]' \
@@ -12,13 +12,13 @@ SGX=1 ./pal_loader bash -c "/opt/jdk8/bin/java -cp \
1212
--conf spark.rpc.message.maxSize=190 \
1313
--conf spark.network.timeout=10000000 \
1414
--conf spark.executor.heartbeatInterval=10000000 \
15-
--properties-file /ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/conf/spark-bigdl.conf \
16-
--py-files local://${BIGDL_HOME}/python/bigdl-orca-spark_${SPARK_VERSION}-${BIGDL_VERSION}-python-api.zip,local://${BIGDL_HOME}/python/bigdl-dllib-spark_${SPARK_VERSION}-${BIGDL_VERSION}-python-api.zip,/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/examples/dllib/lenet/lenet.py \
15+
--properties-file /ppml/trusted-big-data-ml/work/bigdl-2.0.0/conf/spark-bigdl.conf \
16+
--py-files local://${BIGDL_HOME}/python/bigdl-orca-spark_${SPARK_VERSION}-${BIGDL_VERSION}-python-api.zip,local://${BIGDL_HOME}/python/bigdl-dllib-spark_${SPARK_VERSION}-${BIGDL_VERSION}-python-api.zip,/ppml/trusted-big-data-ml/work/bigdl-2.0.0/examples/dllib/lenet/lenet.py \
1717
--driver-cores 2 \
1818
--total-executor-cores 2 \
1919
--executor-cores 2 \
2020
--executor-memory 8g \
21-
/ppml/trusted-big-data-ml/work/bigdl-0.14.0-SNAPSHOT/examples/dllib/lenet/lenet.py \
21+
/ppml/trusted-big-data-ml/work/bigdl-2.0.0/examples/dllib/lenet/lenet.py \
2222
--dataPath /ppml/trusted-big-data-ml/work/data/mnist \
2323
--maxEpoch 2" 2>&1 | tee test-bigdl-lenet-sgx.log && \
2424
cat test-bigdl-lenet-sgx.log | egrep -a "Accuracy"

ppml/trusted-big-data-ml/scala/docker-occlum/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
Pull image from dockerhub.
77

88
```bash
9-
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:0.14.0-SNAPSHOT
9+
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.0.0
1010
```
1111

1212
Also, you can build image with `build-docker-image.sh`. Configure environment variables in `Dockerfile` and `build-docker-image.sh`.

ppml/trusted-big-data-ml/scala/docker-occlum/build-docker-image.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,4 +12,4 @@ sudo docker build \
1212
--build-arg HTTPS_PROXY_PORT=$HTTPS_PROXY_PORT \
1313
--build-arg no_proxy=x.x.x.x \
1414
--build-arg SPARK_JAR_REPO_URL=$SPARK_JAR_REPO_URL \
15-
-t intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:0.14.0-SNAPSHOT -f ./Dockerfile .
15+
-t intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.0.0 -f ./Dockerfile .

ppml/trusted-big-data-ml/scala/docker-occlum/kubernetes/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,12 +3,12 @@
33
## Prerequisite
44

55
* Check Kubernetes env or Install Kubernetes from [wiki](https://kubernetes.io/zh/docs/setup/production-environment)
6-
* Prepare image `intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:0.14.0-SNAPSHOT`
6+
* Prepare image `intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.0.0`
77

88
1. Pull image from Dockerhub
99

1010
```bash
11-
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:0.14.0-SNAPSHOT
11+
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.0.0
1212
```
1313

1414
If Dockerhub is not accessable, we can build docker image with Dockerfile and modify the path in the build-docker-image.sh firstly.

ppml/trusted-big-data-ml/scala/docker-occlum/kubernetes/executor.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ metadata:
66
spec:
77
containers:
88
- name: spark-example
9-
image: intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:0.14.0-SNAPSHOT
9+
image: intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.0.0
1010
imagePullPolicy: Never
1111
volumeMounts:
1212
- name: sgx-enclave

ppml/trusted-big-data-ml/scala/docker-occlum/kubernetes/run_spark_gbt.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ${SPARK_HOME}/bin/spark-submit \
77
--class org.apache.spark.examples.ml.GBTExample \
88
--conf spark.executor.instances=1 \
99
--conf spark.rpc.netty.dispatcher.numThreads=32 \
10-
--conf spark.kubernetes.container.image=intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:0.14.0-SNAPSHOT \
10+
--conf spark.kubernetes.container.image=intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-occlum:2.0.0 \
1111
--conf spark.kubernetes.authenticate.driver.serviceAccountName=spark \
1212
--conf spark.kubernetes.executor.podNamePrefix="sparkgbt" \
1313
--conf spark.kubernetes.executor.deleteOnTermination=false \

0 commit comments

Comments
 (0)