export RANDOMID=$RANDOM
export PROJECT_ID=data-flow-$RANDOMID
export EMAIL_ADDRESS=learnpde@gmail.com
export BUCKET_NAME=dataflow-wordcount-example-$RANDOMID
export DATAFLOW_REGION=us-central1
echo $PROJECT_ID
echo $BUCKET_NAMEgcloud projects create $PROJECT_IDgcloud config set project $PROJECT_IDexport PROJECT_NUMBER=$(gcloud projects list --filter=$PROJECT_ID --format="value(PROJECT_NUMBER)")echo $PROJECT_NUMBERgcloud services enable dataflow compute_component logging storage_component storage_api \
bigquery pubsub datastore.googleapis.com cloudresourcemanager.googleapis.comgcloud auth application-default logingcloud projects add-iam-policy-binding $PROJECT_ID \
--member="user:${EMAIL_ADDRESS}" --role=roles/iam.serviceAccountUsergsutil mb -c STANDARD -l US gs://$BUCKET_NAMEgcloud projects add-iam-policy-binding $PROJECT_ID \
--member="serviceAccount:${PROJECT_NUMBER}-compute@developer.gserviceaccount.com" \
--role=roles/dataflow.admin
gcloud projects add-iam-policy-binding $PROJECT_ID \
--member="serviceAccount:${PROJECT_NUMBER}-compute@developer.gserviceaccount.com" \
--role=roles/dataflow.worker
gcloud projects add-iam-policy-binding $PROJECT_ID \
--member="serviceAccount:${PROJECT_NUMBER}-compute@developer.gserviceaccount.com" \
--role=roles/storage.objectAdminpython --version
python -m pip --versionmkdir dataflow-$RANDOMID
cd dataflow-$RANDOMID
python -m venv envsource env/bin/activatepip install wheel
pip install 'apache-beam[gcp]'python -m apache_beam.examples.wordcount \
--output outputsmore outputs*python -m apache_beam.examples.wordcount \
--region $DATAFLOW_REGION \
--input gs://dataflow-samples/shakespeare/kinglear.txt \
--output gs://${BUCKET_NAME}/results/outputs \
--runner DataflowRunner \
--project $PROJECT_ID \
--temp_location gs://${BUCKET_NAME}/tmp/gcloud projects delete $PROJECT_ID