Budgets & Notifications -> Create Budget Name of Project
Projects - Select
[Services you want to be notified]
Create budget - Value - [Percentages of Progress you want to be alerted]
When you use a service, the first step is to activate the API of this service.
Menu -> Computer Engine -> VM Instance -> Activate API
VM Instance -> Create Instance
[Name; Region; Zone]
[Machine type E2 configuration]
[OS image; Version; storage]
Service account - says which services the machine can access
Good practice: Create a new service account only with the services you will use.
Firewall: [ports].
SSH; local host name;
CREATE
It gives an internal and external IP.
You can access the machine by SSH, Cloud Shell;
gcloud compute instances list
gcloud compute ssh machine_name --zone-us-central1-a
Menu -> Big Data -> Dataproc
Clusters; Jobs; Workflows; AutoScale Police (cluster increases power in acord with a threshold); Components; MetaStore (metadata); Notebooks (Jupyter).
Create bucket (places you can put data, when you execute a code, you point to this bucket)- Cloud Storage
Create range - Type of local - options of storage - Access Control - Cryptography
Create Cluster
Machine type
Personalize Cluster
It created a bucket for specific information of the cluster - Bucket of Cloud Storage Preparation (click to access).
Open cluster - Monitoring Screen - Jobs, VM, Web Interface (component gateway)
Compute Engine - you can access virtual machines.
Menu -> Dataproc -> Send job -> ID, Name, Type, Class (org.apache.spark.examples.SparkPi), jar files (file://usr/lib/spark/examples/jars/spark-example.jar), arguments(1000) - SEND
It can be also executed by ssh or Cloud-shell
It will create a folder with the results inside the Bucket.The folder contains a file with the log of the job and a file with the result.