Difference between revisions of "Cloud Computing"

From Sinfronteras
Jump to: navigation, search
(Allow connections on HTTP 80, HTTPS 443 or other port)
(AWS Platform)
Line 890: Line 890:
  
  
 +
<br />
 +
===Pricing===
 +
https://aws.amazon.com/ec2/pricing/on-demand/
 +
 +
t2.micro 1 Variable 1 GiB EBS Only $0.0116 per Hour:
 +
: $ 100.224 / y
 +
 +
t2.small 1 Variable 2 GiB EBS Only $0.023 per Hour:
 +
: $ 198.72 / y
 +
 +
 +
<br />
 
===Allow connections on HTTP 80, HTTPS 443 or other port===
 
===Allow connections on HTTP 80, HTTPS 443 or other port===
 
For example, if we host a website on an EC2 instance and want apache to serve the pages.
 
For example, if we host a website on an EC2 instance and want apache to serve the pages.

Revision as of 17:09, 15 February 2020

  • There are some virtualization and cloud material in this resume that could be added to the wiki.



Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction. This cloud model is composed of five essential characteristics, three service models, and four deployment models.



Contents

CA1 - Case study - Migration of on-premise network onto a public Cloud provider

CA1 - Physical Server Hardware Research / Case study of one company that successfully migrated their on-premise network onto a public Cloud provider

Media:Physical_Server_Hardware_Research-CA1_Cloud.pdf



Essential Characteristics

On-demand self-service: A consumer can unilaterally provision computing capabilities, such as server time and network storage, as needed automatically without requiring human interaction with each service provider.


Broad network access: Capabilities are available over the network and accessed through standard mechanisms that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones, tablets, laptops, and workstations).


Resource pooling: The provider's computing resources are pooled to serve multiple consumers using a multi-tenant model, with different physical and virtual resources dynamically assigned and reassigned according to consumer demand. There is a sense of location independence in that the customer generally has no control or knowledge over the exact location of the provided resources but may be able to specify location at a higher level of abstraction (e.g., country, state, or datacenter). Examples of resources include storage, processing, memory, and network bandwidth.


Rapid elasticity: Capabilities can be elastically provisioned and released, in some cases automatically, to scale rapidly outward and inward commensurate with demand. To the consumer, the capabilities available for provisioning often appear to be unlimited and can be appropriated in any quantity at any time.


Measured service: Cloud systems automatically control and optimize resource use by leveraging a metering capability1 at some level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth, and active user accounts). Resource usage can be monitored, controlled, and reported, providing transparency for both the provider and consumer of the utilized service.


On-Premise Computing:

  • Requires hardware, space, electricity, cooling
  • Requires managing OS, applications and updates
  • Software Licensing
  • Difficult to scale:
    • Too much or too little capacity
  • High upfront capital costs
  • You have complete control


Cloud Computing:

  • Shared, multi-tenant environment
  • Pools of computing resources
  • Resources can be requested as required
  • Available via the Internet
  • Private clouds can be available via private WAN
  • Pay as you go

Una de las ventajas más importantes del Cloud es que puede ser «scaled» fácilmente (Resources can be requested as required). En otras palabras, a través del cloud la capacidad se puede variar fácilmente con respecto a la demanda. sólo cuando la demanda sea alta, el sistema será configurado para disponer de gran capacidad. On-Premise Computing, lo que generalmente pasa es que se configura el sistema para cubrir una capacidad promedio y generalmente se debe sobreestimar para que, cuando la demanda sea alta, poder todavía cubrirla. Entonces, cuando la demanda es baja estamos a "over capacity" y podría también pasar que cuando la demanda sea muy alta estemos a "under capacity". Esto es solucionado en el cloud. Esto se puede visualizar a través del ejemplo presentado en las siguientes figuras, en el cual se plantea el caso de un Concert ticket web site.

Cloud-Managing demand1.png
Cloud-Managing demand2.png



Service Models

Infrastructure as a Service (IaaS) - Platform as a Service (PaaS) - Software as a Service (SaaS)

Cloud: IaaS - PaaS - SaaS: En esta gráfica se puede ver que la Virtual machine corre directamente sobre el hardware.. No hay un SO sobre el cual corren las VM's


Software as a Service (SaaS): The capability provided to the consumer is to use the provider’s applications running on a cloud infrastructure2. The applications are accessible from various client devices through either a thin client interface, such as a web browser (e.g., web-based email), or a program interface. The consumer does not manage or control the underlying cloud infrastructure including network, servers, operating systems, storage, or even individual application capabilities, with the possible exception of limited user-specific application configuration settings.

Platform as a Service (PaaS): The capability provided to the consumer is to deploy onto the cloud infrastructure consumer-created or acquired applications created using programming languages, libraries, services, and tools supported by the provider.3 The consumer does not manage or control the underlying cloud infrastructure including network, servers, operating systems, or storage, but has control over the deployed applications and possibly configuration settings for the application-hosting environment.

Infrastructure as a Service (IaaS): The capability provided to the consumer is to provision processing, storage, networks, and other fundamental computing resources where the consumer is able to deploy and run arbitrary software, which can include operating systems and applications. The consumer does not manage or control the underlying cloud infrastructure but has control over operating systems, storage, and deployed applications; and possibly limited control of select networking components (e.g., host firewalls).



Deployment Models

Private cloud: The cloud infrastructure is provisioned for exclusive use by a single organization comprising multiple consumers (e.g., business units). It may be owned, managed, and operated by the organization, a third party, or some combination of them, and it may exist on or off premises.

Community cloud: The cloud infrastructure is provisioned for exclusive use by a specific community of consumers from organizations that have shared concerns (e.g., mission, security requirements, policy, and compliance considerations). It may be owned, managed, and operated by one or more of the organizations in the community, a third party, or some combination of them, and it may exist on or off premises.

Public cloud: The cloud infrastructure is provisioned for open use by the general public. It may be owned, managed, and operated by a business, academic, or government organization, or some combination of them. It exists on the premises of the cloud provider.

Hybrid cloud: The cloud infrastructure is a composition of two or more distinct cloud infrastructures (private, community, or public) that remain unique entities, but are bound together by standardized or proprietary technology that enables data and application portability (e.g., cloud bursting for load balancing between clouds).



Containers

https://www.cio.com/article/2924995/software/what-are-containers-and-why-do-you-need-them.html https://cloud.google.com/containers/

Containers are a solution to the problem of how to get software to run reliably when moved from one computing environment to another. This could be from a developer's laptop to a test environment, from a staging environment into production, and perhaps from a physical machine in a data center to a virtual machine in a private or public cloud.

Problems arise when the supporting software environment is not identical. For example: "You're going to test using Python 2.7, and then it's going to run on Python 3 in production and something weird will happen. Or you'll rely on the behavior of a certain version of an SSL library and another one will be installed. You'll run your tests on Debian and production is on Red Hat and all sorts of weird things happen."

And it's not just different software that can cause problems, he added. "The network topology might be different, or the security policies and storage might be different but the software has to run on it."

How do containers solve this problem? Put simply, a container consists of an entire runtime environment: an application, plus all its dependencies, libraries and other binaries, and configuration files needed to run it, bundled into one package. By containerizing the application platform and its dependencies, differences in OS distributions and underlying infrastructure are abstracted away.

Containers offer a logical packaging mechanism in which applications can be abstracted from the environment in which they actually run.



Google Cloud Platform

https://console.cloud.google.com

para ver la Amount remaining del Promotion value of 243.35: Principal Menu (Burger) > Billing > Overview


Lo primero que tenemos que hacer es crear un proyecto. Generalmente hay un proyecto by default llamado «My First Project». Para crear un nuevo proyecto debemos ir a la pestaña que se encuentra justo al lado del Main Menu (Hamburger).



CA2-Interacting with the GCP

Media:Interacting_with_the_GCP-CA2_Cloud.pdf

  1. Creating a storage bucket. Uploading items from a host computer using the command line and the GUI and place them into the bucket
    1. Using the command line
    2. Using the GUI
  2. Upload items from the bucket to a Linux virtual machine using the Google CLI
  3. Creation of a Linux VM, installing Apache and uploading the web page to the web site
  4. Create a Linux VM, install NGINX and upload the web page to the web site
  5. Create a Windows VM, install IIS and upload the web page to the web site
  6. Live Migration of a VirtualBox VM to the GCP
  7. Explain what Live Migration is and identify situations where DigiTech could benefit from it
  8. Research topic: Other services available from Google’s Cloud Launcher
    1. Python tutorial
  9. Challenging research topic - GCSFUSE: It allows you to mount a bucket to a Debian Linux virtual machine
    1. Installing Cloud Storage FUSE and its dependencies
    2. Mounting a Google Cloud Storage Bucket as a local disk
    3. Change the root directory of an apache server



Cloud SDK: Command-line interface for Google Cloud Platform

https://cloud.google.com/sdk/

The Cloud SDK is a set of tools for Cloud Platform. It contains gcloud, gsutil, and bq, which you can use to access Google Compute Engine, Google Cloud Storage, Google BigQuery, and other products and services from the command-line. You can run these tools interactively or in your automated scripts. A comprehensive guide to gcloud can be found in gcloud Overview.



Instalación

El proceso de instalación se encuentra detallado en https://cloud.google.com/sdk/

Logré instalarlo correctamente a través de la opción: Install for Debian / Ubuntu: https://cloud.google.com/sdk/docs/quickstart-debian-ubuntu

La procesimiento de instalación general para Linux: https://cloud.google.com/sdk/docs/quickstart-linux generó errores en mi sistema y no se pudo completar la instalación.



Initialize the SDK

https://cloud.google.com/sdk/docs/quickstart-linux

Luego de instalar SDK, debemos usar the gcloud init command to perform several common SDK setup tasks. These include:

  • Especificar la user account (adeloaleman@gmail.com): authorizing the SDK tools to access Google Cloud Platform using your user account credentials
  • Setting up the default SDK configuration:
  • El proyecto (creado en al GCP) al que vamos a acceder by default
  • ...
gcloud init



Crear una VM Instance

Main Menu (Hamburger) > Compute Engine > VM instances:

  • Create Instance



Linux VM Instance

  • Name: debian9-1
  • Zone: Lo lógico es elegir la zona donde te encuentras, aunque si elijes alguna otra no habrá mucha diferencia (el precio varía dependiendo de la zona)
Elegiremos la primera zona de West Europe propuesta: europe-west4-a
  • Machine type: Is customize. Las características por defecto son:
1vCPU / 3.75 GB memory
Las características más básicas son:
micro (1 shared vCPU 0.6GB memory) (el precio baja considerablemente si elegimos éste) (para Linux podría ser suficiente)
  • Container: Deploy a container image to this VM instance.
Ver Containers
No es, por ahora, necesario activar esta opción para nuestro propósito de prueba.
  • Boot disk:
OS: La opción por defecto es Debian GNU/Linux 9 (stretch). Vamos estar trabajando con éste.
Disk size: 10GB por es la opción por defecto para Linux.
  • Identity and API access :
Service account: Esta opción se encuentra ajustada a "Compute Engine Default Service Account".
Access scopes:
  • Allow default access: Activada por defecto...
  • Allow full access to all Cloud APIs
  • Es apropiado marcar esta opción. Si no se marca no se podrá, por ejemplo, copiar files desde una VM instance a un bucket de esta forma:
gsutil cp file.txt gs://adelostorage/
  • Otra opción es para permitir copiar archivos desde la VM hacia un bucket es configurar la "Service account". Debemos elegir una Service Account que hayamos configurade previamente para permitir este tipo de acciones. Ver Service Accounts para más detalles
  • Set access for each API
  • Firewall:
Add tags and firewall rules to allow specific network traffic from the Internet
  • Allow HTTP traffic: Activaremos esta opción to be able to maange the site through an SSH connection.
  • Allow HTTPS traffic: Activaremos esta opción to be able to maange the site through an SSH connection.
  • Management, disks, networking, SSH keys: Este link nos da la opción de realizar muchas otras configuraciones. Dejaremos todo por defecto por ahora.



Crear storage

Create a storage bucket:

Main Menu > Storage > Browser:

  • Create bucket:
    • Name: mi_storage-1
    • Default storage class: Vamos a dejar la opción by default -- Multi-Regional
    • Location: Europe



Connecting to Instances

https://cloud.google.com/compute/docs/instances/connecting-to-instance

Compute Engine provides tools to manage your SSH keys and help you connect to either Linux and Windows Server instances.



Connecting to Linux instances

You can connect to Linux instances through either:

  • The Google Cloud Platform Console
  • The gcloud command-line tool.
  • Connecting using third-party tools

Compute Engine generates an SSH key for you and stores it in one of the following locations:

  • By default, Compute Engine adds the generated key to project or instance metadata.
  • If your account is configured to use OS Login, Compute Engine stores the generated key with your user account.



The Google Cloud Platform Console

Esta es un Console que pertenece a la GCP a la cual se accede a travé de una Browser window. Para abrirla vamos a:

  • In the GCP (Google Cloud Platform) Console, go to the VM Instances page.
  • In the list of virtual machine instances, click SSH in the row of the instance that you want to connect to.



The gcloud command-line tool

A través del comando gcloud (incluido en Cloud SDK) podemos acceder a la VM Instance desde el Terminal (Console) de nuestra computadora.

gcloud compute ssh [INSTANCE_NAME]



Connecting using third-party tools

https://cloud.google.com/compute/docs/instances/connecting-advanced#thirdpartytools

Ver Linux#SSH

También podemos acceder a una VM Instance sin necesidad de utilizar herramientas de la GCP, como lo son The Google Cloud Platform Console or gcloud. Podemos, en su lugar, utilizar SSH.

Luego de crear una VM Instance, no podremos acceder a la VM a través de SSH debido a que ningún método de autentificación estará disponible. Como explicamos en Linux#SSH. Hay dos approach para la autentificación al usar SSH:

  • Using the password de la «cuenta linux»:
  • Las Google Cloud VM Instance se crean por defecto sin password. Además, en la configuración por defectos de las Google VM Instance, se impide la autentificación con password. Es decir, el archivo «/etc/ssh/sshd_config» presenta la siguiente configuración: «PasswordAuthentication no».
  • Entonces, tenemos primero que encontrar otro método de acceder a las GC VM para:
  1. Configurar «PasswordAuthentication yes»
  2. Configurar un password para el user@VMInstance
  • SSH Keys: Si queremos autentificar la conexión con este método tenemos primero que copiar la «public key» de nuestra computadora en las Google Cloud VM Instances.
  • Como se explica en Linux#SSH, si queremos hacerlo con el comando «ssh-copy-id» tenemos primero que configurar una password para el user@VMInstance y habilitar la autentificación a través de dicho password como explicamos arriba.
  • Entonces, igualmente, tenemos primero que encontrar otro método de acceder a las GC VM para poder copiar la «public key».


Alors, para conectarnos a las GC VM y realizar las configuraciones necesarias para permitir la conexión con SSH podemos usar el comando «gcloud» como ya hemos visto:

gcloud compute ssh [INSTANCE_NAME]

Cuando realizamos la primera conexión a través de «gcloud», el comando generará las keys necesarias para la conexión con «gcloud»:

SSH google keys.png


Lo que he constatado es que, luego de crear la GC VM, si intento ingresar con «ssh» antes de haber ingresado con «gcloud», no será posible el acceso porque no hay método de autentificación. Sin embargo, si primero me conecto con «gcloud», luego sí puedo ingresar con «ssh» sin necesidad de realizar otra configuración ni copiar la «public key». Esto me imagino que es porque, como se ve en la imagen mostrada arriba, «gcloud» ha creado las «key» y copiado la «public key» en la computadora destino; y al parecer éstas mismas funcionan con una conexión normal de ssh (sin acudir a «gcloud»). Esto lo he constatado conectandome desde mi computadora a las GC VM. Sin embargo, siguiendo el mismo procedimiento intenté conectarme entre dos GC VM Instances y no logré estableser la conección a través de SSH. El problema lo solucioné, siguiendo el consejo de un forum, borrando todos los archivos en «~/.ssh/» excepto «known_hosts» y generando nuevas keys (con «ssh-keygen») y copiando la nueva «public key» en la GC VM correspondiente, siguiendo los procedimientos en Linux#SSH.



Copy files


From the host computer to a bucket on Google Cloud

Luego de crear el bucket, podemos copiar archivos al bucket. Hay diferentes formas de hacerlo:

  • Sinple drag and drop
  • Usando el terminal de Google cloud
  • Desde el terminal de la computadora luego de instalar Cloud SDK



Desde el terminal de la computadora luego de instalar Cloud SDK

El siguiente comando copia un archivo que está en mi computadora a un bucket on GC:

gsutil cp probando gs://adelostorage

Where adelostorage is the name of the bucket.



From the host computer to a Linux VM Instance or between VM's

https://cloud.google.com/sdk/gcloud/reference/compute/scp

To copy a remote directory, ~/narnia, from example-instance to the ~/wardrobe directory of your local host, run:

gcloud compute scp --recurse example-instance:~/narnia ~/wardrobe



From a bucket to an VM Instance

Luego de acceder a una VM Instance, puedo copiar un archivo desde un bucket a dicha VM Instance a través de la siguiente orden:

 gsutil cp gs://adelostorage/probando.txt .



Acceding Cloud Storage buckets

He estado buscando una forma para acceder a mis Buckets de forma tal que sea posible compartir el bucket con otras personas:

  • A través de una URL
  • A través de un IP

En los siguientes links se explica como hacer el bucket público y como acceder a él a través de una URL:

Pensé que lo que se explicaba en esos links era la solución. Pero creo que eso sólo permite hacer el bucket público para todos los usuarios de un proyecto (creo); porque la URL a la que creo se refieren en dicho link es:

y pues ese link sólo permitiría el acceso a mi bucket al estar logeados en mi cuenta.



Connecting your domain to a Cloud Storage Bucket

Lo que sí encontré es que hay una forma de asociar el bucket a un dominio que nos pertenezca. Por ejemplo, el siguiente subdominio lo voy a asociar a un bucket:

En el siguiente link se explica como connect your domain to Cloud Storage and Hosting a Static Website en el Bucket: https://cloud.google.com/storage/docs/hosting-static-website



Mounting Cloud Storage buckets as file systems

Cloud Storage FUSE: Mounting Cloud Storage buckets as file systems on Linux or OS X systems: https://cloud.google.com/storage/docs/gcs-fuse

Cloud Storage FUSE is an open source FUSE adapter that allows you to mount Cloud Storage buckets as file systems on Linux or OS X systems.



Installing Cloud Storage FUSE and its dependencies

https://github.com/GoogleCloudPlatform/gcsfuse/blob/master/docs/installing.md

  • Add the gcsfuse distribution URL as a package source and import its public key:
export GCSFUSE_REPO=gcsfuse-`lsb_release -c -s`
echo "deb http://packages.cloud.google.com/apt $GCSFUSE_REPO main" | sudo tee /etc/apt/sources.list.d/gcsfuse.list
curl https://packages.cloud.google.com/apt/doc/apt-key.gpg | sudo apt-key add -
  • Update the list of packages available and install gcsfuse:
sudo apt-get update
sudo apt-get install gcsfuse
  • (Ubuntu before wily only) Add yourself to the fuse group, then log out and back in:
sudo usermod -a -G fuse $USER
exit
  • Future updates to gcsfuse can be installed in the usual way:
sudo apt-get update && sudo apt-get upgrade



Set up credentials for Cloud Storage FUSE

Cloud Storage FUSE auto-discovers credentials based on application default credentials:

  • If you are running on a Google Compute Engine instance with scope storage-full configured, then Cloud Storage FUSE can use the Compute Engine built-in service account. For more information, see Using Service Accounts with Applications.
    • Si se está usando una Google Compute Engine Instance, no se tiene que hacer nada para montar el Bucket. Sin embargo si el Access scopes de dicha Instance no fue configurado correctamente, entonces no podremos copiar arechivos desde la Instance hacia el Bucket montado. Ver en la creación de la VM Instance que es apropiado elegir "Allow full access to all Cloud APIs" o elegir una Service Account que hayamos configurade previamente para permitir este tipo de acciones. Ver Service Accounts para más detalles.

  • If you installed the Google Cloud SDK and ran gcloud auth application-default login, then Cloud Storage FUSE can use these credentials.
  • If you set the environment variable GOOGLE_APPLICATION_CREDENTIALS to the path of a service account's JSON key file, then Cloud Storage FUSE will use this credential. For more information about creating a service account using the Google Cloud Platform Console, see Service Account Authentication.



Mounting a Google Cloud Storage Bucket as a local disk

  • Create a directory:
mkdir /path/to/mount
  • Create the bucket you wish to mount, if it doesn't already exist, using the Google Cloud Platform Console.
  • Use Cloud Storage FUSE to mount the bucket (e.g. example-bucket).
gcsfuse example-bucket /path/to/mount

Es muy importante notar que si montamos el bucket como se muestra arriba, no podremos accederlo a través del Web Brower. Es decir, el Web server no tendrá la permisología para acceder a este directorio.

Esto es un problema si queremos colocar un sitio web en el bucket; al cual se va a acceder a través de la VM Instance.

En este caso, debemos montarlo de la siguiente forma: https://www.thedotproduct.org/posts/mounting-a-google-cloud-storage-bucket-as-a-local-disk.html

sudo gcsfuse -o noatime -o noexec --gid 33 --implicit-dirs -o ro -o nosuid -o nodev --uid 33 -o allow_other mi_storage1 /path/directory/

Explaining the mount options:

  • ro - mount the volume in read-only mode
  • uid=33,gid=33 - grant access to user-id and group-id 33 (usually the web serve user e.g. www-data):
    • Your OS has a web server user whose user-id (UID) and groupd-id (GID) are 33

Para desplegar el web server user-id (UID) and groupd-id: http://www.digimantra.com/linux/find-users-uid-gid-linux-unix/

id -u www-data
id -g www-data
  • noatime - don't update access times of files
  • _netdev - mark this as a network volume so the OS knows that the network needs to be available before it'll try to mount
  • noexec - don't allow execution of files in the mount
  • user - allow non-root users to mount the volume
  • implicit_dirs - treat object paths as if they were directories - this is critical and Fuse-specific
  • allow_other - allow non-root users to mount (Fuse-specific AFAIK)
  • To unmount the bucket:
fusermount -u /home/shared/local_folder/



Firewall Rules

How to open a specific port such as 9090 in Google Compute Engine:

https://cloud.google.com/vpc/docs/using-firewalls

https://stackoverflow.com/questions/21065922/how-to-open-a-specific-port-such-as-9090-in-google-compute-engine

Por defectos, sólo el puerto tcp:80 se encuentra abierto. Los otros están protegidos por los firewalls. Si configuramos, por ejemplo, un Web Server (Apache, NGINX) en un puerto distinto al 80, debemos modificar los Firewall Rules para que el Web Server sea capaz de servir las páginas a través de otro puerto.

Para abrir un puerto:

  • Choose you Project: al lado del Main Menu (burguer)
  • Then, go to: Burguer > VPC network > Firewall rules:
    • Create a firewall rule

Luego, seguir los pasos indicados en: https://cloud.google.com/vpc/docs/using-firewalls: Con las siguientes especificaciones (el resto en default) he sido capaz de hacerlo:

  • Targets
    • All instances in the network
  • Source IP ranges:
    • If you want it to apply to all ranges, specify 0.0.0.0/0
  • Specified protocols and ports:
    • tcp:8080 (para abrir el puerto 8080)



Service Accounts

A service account is a special account whose credentials you can use in your application code to access other Google Cloud Platform services.

Creating and Managing Service Accounts: https://cloud.google.com/iam/docs/creating-managing-service-accounts#creating_a_service_account

Creating and Enabling Service Accounts for Instances: https://cloud.google.com/compute/docs/access/create-enable-service-accounts-for-instances#createanewserviceaccount

Granting Roles to Service Accounts: https://cloud.google.com/iam/docs/granting-roles-to-service-accounts

Por ejemplo, para poder copar arvhivos desde VM Instance hacia un Bucket, you need to Setting up a new instance to run as a service account. Dicha service account tiene que haber sido Granted with the particular Roles que permiten dicha operación de copia.

To grant the roles, no puede hacerlo a través de la línea de comandos pues me generó un error. Lo pude sin embargo hacer desde la página de GCP. para la operación de copia descrita arriba hay que agregar a la service account en la cual corre la VM los roles relacionados con Storage:

  • Storage Admin
  • Storage Object Admin
  • Storage Object Creator
  • Storage Object Viewer

Luego de haber creado la Service Account podemos, al momento de crear la VM Instance, ajustar que la new instance runs as this service account (ver ).



Live Migration of a Virtual Machine

  • The virtualBox machine that you migrate should be using a VMDK (Virtual Machine Disk), not VDI (VirtualBox Disk Image) or VHD (Virtual Hard Disk).
  • Make sure you the NIC (Network Interface Controler) (Network > Adapter 1 > Attached to) is set to NAT so that you have Internet access.
  • Es muy importante tomar en cuenta que para podernos conectar (a través de gcloud compute ssh VMmigrada) a la VM Instance luego de haberla migrado, el paquete openssh-server tiene que haber sido previamente instalado en la VM.

Para realizar la migración:

  • In the GCP
    • Main Menu > Compute engine > VM Instances: Import VM
    • There you will be redirected to the Cloud Endure Platform (VM Migration Service)


  • In the Cloud Endure Platform (VM Migration Service)
https://gcp.cloudendure.com
  • GCP credentials
We need to create GCP credentials (Google Cloud Platform JSON private key (JSON file)). To generate a JSON private key file (could be also PKCS12 format) (Creo que estas credenciales cambiarían sólo si queremos realizar una migración hacia otro proyecto):
  • Click "Find out where to get these credentials"
  • Lo anterior abrirá la página de Documentation with a link to "Open the link of Credentials"
  • Then, you need to choose the project: The target project where your VM will be migrate.
  • Debemos copiar el Project ID porque luego lo vamos a necesitar en la Cloud Endure Platforme
  • Click "Create credentials" and Select "Service Account Key" option
  • Select the JSON option (for key type)
  • The service account should be "Compute engine default service account"
  • Then hit create: the JSON file will be downloaded to your computer.
  • Luego, you go back to the Cloud Endure Platform (VM Migration Service)
  • Enter the Google Cloud Platform project ID
  • Luego cargamos el JSON file (Google Cloud Platform JSON private key). Then "Save".
  • Replication settings
  • Live migration target: Google EU West 2 (London)
  • Hit Save replication settings
  • Luego de esto: Project setup complete Congratulations!
  • Your Live Migration Project is set up. So, the next step is to install the CloudEndure Agent on your Source machine.


  • Install the CloudEndure Agent on your Source machine
How To Add Machines: https://gcp.cloudendure.com/#/project/2be028cc-40ae-4215-ad2c-52e3ce4e325d/machines
In order to add a machine to the GCP, install the CloudEndure Agent on your machine (your VirtualBox VM for example) (data replication (proceso de transferencia de la VM) begins automatically upon CloudEndure Agent installation).

Your Agent installation token: (este código lo genera automáticamente la Cloud Endure Platform (VM Migration Service. No sé si es diferente en cada migración.. Creo podría depender del proyecto; entonces debería cambiar si queremos migrar una VM hacia otro proyecto)

0E65-51C9-50EE-E20E-A583-9709-1653-BA57-457D-92BF-57B9-0DE2-22DC-EA11-3B80-81E6

For Linux machines:

Download the Installer:

wget -O ./installer_linux.py https://gcp.cloudendure.com/installer_linux.py

Antes de ejecutar el siguiente paso tuvimos que instalar:

sudo apt-get install python2.7
sudo apt-get install gcc

Then run the Installer and follow the instructions:

sudo python ./installer_linux.py -t 0E65-51C9-50EE-E20E-A583-9709-1653-BA57-457D-92BF-57B9-0DE2-22DC-EA11-3B80-81E6 --no-prompt

For Windows machines:

Download the Windows installer here https://gcp.cloudendure.com/installer_win.exe, then launch as follows:

installer_win.exe -t 0E65-51C9-50EE-E20E-A583-9709-1653-BA57-457D-92BF-57B9-0DE2-22DC-EA11-3B80-81E6 --no-prompt


  • Data replication
  • Data replication begins automatically once the installation of the CloudEndure Agent is completed. You will see able to see progress on the Cloud Endure Platform (VM Migration Service)
  • Data replication es de hecho el proceso en el cual la VM es copiada hacia la GCP.


  • Launch target Machine
  • Luego de haberse completado la Data replication, se habrá creado una Instance en la página de VM Instace del correspondiente proyecto en la GCP. Esta Instace, cuyo nombre contiene "replicator" no será la VM Instance definitiva. Debemos entonces hacer click en "Launch target Machine" (aún en la Cloud Endure Platform (VM Migration Service)), lo cual generará la VM Instance en la correspondiente página de la GCP. Habrán entonces dos Instance en la Página de la GCP. En mi caso:
  • ce-replicator-goo16-20eb5f29
  • ubuntu-18932dd6
  • Si hacemos cambios en nuestra Local VirtualBox VM, podremos actualizar la correspondiente Google VM Instance fácilmente haciendo click en "Launch target Machine" nuevamente. La ce-replicator-goo16-20eb5f29 creo que tiene la función de actuar en esta actualización.



Live migration of a desktop computer

Migrating your computer to Google Cloud using Cloud Endure



Create Bitnami WordPress Site on Google Cloud

Media:Create_Bitnami_WordPress_Site_on_Google_Cloud-Michael2018.pdf



Setting up IIS web site on Google cloud

Media:Setting_up_IIS_web_site_on_Google_cloud-Michael2018.pdf



AWS

Overview of Amazon Web Services: https://d1.awsstatic.com/whitepapers/aws-overview.pdf



AWS Academy

File:AWS_Academy_on_Vocareum_and_VitalSource_Bookshelf.pdf

AWS Academy uses three learning platforms to provide you with access to your learning resources. This guide shows you how to register on each platform and access the learning resources. There are three steps to registering on your class:

  • Register in the AWS Training & Certification Portal
  • Complete registration in Vocareum
  • Complete registration in VitalSource Bookshelf


  • AWS Training & Certification Portal
The AWS Training & Certificaiton Portal is our Learning Mangement System. It hosts videos, eLearning modules, and knowledge assessments:
Resources:
  • Video introductions
  • Video console demos
  • Narrated lectures
  • Knowledge checks


  • Registering on Your Class:
  • This is the general link: https://www.aws.training/SignIn
  • We can Create an account normally. Creo que no requiere ninguna autorización especial.
  • However, your educator could provide you with a registration link. This is the links provided in the course:
https://www.aws.training/SignIn?returnUrl=%2fUserPreferences%2fRegistration%3ftoken%3d4WQW9dOyY0uFHfKBMIhwSA2-JMU1JT1RLCVP3%26returnUrl%3d%252Foneclickregistration%253Fid%253D35854%2526returnUrl%253D%25252Foneclickregistration%25253Fid%25253D39997


  • Vocareum:
https://labs.vocareum.com/home/login.php
Your educator could provide you with a registration link:
Resources:
  • Lab exercises
  • Projects
  • Activities


  • VitalSource BookShelf:
https://bookshelf.vitalsource.com/
We can Create a VitalSource account normally. Creo que no requiere ninguna autorización especial; aunque la guía dice "Your educator will register you in VitalSource Bookshelf"
Resources:
  • Student materials



AWS Global Infrastructure

AWS Global infrastructure can be brooked down into three elements:

  • Regions
  • Availability zones
  • Edge Locations


AWS-Global infrastructure.png
AWS Region.png


  • AWS Regions:
An AWS Region is a geographical area. Each Region is made up of two or more Availability Zones.
AWS has 18 Regions worldwide.
You enable and control data replication across Regions.
Communication between Regions uses AWS backbone network connections infrastructure.


  • AWS Availability Zones:
Each Availability Zone is:
Made up of one or more data centers.
Designed for fault isolation.
Interconnected with other Availability Zones using high-speed private links.
You choose your availability zones. AWS recommends replicating across Availability Zones for resiliency.


  • AWS Edge Locations:
An Edge Location is where users access AWS services.
It is a global network of 114 points of presence (103 Edge Locations and 11 regional Edge Caches) in 56 cities across 24 countries.
Specifically used with Amazon CloudFront, a Global Content Delivery Network (CDN), to deliver content to end users with reduced latency.
Regional edge caches used for content with infrequent access.


AWS Regions and number of Availability Zones.png
Edge locations-Regional edge caches.png


  • AWS Infrastructure Features:
  • Elastic and Scalable:
Elastic infrastructure; dynamic adaption of capacity.
Scalable infrastructure; adapts to accommodate growth.
  • Fault-tolerant:
Continues operating properly in the presence of a failure.
Built-in redundancy of components.
  • High availability:
High level of operational performance
Minimized downtime
No human intervention



AWS Core Services

AWS Foundational/Core Services

https://www.slideshare.net/AmazonWebServices/aws-core-services-overview-immersion-day-huntsville-2019


AWS has a lot of services, but some of them, provides the foundation for all solutions; we refer to those as the Core Services.


AWS-Foundation services.png


AWS-Foundation services-Services and Categories.png



Compute

  • Amazon Elastic Compute Cloud (EC2)
  • EC2 is essentially a computer in the cloud. EC2 is a Web Service that provides compute capacity in the cloud (Virtual computing environment in the cloud).
  • EC2 allows you to create and configure virtual compute capacity.
  • Elastic refers tot the fact that if properly configured, you can increase or decrease the amount of servers required by an application automatically, according to the current demands on that application.
  • Compute refers to the compute, or server resources that are being presented.


  • AWS Lambda:
  • AWS Lambda is an event-driven, serverless computing platform. It is a computing service that runs code in response to events and automatically manages the computing resources required by that code.
  • AWS Lambda lets you run code without provisioning or managing servers. You pay only for the compute time you consume - there is no charge when your code is not running.
  • The purpose of Lambda, as compared to AWS EC2, is to simplify building smaller, on-demand applications that are responsive to events and new information. AWS targets starting a Lambda instance within milliseconds of an event. Node.js, Python, Java, Go[2], Ruby[3] and C# (through .NET Core) are all officially supported as of 2018. In late 2018, custom runtime support[4] was added to AWS Lambda giving developers the ability to run a Lambda in the language of their choice.


  • Auto Scaling:
  • Scales EC2 capacity as needed
  • Improves availability


  • Elastic Load Balancer:
  • Elastic Load Balancing automatically distributes incoming application traffic across multiple targets, such as Amazon EC2 instances, containers, IP addresses, and Lambda functions. It can handle the varying load of your application traffic in a single Availability Zone or across multiple Availability Zones.


  • AWS Elastic Beanstalk:
  • AWS Elastic Beanstalk is an easy-to-use service for deploying and scaling web applications and services developed with Java, .NET, PHP, Node.js, Python, Ruby, Go, and Docker on familiar servers such as Apache, Nginx, Passenger, and IIS.
You can simply upload your code and Elastic Beanstalk automatically handles the deployment, from capacity provisioning, load balancing, auto-scaling to application health monitoring.



Storage

There are three broad categories of storage:
  • Instance storage: Temporary storage that is added to your Amazon EC2 instance


  • Amazon Elastic Block Store (Amazon EBS): It is persistent, mountable storage, which can be mounted as a device to an Amazon EC2 instance. Amazon EBS can only be mounted to an Amazon EC2 instance within the same Availability zone.
Each Amazon EBS volume is automatically replicated within its Availability Zone to protect you from component failure, offering high availability.
  • Uses:
  • Boot volumes and storage for Amazon EC2 instances
  • Data Storage with a file system
  • Database hosts
  • Enterprise applications.
  • Volume types:
  • Solid-State Drives (SSD)
  • Hard Disk Drives (HDD)


  • Amazon Simple Storage Service (Amazon S3): Similar to Amazon EBS. It is persistent storage; however, it can be accessed from anywhere.
  • Object level storage
  • Amazon S3 con even be configured to support corss-region replication such that data put into an Amazon S3 bucket in one region can be automatically replicated to another Amazon S3 Region.
  • Uses:
  • Storing application assets
  • Static web hosting
  • Backup and Disaster Recovery (DR)
  • Staging area for big data


Another Storage options:
  • Amazon Elastic File System (Amazon EFS)
Provides simple, scalable, elastic file storage for use with AWS services and on-premises resources. It is easy to use and offers a simple interface that allows you to create and configure file systems quickly and easyly.
Amazon EFS is build to elastically scale on demand withou disrupting applications, growing and shrinking automatically as you add and remove files, so your applications have the storage they need, when they need it.


  • Amazon Glacier:
  • Amazon Glacier is a secure, durable, and extremely low-cost cloud storage service for data archiving and long-term backup.
  • Data stored in Amazon Glacier takes several hours to retrieve, which is why it's ideal for archiving.


  • Amazon Relational Database Service (RDS)


  • Amazon DynamoDB


  • Amazon Redshift


  • Amazon Aurora



Block vs Object Storage

What if you want to change one character in a 1 GB file?:

  • With block storage, you only need to change the block that contains the character.
  • With object storage, the entire file must be updated.

This difference has a mejor impact on the throughput latency, and cost of your storage solution. Block storage solutions are typically faster and use less bandwidth, but cost more than object-level storage.

Block vs Object Storage.png



Analytics

  • Amazon Athena is an interactive query service that makes it easy to analyze data in Amazon S3 using standard SQL.


  • Amazon EMR provides a managed Hadoop framework that makes it easy, fast, and cost-effective to process vast amounts of data across dynamically scalable Amazon EC2 instances.


  • Amazon CloudSearch is a managed service in the AWS Cloud that makes it simple and cost-effective to set up, manage, and scale a search solution for your website or application. Amazon CloudSearchsupports 34 languages and popular search features such as highlighting, autocomplete, and geospatial search


  • Amazon Elasticsearch Service makes it easy to deploy, secure, operate, and scale Elasticsearch to search, analyze, and visualize data in real-time. With Amazon Elasticsearch Service,you get easy-to-use APIs and real-time


  • Amazon Kinesis makes it easy to collect, process, and analyze real-time, streaming data so you can get timely insights and react quickly to new information.


  • Amazon Redshift is a fast, scalable data warehouse that makes it simple and cost-effective to analyze all your data across your data warehouse and data lake.


  • Amazon QuickSight is a fast, cloud-powered business intelligence (BI) service that makes it easy for you to deliver insights to everyone in your organization. QuickSight lets you create and publish interactive dashboards that can be accessed from browsers or mobile devices.


  • AWS Data Pipeline is a web service that helps you reliably process and move data between different AWS compute and storage services, as well as on-premises data sources, at specified intervals.


  • AWS Glue is a fully managed extract, transform, and load (ETL)service that makes it easy for customers to prepare and load their data for analytics


  • AWS Lake Formation is a service that makes it easy to set up a secure data lake in days. A data lake is a centralized, curated, and secured repository that stores all your data, both in its original form and prepared for analysis.



AWS Platform

https://aws.amazon.com/

AWS Accounts Include
12 Months of Free Tier Access
Including use of Amazon EC2, Amazon S3, and Amazon DynamoDB
Visit aws.amazon.com/free for full offer terms


  • Connecting to Your Linux Instance Using SSH:
https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/AccessingInstancesLinux.html
@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@
@         WARNING: UNPROTECTED PRIVATE KEY FILE!          @
@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@
Permissions 0644 for 'amazonec2.pem' are too open.
It is recommended that your private key files are NOT accessible by others.
This private key will be ignored.
bad permissions: ignore key: amazonec2.pem
Permission denied (publickey).

The above error can be solved this way: https://stackoverflow.com/questions/8193768/trying-to-ssh-into-an-amazon-ec2-instance-permission-error



Pricing

https://aws.amazon.com/ec2/pricing/on-demand/

t2.micro 1 Variable 1 GiB EBS Only $0.0116 per Hour:

$ 100.224 / y

t2.small 1 Variable 2 GiB EBS Only $0.023 per Hour:

$ 198.72 / y



Allow connections on HTTP 80, HTTPS 443 or other port

For example, if we host a website on an EC2 instance and want apache to serve the pages.

For HTTP 80, HTTPS 443 there are specific type of rules:


For any other port we can add a «Custom TCP» rule type:

Custom TCP Rule    TCP      8082      0.0.0.0/0


AWS inbound rule.png



OVH


  • Dirección IPv4 del VPS: 37.59.121.119
  • Dirección IPv6 del VPS: 2001:41d0:0051:0001:0000:0000:0000:18ae
  • Nombre del VPS: vps109855.ovh.net


Tengo la opción VPS Classic 2:


Estos son los datos relativos a su cuenta de cliente:

  • Su identificador  : va266899-ovh
  • Contraseña  : eptpi...



Add a subdomain

Domains > sinfronteras.ws:

  • DNS zone:
  • Add an entry:
Pointer records: A
Sub-domain: wiki.sinfronteras.ws
Target: 52.212.210.222
  • Add an entry (again) (optional):
Pointer records: CNAME
Sub-domain: www.wiki.sinfronteras.ws
Target: wiki.sinfronteras.ws