Skip to main content

Describing TOAD - TripleO automated deployer

An overview of the TOAD framework and the advantages

What is TOAD?

What components make up TOAD?

Requirements to install TOAD

  • Two different use cases: virtualized and baremetal
  • Only one Jenkins+Nginx VM needed to spin up the platform (8gb at least) 
  • RHEL7 / Centos7 operating system

Virtualized deploys

  • One server for slave and virtualized TripleO deployment:
    • multi-core
    • 16GB mem (better 32GB)
    • 60GB disk
    • external network
  • Needs RHN subscription for OSP jobs.

Baremetal deploys

  • One server for slave and virtualized undercloud:
    • multi-core
    • 16GB mem
    • 60GB disk
    • external network
    • Needs RHN subscription for OSP jobs.
  • One server for controller and one for compute:
    • multi-core
    • 4GB memory
    • 60GB disk
    • IPMI support
    • independent NIC/VLANS for provisioning and administration.
Can be extended with more controller and computes for HA

How users benefit from TOAD?

  • Easy: perform full virtualized and baremetal deployments with one click
  • Automated: avoid manual steps that are prone to human errors
  • Repeatable: job definitions and configs stored in git repos, track changes
  • Battle tested: relies on TripleO quickstart, used for upstream CI
  • Complete: choose between different releases, and RDO/OSP deploys
  • Flexible: extend with customization scripts and templates. Consume local repos with pinned versions and custom packages
  • Visible: logs for all deployment steps are collected and published 

TOAD job deployment workflow 

Comments

  1. Nice blog... This blog is helpful for me to understand OpenStack development. Thanks for sharing information

    ReplyDelete

Post a Comment

Popular posts from this blog

Setup an NFS client provisioner in Kubernetes

Setup an NFS client provisioner in Kubernetes One of the most common needs when deploying Kubernetes is the ability to use shared storage. While there are several options available, one of the most commons and easier to setup is to use an NFS server.
This post will explain how to setup a dynamic NFS client provisioner on Kubernetes, relying on an existing NFS server on your systems.
Step 1. Setup an NFS server (sample for CentOS) First thing you will need, of course, is to have an NFS server. This can be easily achieved with some easy steps:

Install nfs package: yum install -y nfs-utils Enable and start nfs service and rpcbind:
systemctl enable rpcbind
systemctl enable nfs-server
systemctl start rpcbind
systemctl start nfs-server
Create the directory that will be shared by NFS, and change the permissions:
mkdir /var/nfsshare
chmod -R 755 /var/nfsshare
chown nfsnobody:nfsnobody /var/nfsshare
 Share the NFS directory over the network, creating the /etc/exports file:
vi /etc/exports
/var/nfsshare …

Create and restore external backups of virtual machines with libvirt

A common need for deployments in production, is to have the possibility of taking backups of your working virtual machines, and export them to some external storage.
Although libvirt offers the possibility of taking snapshots and restore them, those snapshots are intended to be managed locally, and are lost when you destroy your virtual machines.
There may be the need to just trash all your environment, and re-create the virtual machines from an external backup, so this article offers a procedure to achieve it.
First step, create an external snapshot So the first step will be taking an snapshot from your running vm. The best way to take an isolated backup is using blockcopy virsh command. So, how to proceed?

1. First you need to extract all the disks that your vm has. This can be achieved with domblklist command:
DISK_NAME=$(virsh domblklist {{domain}} --details | grep 'disk' | awk '{print $3}')

This will extract the name of the device that the vm is using (vda, hda, et…

Automating local mirrors creation in RHEL

Sometimes there is a need to consume RHEL mirrors locally, not using the Red Hat content delivery network. It may be needed to speed up some deployment, or due to network constraints.

I create an ansible playbook, rhel-local-mirrors (https://github.com/redhat-nfvpe/rhel-local-mirrors), that can help with that.
What does rhel-local-mirrors do? It is basically a tool that connects to the Red Hat CDN, and syncs the repositories locally, allowing to populate the desired mirrors, that can be accessed by other systems via HTTP.

The playbook is performing several tasks, that can be run together or independently:
register a system on the Red Hat Networkprepare the system to host mirrorscreate the specified mirrorsschedule automatic updates of the mirrors How to use it?It is an Ansible playbook, so start by installing it, in any prefered format. Then continue by cloning the playbook:
git clone https://github.com/redhat-nfvpe/rhel-local-mirrors.gitThis playbook expects a group of servers called