Summary
Overview
Work History
Education
Skills
Websites
Certification
Timeline
Generic

Umair Hussain

Fairfield-California

Summary

Accomplished DevOps Platform Engineer and Cloud Consultant with 15 years of expertise in architecting and managing robust cloud infrastructures on AWS and Azure. Demonstrated success in leading cross-functional teams to deliver high-availability solutions for a diverse range of clients, including enterprise-level organizations and startups. Proficient in cloud architecture, Infrastructure as Code, and CI/CD pipelines, complemented by extensive experience in containerization and monitoring frameworks. Committed to leveraging deep knowledge of Agile methodologies to drive continuous improvement and innovation within cloud operations.

Overview

19
19
years of professional experience
1
1
Certification

Work History

DevOps Engineer

Compulse
Remote
04.2023 - Current
  • Configured and maintained AWS resources such as EC2, RDS, S3, Lambda, and VPCs to support scalable infrastructure, ensuring high availability, and fault tolerance for critical applications.
  • Continuously monitored and optimized AWS usage through services like AWS CloudWatch, AWS Trusted Advisor, and Cost Explorer, achieving a balance between performance and cost efficiency.
  • Managed user onboarding and off boarding processes, configuring Single Sign-On (SSO) for various applications to enhance security and streamline access.
  • Led the design and implementation of automated CI/CD pipelines, primarily using Jenkins and GitHub Actions, to streamline software development and deployment workflows.
  • This automation helped improved delivery times and reduction in production errors
  • Worked on creating CI/CD pipelines for Windows-related workloads using TeamCity and Octopus Deploy.
  • Worked on AWS CodePipeline and CodeDeploy to implement automated, zero-downtime deployment strategies, improving system uptime, and ensuring smooth production releases.
  • Managed cloud infrastructure using Terraform to maintain consistency, scalability, and automation.
  • In some cases, I used CloudFormation to meet specific infrastructure requirements.
  • Designed and implemented a highly available and fault-tolerant infrastructure across multiple AWS Availability Zones using services such as Elastic Load Balancers (ELB), Auto Scaling, and RDS Multi-AZ configuration.
  • Set up comprehensive monitoring with AWS CloudWatch, integrated with Prometheus and Grafana, to track key metrics, troubleshoot issues, and ensure system performance at all times.
  • Deployed AWS CloudWatch Logs and the ELK Stack (Elasticsearch, Logstash, Kibana) for centralized logging, providing real-time insight into application and infrastructure logs to detect and resolve issues faster.
  • Built and deployed containerized applications using Docker on AWS ECS and EKS, enabling efficient management of microservices architecture, and facilitating rapid scaling.
  • Configured Grafana with Prometheus to establish a robust monitoring solution for our infrastructure
  • This involved setting up Prometheus to scrape metrics from various services and applications, using a custom Prometheus configuration file to define the scrape interval and target endpoints.
  • After installing Grafana, I integrated it with Prometheus as a data source, enabling real-time visualization of the collected metrics.
  • Created custom visualizations and alerts to monitor key performance indicators, system health, and application metrics effectively, using Grafana dashboard capabilities.
  • Used Rancher to manage and orchestrate Kubernetes clusters, enabling efficient deployment, scaling, and management of containerized applications.
  • Engaged Rancher's UI and automation capabilities to simplify cluster management across multi-cloud environments, ensuring streamlined operations and enhanced visibility into Kubernetes workloads.
  • Deployed Helm as a package manager for Kubernetes, enabling efficient deployment and management of applications.
  • Experienced in creating and managing Helm charts, utilizing templating for easy configuration.
  • Integrated Wiz with AWS to enhance security practices by enabling continuous monitoring of cloud resources, such as EC2, S3, EKS, and IAM services.
  • Configured a secure, read-only IAM role through AWS CloudFormation, allowing Wiz to scan for vulnerabilities, misconfigurations, and compliance risks across the infrastructure.
  • Utilized VictorOps for managing alerts, ensuring timely notification and response to critical incidents across infrastructure and application environments.
  • Configured alert routing, escalation policies, and integrated VictorOps with monitoring tools to streamline incident management, and improve system reliability.
  • Automated routine operational tasks through scripting in Bash, Python, and PowerShell, enhancing system reliability and team productivity.
  • Administered user and group management on Databricks, including provisioning and de-provisioning access, assigning roles, and configuring workspace permissions.
  • Simplified user onboarding and offboarding processes.

DevOps Engineer

nClouds
San Francisco
01.2018 - 03.2023
  • Designed and deployed scalable infrastructure on AWS and Azure for multiple clients, including enterprises such as healthcare, asset management, and financial institutions, as well as startups.
  • Architected secure cloud environments adhering to HITRUST CSF standards, ensuring the protection of sensitive data.
  • Developed and enforced security policies and procedures to meet HITRUST certification requirements, enhancing data security and compliance.
  • Responsible for end-to-end design and implementation of CI/CD pipelines using Jenkins, CircleCI, AWS CodePipeline, GitLab, and Bamboo.
  • Engineered a CI/CD pipeline using AWS CDK for a healthcare customer, integrating AWS CodePipeline and CodeBuild to reduce deployment time and improve code quality through automated testing.
  • Implemented and integrated SonarQube with the CI/CD pipeline for a New York-based enterprise customer, and also managed SonarQube instances for automated code analysis.
  • Architected Nexus Repository Manager, integrated Nexus with CI/CD tools.
  • Automated the infrastructure of a healthcare customer using AWS CDK, cutting down manual provisioning tasks, and enabling swift and consistent deployment.
  • Automated deployment pipeline using Octopus Deploy for multiple web applications and services.
  • Managed TeamCity build servers, agents, and projects for the automation of CI/CD.
  • Integrated Red Hat Cluster Suite with cloud platforms, including AWS and Azure, to enable hybrid cloud deployments with seamless failover capabilities.
  • Managed and monitored Red Hat Cluster Suite environments, performing tasks like cluster resource configuration, failover testing, and quorum settings.
  • Configured and deployed Red Hat Cluster Suite for high availability across multi-node environments, ensuring fault tolerance, and minimizing downtime.
  • Configured and optimized TeamCity build configuration, including VCS roots, build steps, and triggers, to ensure efficient, reliable builds.
  • Designed and implemented Azure-based solutions for continuous integration and deployment (CI/CD), leveraging services such as Azure DevOps (formerly VSTS) pipelines, Azure Repos, and Azure Artifacts to automate software delivery processes.
  • Implemented Infrastructure as Code (IaC) solutions using tools such as Terraform and AWS CloudFormation for AWS, and Azure Resource Manager (ARM) templates for Azure, automating the provisioning and management of cloud infrastructure to improve scalability, reliability, and consistency across multi-cloud environments.
  • Migrated on-premises infrastructure to the cloud, implementing Infrastructure as Code (IaC) with Terraform to automate the provisioning of compute instances, networking components, and storage resources.
  • Designed and executed data migration for a healthcare client from on-premise to AWS S3, which helped in reducing storage costs by 30%.
  • Used AWS Database Migration Service (DMS) to migrate 65 on-premise databases to AWS RDS, which improved database performance.
  • Migrated 24 on-premise applications to AWS Elastic Beanstalk for a reduction in deployment time and improved application performance.
  • Created a design and implementation of an Azure partner using Docker and Kubernetes to migrate six on-premise applications to Azure Kubernetes Service (AKS), achieving almost a 90% reduction in deployment time and an overall improvement in application scalability.
  • Worked on a hybrid cloud migration for an automotive customer from on-premise to AWS and Azure.
  • Migrated 175 on-premise servers to AWS EC2 for an asset management customer to reduce costs and for high availability.
  • Migrated on-premises VMware workloads to AWS, utilizing VMware Cloud on AWS and its services, including EC2, S3, and RDS.
  • Implemented a Kubernetes cluster for a healthcare customer for high availability and scalability of their applications.
  • Conducted Kubernetes deployments, services, and ingress controllers for the web applications and APIs.
  • Developed and maintained AWS Glue ETL workflows to extract, transform, and load data from S3, Amazon RDS, and AWS DynamoDB.
  • Implemented the AWS Glue job scheduler to automate data processing workflows.
  • Configured and managed VMware vCenter, ESXi hosts, and VMs, including resource allocation, networking, and storage configuration.
  • Integrated VMware NSX with AWS networking services for establishing efficient communication between on-premises and cloud environments.
  • Designed and implemented scalable GraphQL APIs for multiple microservices-based systems.
  • Developed and maintained GraphQL schemas, resolvers, and data models to support complex queries.
  • Designed and executed Grafana dashboard implementation, offering immediate system performance insights, infrastructure status updates, and application metrics to enhance proactive monitoring and problem resolution.
  • Created intuitive and customizable dashboards using Grafana's rich visualization capabilities.
  • Enabled stakeholders to quickly identify trends, anomalies, and optimization opportunities within complex environments.
  • Monitored and optimized GraphQL performance using Prometheus, Grafana, and New Relic.
  • Designed and implemented Azure networking solutions, including virtual networks (VNets), Azure ExpressRoute, and Azure VPN Gateway, to establish secure and scalable connectivity between on-premises and cloud environments.
  • Architected a 6-node Elasticsearch cluster for an e-commerce customer, boosting search query performance.
  • Enhanced search capabilities through creating custom Elasticsearch plugins.
  • Secured sensitive customer data by configuring the Elasticsearch security feature and access controls across six nodes for ensuring compliance and data privacy.
  • I installed, configured, and managed the RHEL system, ensuring compliance with security standards, and optimizing system performance.
  • Configured and managed RHEL clustering and high-availability solutions.
  • Configured and managed AWS Landing Zone core services, including AWS Organization, AWS IAM, AWS S3, and AWS VPC.
  • Successfully implemented Kafka clusters on AWS, utilizing Terraform and Ansible for installation and deployment.
  • Configured and managed Kafka topics, partitions, and replication factors for high availability.
  • Integrated Kafka with Apache Hadoop to streamline data integration across multiple systems.
  • Troubleshot and resolved Kafka issues through the utilization of the Kafka console, the Kafka GUI, and log analysis.
  • Architected and deployed infrastructure automation utilizing Chef, Puppet, and Ansible.
  • Established datadog monitoring system to enhance server performance across multiple platforms
  • Integrated CloudCheckr, Datadog, and Splunk dashboards with AWS accounts.
  • Utilized other monitoring tools, including New Relic, Amazon CloudWatch, and Azure Monitor, as well for clients that were not using Datadog.
  • Created customized Sentinel policies to address specific security and compliance requirements for cloud infrastructure.
  • Created policies and procedures to align the SOC framework, ensuring the consistent application of security practices.
  • Worked on AWS Secrets Manager to securely store and retrieve secrets, ensuring data privacy.
  • Built a serverless infrastructure for healthcare by deploying AWS Lambda functions.
  • Automated real-time data workflows using AWS Lambda for reducing manual data handling, and for timely updates for critical health metrics.
  • Implemented the deployment of containerized applications on various cloud platforms, including ECS, EKS, Azure Container Instances, and AKS.
  • Designed and implemented automated patch management solutions, utilizing AWS Systems Manager and Azure Update Management Center.
  • Designed and implemented disaster recovery solutions of varying complexity in accordance with the RTO/RPO requirements of each workload.
  • Developed custom scripts and integrations to extend the functionality of Nexus, Tosca, and SonarQube, integrating them into existing toolchains and workflows to streamline automation processes.
  • Orchestrated complex data workflows using Apache Airflow, ensuring reliable and timely execution of ETL processes.
  • Successfully deployed data pipelines with Dagster, utilizing its unique abstractions to improve testability and maintainability.
  • Worked on the provisioning and management of Azure infrastructure using tools like Azure Resource Manager (ARM) templates and Azure CLI, ensuring infrastructure as code (IaC) principles for consistency, scalability, and repeatability.
  • I dabbled in database migration using AWS DMS and Percona Xtrabackup.
  • Setup and administration of various database management systems, including MySQL, PostgreSQL, MSSQL, Amazon Aurora, MongoDB, DynamoDB, and Redis.
  • Implemented Azure Monitor and Application Insights to gain visibility into application performance, monitor system health, and proactively identify and troubleshoot issues, ensuring high availability and reliability of Azure-based services.
  • Implemented Azure Automation for automating routine tasks and workflows, leveraging PowerShell scripts and Azure Automation Runbooks to streamline operational processes and reduce manual effort.
  • Implement log aggregation, visualization, and analytics using Logz.io, Datadog, Graylog, and ELK.
  • Created Apache Directory Server for local networks, and integrated RHEL instance with Active Directory in AWS VPC.
  • Developed Puppet manifests and custom modules to securely transfer retrieved secrets from AWS Secrets Manager to Password Manager Pro.
  • Utilized Puppet's built-in file transfer capabilities to ensure the encrypted and secure transmission of sensitive data to Password Manager Pro.
  • Created log collection ELK: Elasticsearch, Logstash, and Kibana installed. Filebeat is installed on all nodes in the cluster to send log data to Logstash.

DevOps Engineer

Harosec
San Francisco
10.2014 - 12.2017
  • Designed and deployed scalable infrastructure on AWS
  • Managed the entire lifecycle of CI/CD pipelines from design to implementation using Jenkins, CircleCI, AWS CodePipeline, GitLab, and Bamboo.
  • Designed and implemented a highly available and scalable Linux-based infrastructure for cloud-native apps.
  • Led the migration of VMware workloads to AWS using AWS Server Migration Services (SMS) and Migration Hub.
  • Performed disaster recovery procedures to ensure seamless operations.
  • Successfully led and supervised a team of 17 senior and junior DevOps engineers.
  • Integrated Opsgenie with Jenkins to streamline operational processes and enhance efficiency.
  • Integrated Sentry with the incident management system.
  • Worked on the IDP development environment setup using Spring Tool Suite, Docker, and Kubernetes technologies for creating Docker containers and deploying them into Kubernetes clusters.
  • Used Sysdig Secure and Sysdig Monitor to continuously monitor the performance, availability, and compliance of multiple EKS and Kubernetes clusters.
  • Used Sentinel policies to automate security and compliance checks, reducing manual effort, and improving the efficiency of cloud operations.
  • Integrated Sysdig Image Vision image scanning in a Jenkins-based CI/CD pipeline to automate the local scanning of images pre-deployment.
  • Automated infrastructure utilizing Chef, Puppet, and Ansible.
  • Setting up infrastructure provisioning using Packer, streamlining deployment processes, and reducing deployment time.
  • Implemented modular configurations, leveraging Packer's templates to create reproducible and consistent machine images across development, testing, and production environments.
  • Used Terraform, Ansible, Chef, and CloudFormation extensively for configuration management and infrastructure as code (IaC).
  • Diagnosed and troubleshot UNIX and Windows processing problems, and applied solutions to increase company efficiency.
  • Deployed containerized apps on ECS, EKS, Azure Container Instances, and AKS.
  • Developed serverless applications using advanced cloud technologies such as AWS Lambda, Step Functions, Azure Functions, and Logic Apps to enhance business efficiency.
  • Deployed applications on Azure App Services and on VMs.
  • Tracked issues with Jira, and documented them using Confluence.
  • Managed applications on Django, PHP, Java, and Tomcat.
  • Enhanced monitoring capabilities through the implementation of tools like Datadog, New Relic, Nagios/Icinga, AWS CloudWatch, and Azure Monitor.
  • Implemented logs aggregation, visualization, and analytics with Logz.io, Datadog, and ELK.

DevOps Engineer

Signet Media
San Francisco
10.2013 - 09.2014
  • Implement security features for multi-tenant environments built on virtualization technology
  • Manage a small, highly-skilled technical team that assists in handling the day-to-day activities and assists in executing the vision provided
  • Gained insight into system requirements by effectively communicating with staff and clients
  • Provide advice on project costs, design concepts, or design changes
  • Implemented AWS Code Pipeline and created CloudFormation JSON templates in Terraform for infrastructure as code
  • Checking code compatibility, standard, and functionality over Windows, UNIX/Linux environment
  • In order to support automated builds and deployments, I wrote Windows CMD shell scripts for automating builds and transfer scripts for promoting binaries to target systems using scripted WinSCP processes
  • Document design specifications, installation instructions, and other system-related information
  • Verified system architecture for stability, interoperability, portability, security, and scalability
  • Collaborate with engineers or software developers to select appropriate design solutions, or ensure the compatibility of system components
  • Evaluate current or emerging technologies to consider factors such as cost, portability, compatibility, or usability
  • Provide technical guidance or support for the development or troubleshooting of systems
  • Identify system data, hardware, or software components required to meet user needs
  • Deploy containerized apps using Docker
  • Infrastructure configuration automation with Chef and Ansible
  • Implement infrastructure monitoring with Nagios
  • Build automated machine images using Packer

DevOps Engineer

Design Reactor
San Francisco
04.2010 - 09.2013
  • Spearheaded engineering and provisioning initiatives for SA projects.
  • Install new or rebuild existing servers, and configure hardware, peripherals, services, settings, directories, storage, etc.
  • In accordance with standards and project/operational requirements.
  • Installed and configured virtualization infrastructure applications and network management applications.
  • Developing and maintaining installation and configuration procedures.
  • Streamlined operations and support processes.
  • Performing regular security monitoring to identify any possible intrusions.
  • Performing daily backup operations, ensuring all required file systems and system data are successfully backed up to the appropriate media.
  • Recovery tapes or disks are created, and media is recycled and sent off-site as necessary.
  • Performing regular file archival and purge as necessary.
  • Maintaining operational, configuration, or other procedures.
  • Performing periodic performance reporting to support capacity planning.
  • Performing ongoing performance tuning, hardware upgrades, and resource optimization as required.
  • Configure the CPU, memory, and disk partitions as required.
  • Maintaining data center environmental and monitoring equipment.

I.T Manager

Info span
Islamabad
03.2006 - 03.2010
  • Manage and supervise IT teams located in multiple geographical locations, with multiple teams in each geographical location, performing multiple IT functions.
  • Manage IT operations, human resource objectives by recruiting, selecting, orienting, training, assigning, scheduling, coaching, counseling, and disciplining employees; communicating job expectations; planning, monitoring, appraising, and reviewing job contributions; planning and reviewing compensation actions; enforcing policies and procedures.
  • IT operations' operational objectives include maintaining the current system, evaluating, recommending, testing, and installing new technology, contributing information and recommendations to strategic plans and reviews, preparing and completing action plans, implementing production, productivity, quality, and customer service standards, resolving problems, completing audits, identifying trends, determining system improvements, and implementing change.
  • Forecasts IT operations' financial requirements to meet objectives, prepares the annual budget, schedules expenditures, analyzes variances, and initiates corrective actions.
  • Determined IT operations service requirements by analyzing the needs of users and departments, prioritizing modifications to core system applications, providing information by collecting, analyzing, and summarizing data and trends, and designing, developing, and implementing a Linux-based call center CRM and predictive dialer solution utilizing both VoIP and PSTN-based networks.
  • Oversee the design and implementation of Cisco-based core, distribution, and access layer network infrastructure, with fully redundant core and distribution layer implementation.

Education

MBA -

Comsats University
Islamabad, Pakistan
07.2005

BS Computer Sciences -

Allama Iqbal Open University
Islamabad, Pakistan
06.2003

Skills

  • AWS Management
  • CI/CD implementation
  • Infrastructure as code
  • Container orchestration
  • Monitoring solutions
  • Security compliance
  • Microservices architecture
  • Linux operating system
  • Cloud migration
  • Scripting languages

Certification

  • Certified AWS Solutions Architect (Associate)
  • Certified AWS Solutions Architect (Professional)

Timeline

DevOps Engineer

Compulse
04.2023 - Current

DevOps Engineer

nClouds
01.2018 - 03.2023

DevOps Engineer

Harosec
10.2014 - 12.2017

DevOps Engineer

Signet Media
10.2013 - 09.2014

DevOps Engineer

Design Reactor
04.2010 - 09.2013

I.T Manager

Info span
03.2006 - 03.2010

MBA -

Comsats University

BS Computer Sciences -

Allama Iqbal Open University
Umair Hussain