Role & Responsibilities:
? Develop / Design effective and scalable solutions to administrate Data clusters, large-scale operations, and infrastructure systems.
? Architect systems, infrastructure and platforms using Linux and Amazon web services to support applications.
? Own and deliver the implementation of new methods for systems, deployment, monitoring, management, and automation.
? Technical depth – Exposure to a wide variety of problem-solving skills and respective automation.
? Devise schemes to transfer, monitor, and verify terabytes of data are moved from diverse locations, securely and reliably.
? Real-time problem diagnosis/resolution on live systems
? Monitor grid health and performance, use critical thinking to find areas for improvement, develop monitoring framework and metrics in order to predict system behaviour pro-actively and take appropriate steps.
? Capacity planning(cloud), provisioning new resources, ability to understand various capacity parameters and its cardinality
? Infrastructure and platform security.
? Infrastructure and platform cost management.
? Participate in oncall rotation using pager duty
? Minimum 5+ years’ experience in DevOps role:
? In-depth Linux/Unix knowledge, good understanding the various Linux kernel subsystems (memory, storage, network etc).
? Amazon Web Services 3. DNS, TCP/IP, Routing, HA & Load Balancing. Configuration management using tools like Ansible, Salt
? HA and Load balancing using tools like the Elastic Load Balancer and HAProxy.
? Monitoring tools like sensu, services like datadog, new relic
? Log management tools like Logstash/Syslog/ElasticSearch or similar.
? Metrics collection tools like Ganglia, Graphite, OpenTSDB or similar.
? Good understanding of distributed systems like kafka, zookeeper
? Good understanding of building immutable infrastructures using packer, terraform
? Automation experience using python/ruby/go
? Good understanding of linux containers(docker, coreos) and orchestration technologies like kubernetes , docker swarm.