We are seeking a Staff Software Engineer to join our team and build a cutting-edge application capable of parsing and analyzing terabytes of logs emitted from hybrid cloud environments. The ideal candidate will design and implement systems to extract meaningful patterns from logs and deliver actionable recommendations for cost savings at both the application and infrastructure levels.
Key Responsibilities:
– Design and build scalable systems to process and analyze logs and metrics from hybrid cloud environments (AWS, Azure, GCP, and on-premises).
– Develop algorithms to extract patterns and insights related to resource utilization, performance bottlenecks and cost-saving opportunities.
– Integrate the application with existing cloud or on-prem monitoring tools (e.g., Datadog, CloudWatch, Prometheus).
– Collaborate with cross-functional team to build dashboards and APIs to present actionable insights to end users.
– Collaborate with DevOps, cloud architects and data scientists to optimize hybrid cloud operations.
– Ensure data security and compliance with relevant regulations in multi-cloud environments.
Required Skills and Qualifications:
– Strong understanding of hybrid cloud environments (AWS, Azure, GCP and on-premises).
– Expertise in log management and processing tools (e.g., ELK stack, Apache Kafka, Fluentd, Splunk, Graylog).
– Proficiency in programming languages like Python, Go or Java for building scalable applications.
– Experience with big data frameworks (e.g., Apache Spark, Hadoop) for processing large-scale logs.
– Strong knowledge of cloud cost management tools and techniques.
– In-depth knowledge about containerization (e.g., Docker, Kubernetes), with hands-on experience deploying production applications using containerization.
– Expertise in observability platforms and tools across Kubernetes and cloud platforms, including monitoring, tracing and telemetry.
– Strong problem-solving skills and a focus on optimizing infrastructure for performance and cost.
Bonus or Good-to-Have Skills:
– Hands on experience with both open source and commercial cloud monitoring tools (e.g., Datadog, New Relic, Cloudwatch, Prometheus).
– Certifications in AWS, Azure, or GCP architecture and operations.
– Knowledge of sustainability practices related to cloud operations and carbon footprint optimization.
– Familiarity with machine learning techniques for anomaly detection and pattern recognition is a plus.
– Hands-on experience with data visualization tools (e.g., Grafana, Tableau)
Why Join Us?
- Be part of a dynamic team focused on transforming cloud operations through intelligent analytics.
- Work on high-impact projects in a fast-paced, innovative environment.
- Opportunity to shape the future of hybrid cloud sustainability and cost optimization.
How to Apply:
- First, read through all of the job details on this page.
- Scroll down and press the Click Here button.
- To be redirected to the official website, click on the apply link.
- Fill the details with the information provided.
- Before submitting the application, cross-check the information you’ve provided.