#https://www.se.com/ww/en/about-us/careers/job-details/lead-cloud-application-engineer/00939Q/
ied individuals regardless of race, religion, color, gender, disability, national origin, ancestry, age, military status, sexual orientation, marital status, or any other legally protected characteristic or conduct**
Job Profile: Lead Cloud Application Engineer
Location: Schneider Electric Bangalore.
Responsibilities:
Design and implement highly available and scalable Azure cloud data solutions that meet business requirements.
Collaborate with cross-functional teams to design and implement solutions that meet business requirements.
Developing, designing, and maintaining cloud applications, data processing applications, layouts, architectures, and relational/non-relational databases for data access.
Participates in the testing process through test review and analysis.
Education
• Bachelor's degree in Computer Science
Project Summary & Must Skillset
• We have azure databricks as data processing & analytics platform to build our solutions to ingest the data, process and store the data and Pyspark is required/must competency to develop data pipelines
• We highly develop different api services to process and expose the processed data and these api services are containerized and deployed on ACA/AKS and so Python is required/must skillset for this position. As we develop multiple api services in different programming languages, experience on c#, java, go-lang is definitely an advantage to the profile
• And for observability, we highly rely on open telemetry collectors and exporters. Open Telemetry skillset is add-on
Required Qualifications:
8+ years of experience as an Azure Cloud Application engineer
Solid programming experience in python to build cloud IIOT applications & api's and strong expertise on Pyspark to work with data processing applications.
Strong experience with Azure cloud resources eg., Iothub, datalake, blob storages, Cosmos DB, Eventhub, Servicebus, Topic, azure functions, AKS, ACA, redis cache, keyvault etc
Experience in streaming data platforms - Kafka
Experience in dockers, Kubernetes, Microservices, containerized apps.
Good working knowledge with Azure Infrastructure, authentication, authorization management
Terraform Experience to build CI/CD - Github actions & workflows
Design data ingestion, data processing solutions such as ETL, ELT's on a wide set of data storage solutions.
Flexible to learn on new programming languages/technologies based on project needs.
Experience working in an Agile development environment.
Familiar with a variety of the field's concepts, practices, and procedures and a strong focus on data integrity.
Demonstrates strong written and oral communications skills. Ability, desire, and focus to meet deadlines Demonstrates ability to work with all levels of individuals.
Demonstrates organizational skills while working on multiple projects; and communicate effectively within the team.
Independent and able to manage and prioritize workload.
Ability to guide or lead junior resources to get desired project results.
Advanced troubleshooting skills to drive to root cause
Ability to adapt quickly and positively to change.
Ability to manage ambiguity and solve undefined problems
Nice to Have:
Observability expertise with Dynatrace is add-on
We develop multiple api services in Python, Java, C#. Expertise on C#, java, go-lang is definitely an advantage, but Python is a must skill for this position.
Schedule: Full-time
Req: 00939Q