• You will design, develop, and maintain data pipelines, including solutions for data collection, cleansing, transformation, and usage, ensuring accurate data ingestion and readiness for downstream analysis, visualization, and AI model training.
• You will build frameworks to validate data quality and completeness, detect anomalies, enhance data pipeline resiliency, and support unit and integration testing.
• You will design and implement end-to-end software life cycles, encompassing design, development, CI/CD, service reliability, recoverability, and participation in agile development practices, including on-call rotation.
• You will write code to implement performance monitoring protocols across data pipelines, building visualizations and aggregations to monitor pipeline health. You’ll also implement solutions and self-healing processes that minimize points of failure across multiple product features.
• You will anticipate data governance needs, designing data modeling and handling procedures to ensure compliance with all applicable laws and policies.
• You will implement and enforce security and access control measures to protect sensitive resources and data.
• You will Perform database administration tasks, including maintenance, and performance monitoring.
Required/Minimum Qualifications
Bachelor's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 2+ years experience in business analytics, data science, software development, data modeling or data engineering work OR Master's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 1+ year(s) experience in business analytics, data science, software development, or data engineering work OR equivalent experience.
• 2+ years of experience in software or data engineering, with proven proficiency in C#, Java, or equivalent.
• 2+ years in one scripting language for data retrieval and manipulation (e.g., SQL or KQL).
• 2+ years of experience in big data scripting languages like Python, SCOPE, SQL or equivalent.
• 2+ years of experience with ETL and data cloud computing technologies, including Azure Data Lake, Azure Data Factory, Azure Synapse, Azure Logic Apps, Azure Functions, Azure Data Explorer, and Power BI or equivalent platforms.
Other Requirements
Ability to meet Microsoft, customer and/or government security screening requirements are required for this role. These requirements include, but are not limited to the following specialized security screenings: Microsoft Cloud Background Check:
Preferred/Additional Qualifications
• 1+ years of demonstrated experience implementing data governance practices, including data access, security and privacy controls and monitoring to comply with regulatory standards.
• Experience with big data technologies such as: Hadoop, Hive, Spark.
• Experience with distributed database engines like Microsoft Synapse.