Arbor Tek SystemsOn-site

Actively looking for a Data Engineer

Project-Based

Description

Role: Data Engineer Location: Washington, DC onsite day 1(Need only locals) Mandatory Key skills: Azure ADF/DBT/Data bricks/ Python Mail: Key Responsibilities Design, develop, and maintain scalable data pipelines using Azure Databricks and Apache Spark Implement complex data transformations using Spark SQL and Python (mandatory) Build and manage analytics models using DBT Core / DBT Cloud, following analytics engineering best practices Develop and orchestrate data ingestion and transformation workflows using Azure Data Factory Ensure data quality, reliability, and performance across batch and near real-time pipelines Integrate Databricks workloads with downstream reporting, analytics, and AI/ML use cases Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions Implement CI/CD pipelines and source control using Azure DevOps Apply secure data engineering practices aligned with WBG data governance and compliance standards Leverage AI-assisted tools to improve developer productivity, code quality, and operational efficiency Participate in design reviews, code reviews, and knowledge sharing sessions Mandatory Skills & Qualifications Strong expertise in Spark SQL and Python programming (Mandatory) Azure Databricks Data Engineer with at least one real-time implementation project Hands-on experience with DBT Core and/or DBT Cloud Experience building and managing pipelines using Azure Data Factory Proficiency with Azure DevOps (Repos, Pipelines, CI/CD) Experience using AI tools (e.g., AI-assisted development, data analysis, or automation tools) Solid understanding of data warehousing concepts, ETL/ELT patterns, and distributed data processing Ability to work in agile delivery models and collaborate with global teams Preferred / Good to Have Skills Databricks Certification (Associate or Professional) strongly preferred Experience with Azure data ecosystem (ADLS Gen2, Azure SQL, Synapse, etc.) Exposure to data governance, metadata management, and security frameworks Familiarity with performance tuning and cost optimization on Databricks Experience supporting analytics or AI/ML workloads on cloud platforms

Skills

ETLMLSecuritydbtDatabricksAzure DevOpsAgileCI/CDAIPythonDevOpsApacheData AnalysisSQLApache SparkData EngineeringComplianceAzureMachine LearningSpark

Want AI to find more roles like this?

Upload your CV once. Get matched to relevant assignments automatically.

Try personalized matching