Arbor Tek SystemsOn-site

Data Engineer

Project-Based

Description

Job Title: Data Engineer Location: Washington, DC onsite day 1 Job Type: contract Duration: 12 months with possible extension Mandatory Key skills: Azure ADF/DBT/Data bricks/ Python:- Key Responsibilities Design, develop, and maintain scalable data pipelines using Azure Databricks and Apache Spark Implement complex data transformations using Spark SQL and Python (mandatory) Build and manage analytics models using DBT Core / DBT Cloud, following analytics engineering best practices Develop and orchestrate data ingestion and transformation workflows using Azure Data Factory Ensure data quality, reliability, and performance across batch and near real-time pipelines Integrate Databricks workloads with downstream reporting, analytics, and AI/ML use cases Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions Implement CI/CD pipelines and source control using Azure DevOps Apply secure data engineering practices aligned with WBG data governance and compliance standards Leverage AI-assisted tools to improve developer productivity, code quality, and operational efficiency Participate in design reviews, code reviews, and knowledge sharing sessions Mandatory Skills & Qualifications Strong expertise in Spark SQL and Python programming (Mandatory) Azure Databricks Data Engineer with at least one real-time implementation project Hands-on experience with DBT Core and/or DBT Cloud Experience building and managing pipelines using Azure Data Factory Proficiency with Azure DevOps (Repos, Pipelines, CI/CD) Experience using AI tools (e.g., AI-assisted development, data analysis, or automation tools) Solid understanding of data warehousing concepts, ETL/ELT patterns, and distributed data processing Ability to work in agile delivery models and collaborate with global teams Preferred / Good to Have Skills Databricks Certification (Associate or Professional) strongly preferred Experience with Azure data ecosystem (ADLS Gen2, Azure SQL, Synapse, etc.) Exposure to data governance, metadata management, and security frameworks Familiarity with performance tuning and cost optimization on Databricks Experience supporting analytics or AI/ML workloads on cloud platforms

Skills

ETLSecurityMLAzure DevOpsData EngineeringSparkPythonApache SparkMachine LearningSQLData AnalysisCI/CDComplianceDatabricksAgileAIdbtApacheDevOpsAzure

Want AI to find more roles like this?

Upload your CV once. Get matched to relevant assignments automatically.

Try personalized matching