Rio de Janeiro, BRA
16 hours ago
Data Engineer - Platform & Pipelines
Data Engineer - Platform & Pipelines Date: Jan 22, 2026 Location: Rio de Janeiro, RJ, BR, 291941 We are looking for the right people — people who want to innovate, achieve, grow and lead. We attract and retain the best talent by investing in our employees and empowering them to develop themselves and their careers. Experience the challenges, rewards and opportunity of working for one of the world’s largest providers of products and services to the global energy industry. Job Duties We are implementing a strictMedallion Architectureto organize petabytes of industrial data. This role is for a Data Engineer who excels at transforming raw chaos into structured, queryable assets. You will build and maintain the ELT pipelines that move data from "Bronze" (Raw) to "Silver" (Cleaned) and "Gold" (Aggregated). You will work with Delta Lake (On-prem/Databricks), PolarsandAirflowto ensure data quality and availability for Data Scientists and the Knowledge Graph. What You’ll Do + Pipeline Development:Develop and maintain robust Airflow DAGs to orchestrate complex data transformations. + Data Transformation:Use Spark (when scale requires) and Polars to clean, enrich, and aggregate data according to business logic. + Architecture Implementation:Enforce the Medallion Architecture patterns, ensuring clear separation of concerns between data layers. + Performance Tuning:Optimize processing workflows (Polars/Spark) jobs and SQL queries to reduce costs and execution time; make intelligent decisions on when to use Polars vs. Spark. + Deployment & Operations:Manage code deployment to on-prem and cloud infrastructure, including containerization and environment configuration. + Data Quality:Implement comprehensive data validation checks and quality gates between medallion layers. + Data Cataloging:Maintain the metadata and catalog entries to ensure all data assets are discoverable and documented. The Technology Stack + Orchestration:Apache Airflow. + Data Processing:Polars (primary for ETL), PySpark/SQL (for massive scale) + Compute: Single-node workers (Polars), Databricks/Spark clustrers (when scale requires) + Storage:Delta Lake, Parquet, S3/Blob Storage, MinIO + Language:Python 3.12+ (w/ Polars), SQL. Qualifications + The Structured Thinker:You love organizing data. You understand the importance of schemas, data typing, and normalization. + Quality Obsessive:You don't just move data; you test it. You implement checks to ensure no bad data reaches the Gold layer. + Pipeline Builder:You view data engineering as software engineering. You write modular, reusable code for your transformations. Knowledge, Skills, and Abilities Must Haves: + 3+ years of experience in Data Engineering. + Strong proficiency inApache AirflowandDatabricks. + Experience implementingMedallion/Delta Lakearchitectures. + StrongSQLandPythonskills. + Advanced English communication skills. Good to Have: + Experience withUnity Catalogor other governance tools. + Familiarity withdbt(data build tool). + Background in processing telemetry or sensor data. Halliburton is an Equal Opportunity Employer. Employment decisions are made without regard to race, color, religion, disability, genetic information, pregnancy, citizenship, marital status, sex/gender, sexual preference/ orientation, gender identity, age, veteran status, national origin, or any other status protected by law or regulation. Location Rua Paulo Emidio Barbosa 485 Q, Rio de Janeiro, Rio de Janeiro, 291941, Brazil Job Details Requisition Number: 205556 Experience Level: Entry-Level Job Family: Engineering/Science/Technology Product Service Line: Landmark Software & Services Full Time / Part Time: Full Time Additional Locations for this position: Compensation Information Compensation is competitive and commensurate with experience.
Confirmar seu email: Enviar Email