123DataEngineer
Position:
Worldwide, Remote Full-time

Senior Data Engineer

Company Headquarter:
Canada
Python
Delta Lake
SQL
Databricks
Synapse
ADF

About the role

Senior Data Engineer (Data Lake + Python + SQL).

We are looking for a Data Engineering Specialist who is passionate about building systems that shape business outcomes. This role will play a key role in designing, building, and optimizing the data platform. You will drive and manage large-scale data migration and modernization initiatives, and work with cross-functional teams to deliver trusted, high-quality data that drives business intelligence, analytics, and innovation.
As a Data Engineering Specialist, you will be at the forefront of building scalable pipelines, architecting data solutions, and ensuring the data platforms can support real-time insights and long-term growth.

Responsibilities and Duties:

• Take ownership of coding and solution design while collaborating with internal and external engineers on design, development, and deployment of data pipelines and solutions.
• Design and implement end-to-end data migration strategies from legacy systems to modern cloud platforms.
• Manage data architecture and data modernization initiatives, ensuring alignment with business goals, data governance policies, and performance requirements.
• Develop and optimize ETL/ELT pipelines using modern tools and frameworks (e.g., Apache Spark, Databricks, Airflow, dbt).
• Work with stakeholders to gather requirements and translate them into scalable and maintainable data solutions.
• Ensure data quality, reliability, and integrity through robust testing and monitoring.
• Drive best practices in DevOps for data – CI/CD.
• Mentor and provide technical guidance to data engineers.

Qualifications Required

:
• 7+ years of relevant experience in Data Engineering / Analytics domain with at least 3 years in Digital Analytics
• 7+ years of experience in SQL
• 4+ years in Python
• Experience with building a data lake over Cloud / Azure, ADF, Synapse, and Databricks – Delta Lake
• Proven expertise with tools like ADF, Qlik Replicate, FiveTran for data integration and ETL processes.
• Strong understanding of Python scripting and use of libraries (NumPy, Pandas)
• Extensive hands-on experience with MS Fabric, including data warehousing, SQL optimization, and cluster management.
• Strong knowledge of Big Data technologies such as Data Bricks, Event driven data processing using Function Apps/Lambda is preferred
• Understanding of Web services (SOAP, XML, UDDI, WSDL)
• Strong knowledge and experience in event driven architecture using standard message queues (i.e.RabbitMQ, SQS or Kafka)
• Hands on Experience parsing NOSQL file systems such as JSON, XML, AVRO, Parquet.
• Proven track record of using Apache Spark Data Bricks.
• Hands-on experience with Azure Data Factory (ADF), Logic Apps, and Runbooks.
• Exposure to traditional BI Tools (Tableau, Power BI, Qlik, SSRS, etc.)
• Experience building web-services is a plus.

Full-time remote job

Apply for this job

Want to receive updates on new job offerings?

If this job doesn't fit you, please apply as talent to receive notifications about new jobs, relevant to you
Subscribe

Check other vacancies