Excelra hiring Data Engineer : Hiring @ Excelra
Remote Role - www.excelra.com
- strong on Python - Ready to take assessment on Python
Job Summary
At Excelra ,We are seeking a highly skilled Data Engineer with expertise in Databricks, SQL, Python, and data flow pipelines. The ideal candidate will have experience in creating Azure Data Factory (ADF) pipelines, deploying code in Azure DevOps using Git Repos and YAML files, and exposure to Snowflake. This role involves designing, implementing, and maintaining robust data solutions to support various business needs.
Key Responsibilities
Databricks Expertise:
Develop and maintain data processing workflows using Databricks.
Optimize Databricks jobs for performance and cost efficiency.
Implement data transformations and aggregations within the Databricks environment.
Collaborate with data scientists and analysts to integrate and process large datasets.
SQL Proficiency
Write complex SQL queries to retrieve and manipulate data from various databases.
Ensure data accuracy and consistency through effective querying and data validation techniques.
Optimize SQL queries for performance and scalability.
Python Programming
Develop data pipelines and ETL processes using Python.
Create reusable code modules and libraries for data processing tasks.
Perform data cleaning, transformation, and analysis using Python-based tools and frameworks.
Data Flow Pipelines
Design and implement data flow pipelines to move and transform data across different systems.
Ensure data integrity and reliability in data pipelines.
Monitor and troubleshoot data pipelines to resolve issues promptly.
Azure Data Factory (ADF) Pipelines
Create and manage ADF pipelines to orchestrate data movement and transformation activities.
Develop custom ADF activities using Azure Functions or other services.
Implement logging, monitoring, and alerting mechanisms for ADF pipelines.
Azure DevOps Deployment Pipelines
Set up and maintain deployment pipelines in Azure DevOps.
Use Git Repos for version control of data flow related code and Databricks notebooks.
Create and manage YAML files for continuous integration and deployment (CI/CD) processes.
Ensure seamless deployment of Databricks and other data-related code through Azure DevOps.
Snowflake Exposure
Work with Snowflake data warehouse to store and manage large volumes of data.
Write and optimize SQL queries for Snowflake.
Integrate Snowflake with other data processing tools and platforms.
Qualifications
Education:
Bachelor’s degree in Computer Science, Information Technology, Engineering, or a related field. A master’s degree is a plus.
Experience
Proven experience as a Data Engineer or similar role.
Hands-on experience with Databricks and Spark.
Strong SQL Skills With Experience In Query Optimization.
Proficiency in Python for data processing and automation.
Experience in building and managing ADF pipelines.
Familiarity with Azure DevOps, Git, and YAML for CI/CD processes.
Exposure to Snowflake and understanding of its architecture and capabilities.
Skills
Strong analytical and problem-solving skills.
Excellent communication and collaboration abilities.
Ability to work in a fast-paced, dynamic environment.
Attention to detail and commitment to data quality.
Additional Desirable Skills
Experience with other cloud platforms (AWS, GCP).
Knowledge of data governance and security best practices.
Familiarity with data visualization tools (e.g., Power BI, Tableau).
Benefits
Competitive salary and benefits package.
Opportunities for professional growth and development.
Collaborative and innovative work environment.
Access to the latest technologies and tools.