Data Engineer

Overview

Acuity Knowledge Partners, a leading provider of high-value research, analytics, and business intelligence to over 500 financial institutions and consulting companies through our specialist workforce of over 6,000 analysts and delivery experts across our global delivery network.

Job Description

Job Purpose:

Design, construct, and maintain scalable data management systems using Azure Databricks, ensuring they meet end-user expectations. Supervise the upkeep of existing data infrastructure workflows to ensure continuous service delivery. Create data processing pipelines utilizing Databricks Notebooks, Spark SQL, Python and other Databricks tools. Oversee and lead the module through planning, estimation, implementation, monitoring and tracking.

Key Responsibilities:

Interpret business requirements, either gathered or acquired. 

Work with internal resources as well as application vendors 

Designing, developing, and maintaining Databricks Solution and Relevant Data Quality rules

Troubleshooting and resolving data related issues.

Configuring and Creating Data models and Data Quality Rules to meet the needs of the customers.

Hands on in handling multiple database platforms, like Microsoft SQL Server, Oracle etc

Reviewing and analysing data from multiple internal and external sources

Analyse existing PySpark/Python code and identify areas for optimization.

Write new optimized SQL queries or Python Scripts to improve performance and reduce run time.

Identify opportunities for efficiencies and innovative approaches to completing scope of work.

Write clean, efficient, and well-documented code that adheres to best practices and Council IT coding standards.

Maintenance and operation of existing custom codes processes

Participate in team problem solving efforts and offer ideas to solve client issues.

Query writing skills with ability to understand and implement changes to SQL functions and stored procedures.

Effectively communicate with business and technology partners, peers and stakeholders

Ability to deliver results under demanding timelines to real-world business problems.

Ability to work independently and multi-task effectively.

Configure system settings and options and execute unit/integration testing. 

Develop end-user Release Notes, training materials and deliver training to a broad user base.

Identify and communicate areas for improvement Demonstrate high attention to detail, should work in a dynamic environment whilst maintaining high quality standards, a natural aptitude to develop good internal working relationships and a flexible work ethic.

Responsible for Quality Checks and adhering to the agreed Service Level Agreement (SLA) / Turn Around Time (TAT)

Key Competencies:

Over 8 + years of experience in data engineering, with expertise in Azure Databricks, MSSQL, LakeFlow, Python and supporting Azure technology.

Design, build, test, and maintain highly scalable data management systems using Azure Databricks.

Create data processing pipelines utilizing Databricks Notebooks, Spark SQL.

Integrate Azure Databricks with other Azure services like Azure Data Lake Storage, Azure SQL Data Warehouse.

Design and implement robust ETL pipelines using Databricks, ensuring data quality and integrity.

Design and implement effective data models, schemas and data governance using the Databricks environment.

Develop and optimize PySpark/Python code for data processing tasks.

Assist stakeholders with data-related technical issues and support their data infrastructure needs.

Develop and maintain documentation for data pipeline architecture, development processes, and data governance.

Data Warehousing: In-depth knowledge of data warehousing concepts, architecture, and implementation, including experience with various data warehouse platforms.

Data Quality – implement data quality rules using Databricks and external platforms like IDQ.

Extremely strong organizational and analytical skills with strong attention to detail

Strong track record of excellent results delivered to internal and external clients.

Excellent problem-solving skills, with ability to work independently or as part of team.

Strong communication and interpersonal skills, with ability to effectively engage with both technical and non-technical stakeholders.

Able to work independently without the needs for close supervision and collaboratively as part of cross-team efforts.

Skills & Requirements

Azure Databricks, PySpark, SQL (MSSQL, Spark SQL), Azure Data Lake Storage, Data Modeling and Governance, ETL Pipeline Development, Data Warehousing, Python, Data Quality Implementation, Problem-Solving and Analytical Skills.

Apply Now

Join Our Community

Let us know the skills you need and we'll find the best talent for you