Data Engineer (Databricks) AI/Data Consulting Firm
We are an ambitious consulting firm focused on delivering cutting-edge solutions in data and AI. Our mission is to empower organisations to unlock the full potential of their data by leveraging platforms like Databricks alongside other emerging technologies. As a Data Engineer, you will play a crucial role in building and optimising data solutions, ensuring scalability, performance, and reliability for our clients' complex data challenges.
As a Data Engineer (Databricks), you will be responsible for designing, implementing, and optimising large-scale data processing systems. You will work closely with clients, data scientists, and solution architects to ensure efficient data pipelines, reliable infrastructure, and scalable analytics capabilities. Develop, implement, and optimise data pipelines and ETL processes on Databricks.
Design and implement scalable, high-performance data architectures.
Ensure data integrity, quality, and security through robust engineering practices.
Monitor, troubleshoot, and optimise data workflows for efficiency and cost-effectiveness.
Collaborate with data scientists and analysts to facilitate machine learning and analytical solutions.
Contribute to best practices, coding standards, and documentation to improve data engineering processes.
Design, build, and maintain scalable data pipelines using Databricks, Spark, and Delta Lake.
Develop efficient ETL/ELT workflows to process large volumes of structured and unstructured data.
Implement data governance, security, and compliance standards.
Work with cloud platforms such as AWS, Azure, or GCP to manage data storage and processing.
Collaborate with cross-functional teams to enhance data accessibility and usability.
Optimise data warehouse and lakehouse architectures for performance and cost efficiency.
Maintain and improve CI/CD processes for data pipeline deployment and monitoring.
5+ years of experience in data engineering or related roles.
~ Strong expertise in Databricks, Spark, Delta Lake, and cloud data platforms (AWS, Azure, or GCP).
~ Proficiency in Python and SQL for data manipulation and transformation.
~ Knowledge of data modelling, data warehousing, and lakehouse architectures.
~ Familiarity with DevOps practices, CI/CD pipelines, and infrastructure-as-code.
~ Experience with machine learning data pipelines and MLOps practices.
Knowledge of data streaming technologies such as Kafka or Kinesis.
Competitive compensation, including performance-based incentives.
Opportunities for professional growth and development in a fast-growing firm.
If youre passionate about data engineering and ready to make an impact in AI-driven consulting, wed love to hear from you!