Job Description
Summary
We’re hiring a Senior Data Engineer with strong experience in AWS and Databricks to build scalable data solutions that power next-gen AI and machine learning. Join our fast-growing team to work on impactful projects, collaborate with top talent, and drive innovation at scale.
Key Responsibilities:
- Design, build, and manage large-scale data infrastructures using a variety of AWS technologies such as Amazon Redshift, AWS Glue, Amazon Athena, AWS Data Pipeline, Amazon Kinesis, Amazon EMR, and Amazon RDS.
- Design, develop, and maintain scalable data pipelines and architectures on Databricks using tools such as Delta Lake, Unity Catalog, and Apache Spark (Python or Scala), or similar technologies.
- Integrate Databricks with cloud platforms like AWS to ensure smooth and secure data flow across systems.
- Build and automate CI/CD pipelines for deploying, testing, and monitoring Databricks workflows and data jobs.
- Continuously optimize data workflows for performance, reliability, and security, applying Databricks best practices around data governance and quality.
- Ensure the performance, availability, and security of datasets across the organization, utilizing AWS’s robust suite of tools for data management.
- Collaborate with data scientists, software engineers, product managers, and other key stakeholders to develop data-driven solutions and models.
- Translate complex functional and technical requirements into detailed design proposals and implement them.
- Mentor junior and mid-level data engineers, fostering a culture of continuous learning and improvement within the team.
- Identify, troubleshoot, and resolve complex data-related issues.
- Champion best practices in data management, ensuring the cleanliness, integrity, and accessibility of our data.
- Optimize and fine-tune data queries and processes for performance. Evaluate and advise on technological components, such as software, hardware, and networking capabilities, for database management systems and infrastructure.
- Stay informed on the latest industry trends and technologies to ensure our data infrastructure is modern and robust.
Qualifications:
- 5-7 years of hands-on experience with AWS data engineering technologies, such as Amazon Redshift, AWS Glue, AWS Data Pipeline, Amazon Kinesis, Amazon RDS, and Apache Airflow.
- Hands-on experience working with Databricks, including Delta Lake, Apache Spark (Python or Scala), and Unity Catalog.
- Demonstrated proficiency in SQL and NoSQL databases, ETL tools, and data pipeline workflows.
- Experience with Python, and/or Java.
- Deep understanding of data structures, data modeling, and software architecture.
- Experience with AI and machine learning technologies is highly desirable.
- Strong problem-solving skills and attention to detail.
- Self-motivated and able to work independently, with excellent organizational and multitasking skills.
- Exceptional communication skills, with the ability to explain complex data concepts to non-technical stakeholders.
- Bachelor's Degree in Computer Science, Information Systems, or a related field. A Master's Degree is preferred.
Skills
- Attention to Detail
- AWS
- Communications Skills
- Database Management
- Java
- Problem Solving
- SQL
- Team Collaboration