About the Role
We’re hiring two Databricks Data Engineers to build core pipelines that power LakeFusion’s AI-native MDM platform. You’ll be hands-on with ingestion, transformation, and optimization of large datasets within the Databricks Lakehouse environment.
Responsibilities:
Build and maintain scalable data pipelines using Databricks (PySpark, Delta Lake, SQL).
Develop ETL workflows for structured and unstructured data ingestion.
Collaborate with QA and AI/ML engineers to ensure clean, reliable datasets for entity resolution.
Implement schema evolution handling and data validation frameworks.
Optimize pipeline performance, cluster configurations, and cost efficiency.
Troubleshoot and resolve data quality and lineage issues across environments.
Requirements
Key Requirements:
4+ years of experience in data engineering, with at least 1+ year in Databricks.
Proficiency in PySpark, SQL, and Delta Lake.
Experience with data lake/lakehouse architectures.
Familiarity with CI/CD pipelines for data workflows.
Strong debugging and performance tuning skills.
Bonus: experience with real-time data ingestion (Kafka, Kinesis, or EventHub).
About the Company
Frisco Analytics is a forward-thinking data consulting firm dedicated to empowering businesses with cutting-edge analytics and insights. We specialize in transforming complex data into actionable strategies that drive growth and innovation. Our expert team leverages advanced technologies and a deep understanding of industry trends to deliver tailored solutions that meet the unique needs of our clients. At Frisco Analytics, we believe in the power of data to unlock potential and create lasting impact, partnering with businesses to navigate the ever-evolving landscape of modern analytics.