The successful candidate will be part of the Enterprise Data Warehouse team. The engineer will have professional experience in data integration, transformation, developing data pipelines. This individual will need experience working to get API data from different sources/vendors, along with internal data, enriching and anonymizing the data -- then prepping/transforming the data into a readable format (gold) for visualizations, and maintaining current data feeds using streaming practices. Strong experience with Databricks, as well as spark/python or scala, as well as SQL. They need to have a good understanding of streaming and spark. Architecture concepts, Following medallion architecture (bronze/silver/gold) -- they will be operating in AWS (S3) and potentially Azure (blob), once again - heavy use of databricks -- sole contributor and self-starter/confident in their abilities, and strong communication skills. Candidates must be local to DFW. 3 Days onsite in Dallas HQ preferred 2 days remote.
Bachelor's Degree Computer Science, Information Management, or related field; Masters preferred.
7+ years relevant experience or 4+ years with Masters
Databricks experience building data pipelines and workflows is a must
Proven record of stand alone projects
Proficient in programming language (SQL and python or scala -- pyspark, pysql)
Streaming Experience
Understanding of cloud infrastructure and technologies specifically AWS
Benefit packages for this role will start on the 1st day of employment and include medical, dental, and vision insurance, as well as HSA, FSA, and DCFSA account options, and 401k retirement account access with employer matching. Employees in this role are also entitled to paid sick leave and/or other paid time off as provided by applicable law.