Job Summary
We are seeking a dynamic and innovative Data Engineer to join our team and drive the development of robust data solutions. In this role, you will be responsible for designing, building, and maintaining scalable data pipelines and architectures that empower data-driven decision-making across the organization. Your expertise will help transform complex data sets into actionable insights, supporting strategic initiatives and operational excellence. This position offers an exciting opportunity to work with cutting-edge technologies and collaborate with cross-functional teams to shape our data landscape.
Responsibilities
- Develop, implement, and optimize ETL (Extract, Transform, Load) processes to ensure efficient data flow from diverse sources such as AWS, Azure Data Lake, Hadoop, and Oracle databases.
- Design and maintain scalable data architectures including data warehouses, lakes, and linked data repositories using tools like Apache Hive, Spark, and Microsoft SQL Server.
- Collaborate with data analysts and scientists to facilitate model training, analytics, and reporting using Looker, Talend, and other BI tools.
- Write clean, efficient code in Python, Java, Bash (Unix shell), Shell Scripting, or VBA to automate workflows and improve system performance.
- Integrate various data sources through RESTful APIs and ensure seamless connectivity across platforms.
- Support database design efforts by creating optimized schemas that support high-performance queries and analytics.
- Participate in Agile development cycles to deliver iterative improvements on data solutions while adhering to best practices in version control and documentation.
Experience
- Proven experience as a Data Engineer or similar role with extensive knowledge of Big Data technologies such as Hadoop, Spark, and Apache Hive.
- Strong proficiency in SQL programming with experience in Microsoft SQL Server, Oracle, or similar relational databases.
- Hands-on experience with cloud platforms like AWS or Azure Data Lake for scalable storage and processing solutions.
- Familiarity with ETL tools such as Informatica or Talend for data integration tasks.
- Demonstrated ability to design and implement complex database schemas and optimize queries for performance analysis skills.
- Knowledge of model training processes and analytics techniques to support data science initiatives.
- Experience working within Agile teams using Scrum or Kanban methodologies is preferred. Join us to leverage your technical expertise in a fast-paced environment where innovation meets impact!
Pay: $117,292.65 - $141,255.66 per year
Benefits:
- 401(k)
- Dental insurance
- Flexible schedule
- Tuition reimbursement
Work Location: Remote