Participate in the development and maintenance of data pipelines
Work on data modeling, ETL processes and data transformation tasks to support the data warehousing infrastructure
Contribute to data quality assurance efforts by identifying and rectifying data anomalies and inaccuracies
Maintain clear and concise documentation of data pipelines, processes and best practices
Work across departmental teams to support the data needs of internal partners and stakeholders, clients and regulatory bodies
Stay up-to-date with emerging trends and technologies in Business Intelligence, Data Warehousing, Data Lakes and related fields
Create ad-hoc reports from our data warehouse using SQL/notebooks
Familiarity with Kafka, Flink/Spark, HBase and/or other relevant Big Data technologies
First knowledge in database systems, SQL and data modeling
Proficiency in at least one programming language (e.g., Python, Java, Scala, Spark) is desirable
Experience with Cloudera solutions, both operational and application development, is a plus
You have relevant degrees in a technical field of study (e.g. data or computer science)
You are curious about the world of data engineering and enjoy finding solutions to arising challenges
Excellent communication and teamwork abilities
Detail-oriented with a commitment to delivering high-quality work