Developing and maintaining ETL pipelines from various data sources.
Developing data cleansing mechanism.
Build efficient & high performing data pipelines that could handle huge amount of traffics/data.
Work closely with other engineers and leads for ensuring the success of data related projects implementation.
Testing developed features or functions to validate that they are working as expected.
Accountable for all software development phases (SDLC) to ensure on-time completion.
Create a conceptual model and system specification.
Coordination with user internal and external.
Analyze technical specifications from client’s needs.
Create update report project.
What you will need:
Have been involved in an IT project with roles as Data Engineer for at least 2 years.
Experienced in managing and designing data pipelines, debugging data issues on various tools or cloud vendor e.g. GCP/Azure/AWS.
Background with distributed framework (e.g. Spark, Hadoop, Kafka Streams).
Workflow management or data pipelines (e.g. Airflow, Azure Data Factory, SSIS).
Experience with relational databases (e.g. SQL Server, MySQL), exposure to no-SQL appliances and a variety of databases (e.g. Cosmo DB, Elastic Search, Kafka, Hive, etc.).
Experience in writing software for data extraction and transformation in one or more languages (e.g. Python, Scala, or Java).
Good understanding of Windows/UNIX/LINUX environments.
Highly skilled in writing and optimizing SQL queries.
Experienced in building data model for Data Warehouse, Data Lake.
Familiar with SDLC, Object-Oriented Design Patterns.
Highly motivated with spirit to learn new things.
Good attitude, commitment to work and good team player.
Good problem solving, analytical and leadership skill.
Ability to work independently or in a group.
Able to manage task and deadline.
Ability to adapt and have flexibility within a high-growth organization.
Must be able to clearly communicate to product development stakeholders: Other Engineers, Testers, Project Managers, Team Lead, etc.
Anticipate issues and act proactively to address potential issues serve as a subject matter expert.
Additional Qualities:
Experience in ETL (extract, transform, load) data pipeline (e.g Azure Data Factory; SSIS) and,
Experience in Business Intelligence Platforms (e.g Microsoft Power BI, Tableu, QlikView).