JOB SUMMARY
We are looking for a skilled and experienced Data Engineerto join our growing team. The ideal candidate will be responsible for designing, building, and maintaining scalable and efficient data pipelines, databases, and data infrastructure. You will work closely with our data scientists, analysts, and software engineers to support data-driven decision-making and deliver high-quality data solutions.
JOB DUTIES
- Design, build, and maintain robust and scalable data pipelines to collect, process, and store structured and unstructured data from various sources.
- Develop and optimize database schemas, tables, and queries to support data storage, retrieval, and analysis.
- Collaborate with data scientists and product managers to understand data requirements, perform data profiling, and implement data transformation and cleaning processes.
- Implement ETL processes to integrate data from various sources, ensuring data quality, consistency, and integrity.
- Optimize data pipelines and queries for performance, scalability, and cost-efficiency.
- Integrate analytical tools such as Tableau, Power BI, or Looker to enable data analysis, visualization, and reporting capabilities.
- Monitor and troubleshoot data pipelines and infrastructure issues, and provide timely resolutions.
- Work closely with cross-functional teams to identify data needs and requirements, and deliver data solutions that meet business objectives.
- Ensure data security and compliance with data governance and privacy regulations.
- Document data pipelines, processes, and best practices, and contribute to the development of data engineering standards and guidelines.
- Stay up-to-date with the latest data engineering technologies, tools, and best practices, and recommend their adoption as appropriate.
JOB REQUIREMENTS
- Bachelor’s degree in Computer Science, Information Systems, or equivalent work experience.
- Proven experience as a Data Engineer or similar role, with a strong background in designing and implementing data pipelines and databases.
- Proficiency in programming languages commonly used in data engineering such as Python, Java.
- Hands-on experience with big data technologies and frameworks such as Hadoop, Spark, Kafka, and Hive.
- Hands-on experience with analytical tools such as Tableau, Power BI, Looker, or similar platforms.
- Solid understanding of relational and NoSQL databases, and experience with database design, optimization, and SQL querying.
- Experience with cloud platforms (e.g., AWS, Azure, GCP) and data warehousing solutions (e.g., Snowflake, Redshift, BigQuery).
- Experience with real-time data processing and streaming technologies (e.g., Apache Flink, Kafka Streams).
- Strong analytical and reasoning skills with an ability to anticipate outcomes of a solution.
- Strong communication and collaboration skills, with the ability to work effectively in a cross-functional team environment.
- Strong sense of responsibility, initiative, and commitment.
- Self-motivated, proactive, and able to work independently with minimal supervision.
- Knowledge of data governance, data security, and compliance standards and best practices.
- Maintain confidentiality of information processed or prepared.
- Ability to lead, convince and inspire team members in a fast-paced environment.
- Willingness to travel to meet and work with offshore team members.
- Fluency in written and spoken English and Mandarin is highly desirable to liaise with Chinese speaking clients and counterparts.