工作描述
52 天前
Why Work at Lenovo
We are Lenovo. We do what we say. We own what we do. We WOW our customers.
Lenovo is a US$57 billion revenue global technology powerhouse, ranked #248 in the Fortune Global 500, and serving millions of customers every day in 180 markets. Focused on a bold vision to deliver Smarter Technology for All, Lenovo has built on its success as the world's largest PC company with a full-stack portfolio of AI-enabled, AI-ready, and AI-optimized devices (PCs, workstations, smartphones, tablets), infrastructure (server, storage, edge, high performance computing and software defined infrastructure), software, solutions, and services. Lenovo's continued investment in world-changing innovation is building a more equitable, trustworthy, and smarter future for everyone, everywhere. Lenovo is listed on the Hong Kong stock exchange under Lenovo Group Limited (HKSE: 992) (ADR: LNVGY).
This transformation together with Lenovo's world-changing innovation is building a more inclusive, trustworthy, and smarter future for everyone, everywhere. To find out more visit www.lenovo.com, and read about the latest news via our StoryHub.
Description and Requirements
Summary
Data Asset team is looking for a data engineer with a solid technical background to build and support cloud based data asset platform. Our platform enables internal engineers and analysts to process SQL or transformations using Apache Spark, Python on compute for batch, streaming and ML processing, executing over schemaed data stored in data lake and data warehouse. Our team manages data ingestions from various data sources, build and support ETL pipelines, and continuously look for solutions to enhance the robustness and scalability of the data architecture.
The Job
You will be responsible for application development and ongoing maintenance support of the Data Analytics area with the following scope of duties:
• Manage cloud resources including EC2, EMR, S3, Redshift, Glue, etc.
• Design data schema and operate internal data warehouses and SQL/NOSQL database systems
• Interface with other technology teams to extract, transform, and load data from a wide variety of data sources using SQL and cloud big data technologies
• Build robust and scalable data integration (ETL) pipelines using SQL, PySpark and cloud data services like EMR, Redshift, etc.
• Implement suitable data model for data warehouse in order to achieve efficient data retrieval
• Gather business and functional requirements and translate these requirements into robust, scalable, operable solutions with a flexible and adaptable data architecture.
• Collaborate with other Engineers and Product Managers to recognize and help adopt best practices in data engineering and analytics: data integrity, test design, analysis, validation, and documentation
The Person
• 5+ years working experience in software development, data engineering, business intelligence or related fields
• 2+ years data modeling, ETL development, and data warehousing experience
• 2+ years of experience using big data technologies (Hadoop, Spark, Hive, Impala, etc.)
• Good understanding of cloud infrastructure
• Advanced SQL and query performance tuning skills
• Experience with one of the functional scripting languages (Python, Scala, Java etc.) to process semi-structured or unstructured data inputs.
• Knowledge of professional software engineering practices & best practices for the full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations.
• Strong customer focus, ownership, urgency, and drive.
• University graduate in Computer Science, Information Technology, Data Science or related disciplines
• Effective verbal and written communication in English and Chinese
• Good team player with strong interpersonal and communication skills
• Forward thinking, embracing changes, fast-learning, energetic, able to perform in challenging environment
• Peferred Qualifications
• Working experience with AWS or GCP big data technologies (S3, Redshift, EMR, Kinesis, DynomoDB, BigQuery, Data Flow)
• Experience with HBase, Kafka, Flink.
• Database design and administration experience with Oracle, Teradata or other large commercial relational database systems
• Working experience with NOSQL databases
• Experience contributing to the system design or architecture (architecture, design patterns, reliability and scaling) of new and current systems
• Proven success in communicating with users, other technical teams, and senior management to collect requirements, describe data modeling decisions and data engineering strategy.
• Experience providing technical leadership and educating other engineers for best practices on data engineering.
• Candidates with less experience will also be considered
#LPS
Additional Locations:
• Hong Kong
We are Lenovo. We do what we say. We own what we do. We WOW our customers.
Lenovo is a US$57 billion revenue global technology powerhouse, ranked #248 in the Fortune Global 500, and serving millions of customers every day in 180 markets. Focused on a bold vision to deliver Smarter Technology for All, Lenovo has built on its success as the world's largest PC company with a full-stack portfolio of AI-enabled, AI-ready, and AI-optimized devices (PCs, workstations, smartphones, tablets), infrastructure (server, storage, edge, high performance computing and software defined infrastructure), software, solutions, and services. Lenovo's continued investment in world-changing innovation is building a more equitable, trustworthy, and smarter future for everyone, everywhere. Lenovo is listed on the Hong Kong stock exchange under Lenovo Group Limited (HKSE: 992) (ADR: LNVGY).
This transformation together with Lenovo's world-changing innovation is building a more inclusive, trustworthy, and smarter future for everyone, everywhere. To find out more visit www.lenovo.com, and read about the latest news via our StoryHub.
Description and Requirements
Summary
Data Asset team is looking for a data engineer with a solid technical background to build and support cloud based data asset platform. Our platform enables internal engineers and analysts to process SQL or transformations using Apache Spark, Python on compute for batch, streaming and ML processing, executing over schemaed data stored in data lake and data warehouse. Our team manages data ingestions from various data sources, build and support ETL pipelines, and continuously look for solutions to enhance the robustness and scalability of the data architecture.
The Job
You will be responsible for application development and ongoing maintenance support of the Data Analytics area with the following scope of duties:
• Manage cloud resources including EC2, EMR, S3, Redshift, Glue, etc.
• Design data schema and operate internal data warehouses and SQL/NOSQL database systems
• Interface with other technology teams to extract, transform, and load data from a wide variety of data sources using SQL and cloud big data technologies
• Build robust and scalable data integration (ETL) pipelines using SQL, PySpark and cloud data services like EMR, Redshift, etc.
• Implement suitable data model for data warehouse in order to achieve efficient data retrieval
• Gather business and functional requirements and translate these requirements into robust, scalable, operable solutions with a flexible and adaptable data architecture.
• Collaborate with other Engineers and Product Managers to recognize and help adopt best practices in data engineering and analytics: data integrity, test design, analysis, validation, and documentation
The Person
• 5+ years working experience in software development, data engineering, business intelligence or related fields
• 2+ years data modeling, ETL development, and data warehousing experience
• 2+ years of experience using big data technologies (Hadoop, Spark, Hive, Impala, etc.)
• Good understanding of cloud infrastructure
• Advanced SQL and query performance tuning skills
• Experience with one of the functional scripting languages (Python, Scala, Java etc.) to process semi-structured or unstructured data inputs.
• Knowledge of professional software engineering practices & best practices for the full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations.
• Strong customer focus, ownership, urgency, and drive.
• University graduate in Computer Science, Information Technology, Data Science or related disciplines
• Effective verbal and written communication in English and Chinese
• Good team player with strong interpersonal and communication skills
• Forward thinking, embracing changes, fast-learning, energetic, able to perform in challenging environment
• Peferred Qualifications
• Working experience with AWS or GCP big data technologies (S3, Redshift, EMR, Kinesis, DynomoDB, BigQuery, Data Flow)
• Experience with HBase, Kafka, Flink.
• Database design and administration experience with Oracle, Teradata or other large commercial relational database systems
• Working experience with NOSQL databases
• Experience contributing to the system design or architecture (architecture, design patterns, reliability and scaling) of new and current systems
• Proven success in communicating with users, other technical teams, and senior management to collect requirements, describe data modeling decisions and data engineering strategy.
• Experience providing technical leadership and educating other engineers for best practices on data engineering.
• Candidates with less experience will also be considered
#LPS
Additional Locations:
• Hong Kong
更多來自 Lenovo
產品/項目經理
中西區, 香港
7 天前
全職
辦公室工作
科技、資訊和媒體
數據科學
中西區, 香港
7 天前
全職
辦公室工作
科技、資訊和媒體
IT支援
中西區, 香港
7 天前
全職
辦公室工作
科技、資訊和媒體
產品/項目經理
中西區, 香港
7 天前
全職
辦公室工作
科技、資訊和媒體
網絡安全
中西區, 香港
7 天前
全職
辦公室工作
科技、資訊和媒體
更多相似工作
Data Engineer - AI & Data - Hong Kong Data Engineer - AI & Data - Hong Kong
Walt Disney Imagineering
中西區, 香港
Consultant/ Senior Consultant (Data Engineer) - AI & Data - Hong Kong(313067)
Deloitte Touche Tohmatsu
中西區, 香港
🎉 Got an interview?