Skip to main content

This job has expired

Data Engineer

Employer
Ralph Lauren
Location
Bangalore, India
Closing date
4 Feb 2023

View more

Function
Technology
Level
Manager

Job Details

Company Description
Ralph Lauren Corporation (NYSE:RL) is a global leader in the design, marketing and distribution of premium lifestyle products in five categories: apparel, accessories, home, fragrances, and hospitality. For more than 50 years, Ralph Lauren's reputation and distinctive image have been consistently developed across an expanding number of products, brands and international markets. The Company's brand names, which include Ralph Lauren, Ralph Lauren Collection, Ralph Lauren Purple Label, Polo Ralph Lauren, Double RL, Lauren Ralph Lauren, Polo Ralph Lauren Children, Chaps, among others, constitute one of the world's most widely recognized families of consumer brands. At Ralph Lauren, we unite and inspire the communities within our company as well as those in which we serve by amplifying voices and perspectives to create a culture of belonging, ensuring inclusion, and fairness for all. We foster a culture of inclusion through: Talent, Education & Communication, Employee Groups and Celebration.

Position Overview
Based in Bengaluru, India this Data Engineer will work with the Global Analytics team to build, maintain, and optimize data pipelines for key data and analytics consumers including business and data analysts and data scientists covering our digital and physical channels and value chain. Data engineers also need to guarantee compliance with data governance and data security requirements while creating, improving and operationalizing these integrated and reusable data pipelines. This would enable faster data access, integrated data reuse and vastly improved time-to-solution for Ralph Lauren's data and analytics initiatives. The data engineer will be measured on their ability to integrate analytics and (or) data science results with Ralph Lauren's business processes. This role will require both creative and collaborative working with IT and the wider business. It will involve evangelizing effective data management practices and promoting better understanding of data and analytics. The data engineer will also be tasked with working with key business stakeholders, IT experts and subject-matter experts to plan and deliver optimal enterprise data assets.​

Essential Duties & Responsibilities
Build data pipelines: The primary responsibility of data engineers is to architect, build, and maintain data pipelines that will provision high quality data ready for analysis. This includes ingestion, exploration, modeling, and curation of high value data. Drive Automation through effective metadata management: The data engineer will be responsible for using innovative and modern tools, techniques and architectures to partially or completely automate the most-common, repeatable and tedious data preparation and integration tasks in order to minimize manual and error-prone processes and improve productivity. Learning and using modern data preparation, integration and AI-enabled metadata management tools and techniques. Tracking data consumption patterns. Performing intelligent sampling and caching. Monitoring schema changes. Recommending - or sometimes even automating - existing and future integration flows. Educate and train: The data engineer should be curious and knowledgeable about new data initiatives and how to address them. This includes applying their data and/or domain understanding in addressing new data requirements. They will also be responsible for proposing appropriate (and innovative) data ingestion, preparation, integration and operationalization techniques in optimally addressing these data requirements. The data engineer will be required to train counterparts such as data scientists, data analysts, LOB users or any data consumers in these data pipelining and preparation techniques, which make it easier for them to integrate and consume the data they need for their own use cases. Participate in ensuring compliance and governance during data use: It will be the responsibility of the data engineer to ensure that the data users and consumers use the data provisioned to them responsibly through data governance and compliance initiatives. Data engineers should work with data governance teams (and information stewards within these teams) and participate in vetting and promoting content created in the business and by data scientists to the curated data catalog for governed reuse. Become a data and analytics evangelist: The data engineer will be considered a blend of data and analytics "evangelist," "data guru" and "fixer." This role will promote the available data and analytics capabilities and expertise to business unit leaders and educate them in leveraging these capabilities in achieving their business goals. What you bring (Qualifications): Education and Experience A bachelor's or master's degree in computer science, statistics, applied mathematics, data management, information systems, information science or a related quantitative field is required. The ideal candidate will have a combination of IT skills, data governance skills, analytics skills and Retail industry knowledge with a technical or computer science degree. Multiple years of work experience in data management disciplines including data integration, modeling, optimization and data quality, and/or other areas directly relevant to data engineering responsibilities and tasks. At least 2 years of experience working in cross-functional teams and collaborating with business stakeholders in Retail in support of a departmental and/or multi-departmental data management and analytics initiative. Deep Retail Industry knowledge or previous experience working in the business would be a plus.

Experience, Skills & Knowledge

* Strong experience with advanced analytics tools for Object-oriented/object function scripting using languages such as R, Python, Scala, or similar.
* Strong ability to design, build and manage data pipelines in PySpark and related technologies for data structures encompassing data transformation, data models, schemas, metadata and workload management. The ability to work with both IT and business in integrating analytics and data science output into business processes and workflows.
* Strong experience with popular database programming in relational and nonrelational environments including on AWS Redshift, AWS Aurora, SQL Server and similar platforms.
* Experience in working with large, heterogeneous datasets in building and optimizing data pipelines, pipeline architectures and integrated datasets using traditional data integration technologies. These should include ETL/ELT, data replication/CDC, message-oriented data movement and upcoming data ingestion and integration technologies such as stream data integration and data virtualization.
* Strong experience in working with and optimizing existing ETL processes and data integration and data preparation flows and helping to move them in production.
* Experience in working with both open-source and commercial message queuing technologies such as Kafka, Amazon Simple queuing Service, stream data integration technologies such as Apache Nifi, Apache Kafka Streams, Amazon Kinesis and stream analytics technologies such as Apache Kafka KSQL.
* Basic experience working with popular data discovery, analytics and BI software tools like MicroStrategy, Tableau, Qlik, PowerBI and others for semantic-layer-based data discovery.
* Basic understanding of popular open-source and commercial data science platforms such as Python, R, KNIME, Alteryx, others is a strong plus but not required/compulsory.
* Basic experience in working with data governance, data quality, and data security teams and specifically and privacy and security officers in moving data pipelines into production with appropriate data quality, governance and security standards and certification.
* Demonstrated ability to work across multiple deployment environments including cloud, on-premises and hybrid, multiple operating systems and through containerization techniques such as Docker, Kubernetes, AWS Elastic Container Service and others.
* Experienced in agile methodologies and capable of applying DevOps and increasingly DataOps principles to data pipelines to improve the communication, integration, reuse and automation of data flows between data managers and consumers across an organization Interpersonal Skills
* Strong experience supporting and working with cross-functional teams in a dynamic business environment.
* Required to be highly creative and collaborative. An ideal candidate would be expected to collaborate with both the business and IT teams to define the business problem, refine the requirements, and design and develop data deliverables accordingly. The successful candidate will also be required to have regular discussions with data consumers on optimally refining the data pipelines developed in nonproduction environments and deploying them in production.
* Required to have the accessibility and ability to interface with, and gain the respect of, stakeholders at all levels and roles within the company.
* Is a confident, energetic self-starter, with strong interpersonal skills.
* Has good judgment, a sense of urgency and has demonstrated commitment to high standards of ethics, regulatory compliance, customer service and business integrity. #LI-NP1

Company

Learn more about this company

Visit this company’s hub to learn about their values, culture, and latest jobs.

Visit employer hub

Learn more about this company

Visit this company’s hub to learn about their values, culture, and latest jobs.

Visit employer hub

Get job alerts

Create a job alert and receive personalised job recommendations straight to your inbox.

Create alert