Associate Manager - Data Engineer
Hyderābād (Hyderābād) IT development
Job description
Overview
PepsiCo operates in an environment undergoing immense and rapid change. Big-data and digital technologies are driving business transformation that is unlocking new capabilities and business innovations in areas like eCommerce, mobile experiences and IoT. The key to winning in these areas is being able to leverage enterprise data foundations built on PepsiCo’s global business scale to enable business insights, advanced analytics and new product development. PepsiCo’s Data Management and Operations team is tasked with the responsibility of developing quality data collection processes, maintaining the integrity of our data foundations and enabling business leaders and data scientists across the company to have rapid access to the data they need for decision-making and innovation.
What PepsiCo Data Management and Operations does:
· Maintain a predictable, transparent, global operating rhythm that ensures always-on access to high-quality data for stakeholders across the company
· Responsible for day-to-day data collection, transportation, maintenance/curation and access to the PepsiCo corporate data asset
· Work cross-functionally across the enterprise to centralize data and standardize it for use by business, data science or other stakeholders
· Increase awareness about available data and democratize access to it across the company
Job Description
As a Data Architect, you will be the key technical expert overseeing PepsiCo's data product build & operations and drive a strong vision for how data engineering can proactively create a positive impact on the business. As a member of the data engineering team, you will help lead the development of very large and complex data applications into public cloud environments directly impacting the design, architecture, and implementation of PepsiCo's flagship data products around topics like revenue management, supply chain, manufacturing, and logistics. The primary responsibilities of this role are to work with business users, data product owners, platform owners, enterprise architects, data management owners, and data engineering teams to ensure the data supply chain and the enterprise data products are built with high performance, high availability, and maintainability standards using current and emerging big data technologies. You'll be working in a hybrid environment with in-house, on-premise data sources as well as cloud and remote systems. You will establish data design patterns that will drive flexible, scalable, and efficient data models to maximize value and reuse. You’ll make tactical architecture decisions to support immediate projects but will be a key expert informing long term data architecture strategy.
Important Disclaimer :
· The candidate is required to work for 4 weekends in a quarter (you may be required to work only on Saturday or Sunday or Sat & Sun) and basis that you'll get compensatory off.
· Please note that this role will be based ONLY in India. The role does not involve any movement to other PepsiCo offices outside India in future
Responsibilities
.
Qualifications
· 11+ years of overall technology experience that includes at least 5+ years of hands-on software development, data engineering, and systems architecture.
· 5+ years of experience with Data Lake Infrastructure, Data Warehousing, and Data Analytics tools.
· 5+ years of experience developing enterprise data models.
· Experience in at least one of data modeling tool (ER/Studio, Erwin) for three years.
· 5+ years in cloud data engineering experience in at least one cloud (Azure, AWS, GCP).
· 5+ years’ experience with python & pyspark
· 5+ years of experience with SQL
· Experience with data modeling, data warehousing, and building high-volume ETL/ELT pipelines.
· 5+ years of experience in developing programs for high volume ETL/ELT pipelines using Spark, Hadoop.
· 3+ years of hands-on experience on-premises to cloud migrations, adept in planning, execution, and optimization of end-to-end data migration projects
· Experience with integration of multi cloud services with on-premises technologies.
· Experience with building solutions in the retail or in the supply chain space is preferred.
· Experience with data profiling and data quality tools like Apache Griffin, Deequ, and Great Expectations.
· Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets.
· Experience with at least one MPP technology such as Redshift, Synapse, BigQuery or SnowFlake.
· Experience with running and scaling applications on the cloud infrastructure and containerized services like Docker and Kubernetes.
· Experience with version control systems like Azure DevOps, and deployment & CI tools.
· Experience with Azure Data Factory, Databricks and Azure Machine learning is a plus.
· Experience in building API services with high volume is preferred.
· Certified in Data engineering/Analytics in one or more clouds - Azure
· Experience in infrastructure automation using Infrastructure as a Service (IAAS) tools such as Terraform.
· Experience with building Data Observability systems
· Proficient in designing, implementing, and managing infrastructure code for efficient and scalable deployment of resources.
· Experience with metadata management, data lineage, and data glossaries.
· Working knowledge of agile development, including DevOps and DataOps concepts.
· Familiarity with business intelligence tools (such as PowerBI/Tableau).
· B. Tech /MCA in Computer Science a must
Skills, Abilities, Knowledge
· Excellent communication skills, both verbal and written, along with the ability to influence and demonstrate confidence in communications with senior level management.
· Proven track record of leading, mentoring, hiring and scaling data teams.
· Strong change manager. Comfortable with change, especially that which arises through company growth.
· Ability to understand and translate business requirements into data and technical requirements.
· High degree of organization and ability to manage multiple, competing projects and priorities simultaneously.
· Positive and flexible attitude to enable adjusting to different needs in an ever-changing environment.
· Strong leadership, organizational and interpersonal skills; comfortable managing trade-offs.
· Foster a team culture of accountability, communication, and self-management.
· Proactively drives impact and engagement while bringing others along.
· Consistently attain/exceed individual and team goals
· Ability to lead others without direct authority in a matrixed environment.
Education
· BE/BTech in Computer Science a must, Master’s in computer science and Data engineering and Data science related fields preferred
Competencies
· Highly influential and having the ability to educate challenging stakeholders on the role of data and its purpose in the business.
· Understands both the engineering and business side of the Data Products released.
· Places the user in the center of decision making.
· Teams up and collaborates for speed, agility, and innovation.
· Experience with and embraces agile methodologies.
· Strong negotiation and decision-making skill.