At DXC we use the power of technology to deliver mission-critical IT services that our customers need to modernize operations and drive innovation across their entire IT estate. We provide services across the Enterprise Technology Stack for business process outsourcing, analytics and engineering, applications, security, cloud, IT outsourcing, and modern workplace.
Our DXC Application services team works with our customers as a trusted partner to help simplify, modernize, and accelerate mission-critical applications that support business agility and growth. Customers work with us to take advantage of the latest digital platforms with both custom and packaged applications, ensure resiliency, launch new products and enter new markets with minimal disruption.
About this role:
DXC team is seeking a highly skilled and analytical Data Engineer to join our team. As a Lead Data Engineer, you will be responsible for leading the design, development, and maintenance of our data pipelines and infrastructure. You will work closely with our data scientists, analysts, and IT team to ensure that our data infrastructure is optimized for fast and accurate data processing, while also adhering to company policies and standards. You will also play a key role in developing our data strategy and roadmap, and in mentoring and training junior data engineers on the team.
What You’ll Do:
In this role your key responsibilities will be:
Primary responsibility will be to develop Big Data Hadoop code for new or existing data products, test data products, complete production deployment activities, support production runs, and fix production issues.
Creation of data products for all finance consumers – business users, analysts, and modelers. Explore and understand data sets.
Visualize the data set; determine whether the data set has enough information to answer the question that the business is asking.
Work with Data Analysts to understand the requirements for building new data products or enhancing existing data products.
Building data pipelines, integrating and scheduling Hadoop ETL jobs using CI/CD framework.
Implement required data transformation in the data lake.
Work with IT support to create ETL / ELT interfaces to the data lake and create and visualize the data and data products on the data lake.
Configure required security and data masking to a data set. Implement the Data and IT controls.
Support testing of data acquisition, data set correlation, and/or product development. Investigate and resolve interface issues
Represent Data product changes in CAB meetings and provide updates on development work in Sprint meetings
Actively support Data products production runs, validate the data products for completeness & accuracy and coordinate with Production support teams
Work with IT to harden and produce the model, model interfaces, data products, and business procedures.
Who You Are:
In this role your key responsibilities will be:
Must have domain technical experience in delivering data engineering solutions and developing data products using data lake technology
Experience with the following: Hadoop (CDH), relational databases and SQL, ETL development, spark, data validation and testing (Data Warehousing, ETL/ELT to the Data Lake, Using the Data Lake for data analysis (Hadoop tools – Hive, Impala, Pig, Sqoop, Hue, Kafka, etc., Spark, Python, R, java, Docker, Dakota).
Working experience in Shell Scripts, Oozie workflows, and scheduling tools (Stone branch or CA7).
Knowledge of Cloud platform implementation (Azure or Amazon).
Knowledge of Tableau data visualization tools is a plus
Excellent written, verbal, and interpersonal skills, are a must as there will be significant collaboration with the business and IT
Experience with collaborative development workflows (e.g., Microsoft DevOps Tools).
In addition to the basic qualifications noted above, we find that individuals who are successful in this role have:
Master’s/ Bachelor’s Degree in one of the following: Engineering, Computer Science, Statistical Analytics, Data Science, or Actuarial Science.
At least 8 years of relevant work experience in implementing data and analytics projects. Hadoop ecosystem experience of at least 4+ years is required
Joining DXC connects you to brilliant people who embrace change and seize opportunities to advance their careers and amplify customer success. At DXC we support each other and work as a team — globally and locally. Our achievements demonstrate how we deliver excellence for our customers and colleagues. You will be joining a team that works to create a culture of learning, diversity, and inclusion and are dedicated to strong ethics and corporate citizenship.
At DXC we put our people first. In managing COVID-19, our actions are focused on the health, safety, and well-being of our colleagues and their families and our approach is to encourage and support masking, testing, and vaccination. With our Virtual First strategy, the majority of our workforce now works remotely and will continue to do so. We recognize that requirements and availability around masking, testing, and vaccination vary by location, and we continue to monitor and conform with government regulations and customer requirements specific to each location.