Job Description: POC: Sam Chavez
ATTENTION ALL SUPPLIERS!!!
READ BEFORE SUBMITTING
• UPDATED CONTACT NUMBER and EMAIL ID is a MANDATORY REQUEST from our client for all the submissions
• Limited to 1 submission per supplier. Please submit your best.
• We prioritize endorsing those with complete and accurate information
• Avoid submitting duplicate profiles. We will Reject/Disqualify immediately.
• Make sure that candidate’s interview schedules are updated. Please inform the candidate to keep their lines open.
• Please submit profiles within the max proposed rate***
MANDATORY: Please include in the resume the candidate’s complete & updated contact information (Phone number, Email address and Skype ID) as well as a set of 5 interview timeslots over a 72-hour period after submitting the profile when the hiring managers could potentially reach to them. PROFILES WITHOUT THE REQUIRED DETAILS and TIME SLOTS will be REJECTED.
Job Title: Technology Analyst | Big Data – Data Processing | Spark — Spark Data Engineer
Work Location & Reporting Address: Plano, TX 75024 (Onsite-Hybrid. Open to relocating candidates)
Contract duration: 12
MAX VENDOR RATE: *** per hour max (NOT NEGOTIABLE)
Target Start Date: 26 Feb 2025
Does this position require Visa independent candidates only? No
Must Have Skills:
• Spark
• Databricks
• AWS
Nice to Have Skills:
Detailed Job Description:
• Develop and maintain data pipelines using Apache Spark and Scala to process large volumes of data efficiently.
• Implement data transformation logic to cleanse, normalise, and aggregate data from various sources, ensuring high data quality and consistency.
• Integrate data from different systems and sources, including structured and unstructured data, databases, and Mainframe systems, ensuring seamless data flow and compatibility.
• Implement data validation and quality checks within data pipelines to ensure data accuracy, consistency, and integrity, and handle exception handling and error logging.
• Develop and execute unit tests, integration tests, and end-to-end tests to ensure the reliability and correctness of Spark applications.
• Debug and troubleshoot issues during development and production phases.
• Contribute to the overall DevOps for SBT builds, application deployment stages and releases.
• Designing, Implementing, and maintain Continuous Integration and Continuous Deployment pipelines.
• Automate/Optimise the build, test, and deployment processes to enable rapid and reliable software delivery.
• Collaborate with Walmart cross-functional teams, including Product managers, Architects, and other developers, to understand data requirements and deliver optimal solutions.
• Document data engineering processes, data flows, and technical specifications.
• Keep up-to-date with the latest advancements in Spark, Scala, and related technologies, frameworks, and tools.
• Continuously improve skills and knowledge through self-learning and professional development.
Minimum Years of Experience: 6 years
Certifications Needed:
No
Top 3 responsibilities you would expect the Subcon to shoulder and execute:
• Strong programming skills in Java, Scala, or Python
• Familiarity with big data processing tools and techniques
• Experience with the Hadoop ecosystem
Interview Process (Is face to face required?)
No
Any additional information you would like to share about the project specs/ nature of work:
Drug test details: No