Senior Data Engineer (Python/PySpark/AWS)

Infinitive Inc

Ashburn, VA

JOB DETAILS
JOB TYPE
Full-time
SKILLS
Amazon Web Services (AWS), Analysis Skills, Artificial Intelligence (AI), Best Practices, Cloud Computing, Communication Skills, Computer Programming, Computer Science, Consulting, Continuous Deployment/Delivery, Continuous Integration, Cross-Functional, Customer Support/Service, Customer/Client Research, Data Analysis, Data Management, Data Modeling, Data Processing, Data Quality, Data Recovery, Data Science, Data Storage, Data Warehousing, Database Design, Database Extract Transform and Load (ETL), Develop and Maintain Customers, Docker, Git, GitHub, Information Technology & Information Systems, Information/Data Security (InfoSec), Jenkins, Microsoft Windows Azure, Performance Tuning/Optimization, Privacy Controls, Problem Solving Skills, Python Programming/Scripting Language, Return on Investment (ROI), Scripting (Scripting Languages), Software Engineering, Source Code/Configuration Management (SCM), Team Player, Test Automation
LOCATION
Ashburn, VA
POSTED
30+ days ago

*Candidates must be local to the Washington D.C. metro area. 

About Infinitive:

Infinitive is a data and AI consultancy that enables its clients to modernize, monetize and operationalize their data to create lasting and substantial value. . We possess deep industry and technology expertise to drive and sustain adoption of new capabilities. We match our people and personalities to our clients' culture while bringing the right mix of talent and skills to enable high return on investment.

Infinitive has been named “Best Small Firms to Work For” by Consulting Magazine 7 times most recently in 2024. Infinitive has also been named a Washington Post “Top Workplace”, Washington Business Journal “Best Places to Work”, and Virginia Business “Best Places to Work.”

We are seeking a highly skilled and motivated Data Engineer to join our dynamic team. As a Data Engineer, you will play a crucial role in designing, developing, and maintaining our clients data infrastructure. Your expertise in Python, PySpark, ETL processes, CI/CD (Jenkins or GitHub), and experience with both streaming and batch workflows will be essential in ensuring the efficient flow and processing of data to support our clients.

Responsibilities:

Data Architecture and Design:

  • Collaborate with cross-functional teams to understand data requirements and design robust data architecture solutions.
  • Develop data models and schema designs to optimize data storage and retrieval.

ETL Development:

  • Implement ETL processes to extract, transform, and load data from various sources.
  • Ensure data quality, integrity, and consistency throughout the ETL pipeline.

Python and PySpark Development:

  • Utilize your expertise in Python and PySpark to develop efficient data processing and analysis scripts.
  • Optimize code for performance and scalability, keeping up-to-date with the latest industry best practices.

Data Integration:

  • Integrate data from different systems and sources to provide a unified view for analytical purposes.
  • Collaborate with data scientists and analysts to implement solutions that meet their data integration needs.

Streaming and Batch Workflows:

  • Design and implement streaming workflows using PySpark Streaming or other relevant technologies.
  • Develop batch processing workflows for large-scale data processing and analysis.

CI/CD Implementation:

  • Implement and maintain continuous integration and continuous deployment (CI/CD) pipelines using Jenkins or GitHub Actions.
  • Automate testing, code deployment, and monitoring processes to ensure the reliability of data pipelines.

Qualifications:

  • Bachelor's or Master's degree in Computer Science, Information Technology, or a related field.
  • 7+ years of proven experience as a Data Engineer or similar role.
  • Strong programming skills in Python and expertise in PySpark for both batch and streaming data processing.
  • Hands-on experience with ETL tools and processes.
  • Familiarity with CI/CD tools such as Jenkins or GitHub Actions.
  • Solid understanding of data modeling, database design, and data warehousing concepts.
  • Excellent problem-solving and analytical skills.
  • Strong communication and collaboration skills.

Preferred Skills:

  • Knowledge of cloud platforms such as AWS, Azure, or Google Cloud.
  • Experience with version control systems (e.g., Git).
  • Familiarity with containerization and orchestration tools (e.g., Docker, Kubernetes).
  • Understanding of data security and privacy best practices.

Applicants for employment in the U.S. must possess work authorization which does not require sponsorship by the employer for a visa. Infinitive is an Equal Opportunity Employer.

Powered by JazzHR

About the Company

I

Infinitive Inc