| Role Description: | Lead the design, development, and deployment of PySpark-based big data solutions. Architect and optimize ETL pipelines for structured and unstructured data. Collaborate with Client, data engineers, data scientists, and business teams to understand requirements and provide scalable solutions. Optimize Spark performance through partitioning, caching, and tuning. Implement best practices in data engineering (CI/CD, version control, unit testing). Work with cloud platforms like AWS Ensure data security, governance, and compliance. Mentor junior developers and review code for best practices and efficiency. |