Skill Set: AWS, Snowflake, Kafka, Airflow, GitHub, PySpark Key Responsibilities: Design, develop, and maintain scalable ETL/ELT pipelines Ingest data from various sources (APIs, databases, files, etc.) Implement both real-time and batch processing solutions based on use case requirements Ensure data quality through validation and cleansing processes Collaborate with Product Managers and Business Stakeholders to gather and understand data requirements Translate business needs into technical specifications Ensure data security, access control, and compliance with relevant policies Maintain documentation and follow best practices for data engineering Ideal candidate profile: 4-8 years of hands-on experience with Snowflake, Airflow and Pyspark in any cloud platform (AWS/Azure/GCP). The candidate should be serving notice with LWD in the next 3 weeks.