Responsibilities: ● Design, develop, and maintain scalable ETL/ELT pipelines to process structured and unstructured data. ● Manage and optimize databases, data warehouses, and data lakes (SQL/NoSQL, BigQuery, etc.). ● Ensure data quality, governance, and reliability through validation, monitoring, and automation. ● Optimize pipelines and queries for large datasets and high-volume transactions. ● Explore and analyze datasets using statistical methods to identify trends and insights. ● Build, validate, and fine-tune predictive and prescriptive machine learning models relevant to the business. ● Deploy models into production by integrating them with data pipelines and business applications. ● Communicate findings and recommendations through dashboards, visualizations, and reports. ● Set up, maintain, and monitor Google Cloud Platform services (BigQuery, Dataflow, Cloud Storage, Pub/Sub, Composer). ● Manage ETL workflows on GCP, ensuring reliability, scalability, and cost efficiency. ● Implement security, access control, and compliance in cloud-based data systems. ● Collaborate with cross-functional teams (business, sales, operations) to ensure data is actionable and accessible.