The ideal candidate will use their passion for big data and analytics to provide insights into the business covering a range of topics. They will be responsible for developing an in-house ETL-driven toolchain and conducting both recurring as well as on-demand analyses for business users, research folks, and customers.
Responsibilities
- Build optimized Spark-based data processing jobs to generate analytical models and
deploy them into Airflow
- Build Analytical models and Data Models to align with Product strategy
- Understand the day-to-day issues that our business faces, by closely communicating
with stakeholders across the board
- Build data pipelines to facilitate quality checks on datasets across the board; thereby
ensuring a seamless flow of high-quality data into the platform
- Develop a diverse range of visualizations to convey complicated data in a
straightforward fashion, for both internal and external audiences
Qualifications
- Bachelor's or Master's degree.
- 2 - 4 years of experience in the Analytics/engineering domain.
- Proficient in PySpark, SQL (Google BigQuery, etc.), and CI/CD driven deployment
- Proficient in Big-Data related toolkits, Kubernetes and Docker
- Experience in working with the Airflow orchestration engine.
- Redash or Tableau or PowerBI or equivalent visualization tools.
- Problem-solving skills
- Strong communication/interpersonal skills