Wateen is Pakistan’s leading information and communications technology company with a diverse portfolio of customers across industries. A pioneer in the telecommunications industry, Wateen is proud to have laid the foundation of Pakistan’s internet connectivity with the fastest-growing fiber-optic network.
Wateen is credited with transforming the landscape of communications technology in Pakistan through our efforts to revolutionize businesses with cutting-edge technology. With the country’s largest team of highly experienced certified resources, our subject matter experts ensure end-to-end support uniquely optimized for businesses and enterprises across various verticals.
Key Responsibilities
- Build and maintain scalable data pipelines and architecture on Apache Spark clusters, ensuring high availability and performance.
- Use PySpark to develop robust Extract‑Transform‑Load (ETL) jobs that ingest data from diverse sources and load it into data lakes or warehouses.
- Create and schedule Apache Airflow DAGs to automate end-to-end data workflows, including monitoring and alerting.
- Design and develop interactive dashboards and reports in Power BI that align with business requirements, enabling drill-through analysis and actionable insights.
- Build data models to support downstream analytics in tools like Power BI and Cognos, ensuring data is clean, reliable, and business-ready.
- Optimize PySpark jobs, SQL queries, Airflow schedules, and Power BI dashboards for speed, usability, and reliability.
- Perform data profiling, cleansing, and root‑cause analysis; implement durable fixes and document best practices.
- Work closely with analysts, product owners, and cross-functional teams to translate business requirements into reliable datasets and clear metrics.
Location
Lahore
Experience & Skills Required
- 1+ years of experience in a Data Engineering or related role.
- Hands-on expertise running PySpark workloads on on-prem or cloud-based Apache clusters.
- Familiarity with Informatica, DataStage, or equivalent—plus the ability to replicate those pipelines using Spark.
- Experience building, scheduling, and managing Airflow DAGs.
- Strong SQL skills and understanding of dimensional/analytical data modeling.
- Proficient in designing and developing dashboards in Power BI. Experience with Cognos, SSRS, SSIS, or SSAS is a plus.
- Exposure to AWS or Azure data services.
- Strong communication, analytical, and critical-thinking skills; able to clearly present insights to stakeholders and management.