Are you ready to take the next step in your career? We’re looking for a passionate and driven PySpark Developer to join our growing team .
We’re looking for a talented PySpark Developer to join our data engineering team and help build scalable, high-performance data pipelines. This role is ideal for someone who thrives in distributed computing environments and is passionate about transforming raw data into actionable insights.
What you will do :
– Design, develop, and maintain ETL pipelines using PySpark and other big data technologies.
– Optimize PySpark jobs for performance, scalability, and reliability in distributed environments.
– Collaborate with data engineers, architects, and analysts to transform business requirements into technical solutions.
– Implement data validation, cleansing, and transformation processes.
– Monitor and troubleshoot data processing workflows and ensure data integrity.
– Work with cloud platforms (AWS, Azure, or GCP) to deploy and manage data solutions.
– Document processes and contribute to best practices in data engineering.
What you bring :
– Bachelor’s degree in Computer Science, Engineering, or a related field.
– 6 – 10 years of experience in Python and PySpark development.
– Strong understanding of Apache Spark, Hadoop, and distributed computing principles.
– Experience with SQL, Hive, Kafka, and NoSQL databases (e.g., MongoDB, HBase).
– Familiarity with cloud services like AWS Glue, Azure Data Factory, or GCP Dataflow.
Skills –
Programming – Python, PySpark
Big Data Frameworks – Apache Spark, Hadoop
Data Integration – Kafka, Hive, HBase, MongoDB
Cloud Platforms – AWS, Azure, GCP
Data Formats – JSON, XML, Parquet, Avro
DevOps & Monitoring – Git, Jenkins, Airflow, Datadog
📩 How to Apply
Excited to build the future of data with us? Send your resume to talent@strat-link.com
PySpark Developer
Job Type: Full Time
Job Location: Remote