Sand Technologies
Nairobi, Cairo, Lagos, Johannesburg
Egypt, Kenya, South Africa, Nigeria
Min 7 yrs experience
70k-95k USD per annum
What we are looking for
Sand Technologies is looking for a Senior Data Engineer. the Senior Engineer will focus on building the data infrastructure for a best-in-class intelligence platform and be responsible for ensuring data availability, quality, and security.
What does the job involve
-
Data Pipeline Development: Lead the design, implementation, and maintenance of scalable data pipelines using tools such as Databricks, Python, and PySpark.
-
Data Architecture: Architect scalable and efficient data solutions using modern architectures.
-
Data Modeling: Design and optimize data models and schemas for efficient storage, retrieval, and analysis of structured and unstructured data.
-
ETL Processes: Develop, optimize, and automate ETL workflows.
-
Big Data Technologies: Utilize technologies like Spark, Kafka, and Flink for distributed data processing and analytics.
-
Cloud Platforms: Deploy and manage data solutions on AWS, Azure, or Google Cloud Platform (GCP).
-
Data Quality and Governance: Implement and oversee data governance, quality, and security measures.
-
Monitoring, Optimization, and Troubleshooting: Monitor and optimize pipeline performance, troubleshoot issues.
-
DevOps: Build and maintain CI/CD pipelines, commit code to version control, and deploy data solutions.
-
Best Practices: Stay current with emerging technologies and best practices in data engineering, cloud architecture, and DevOps.
-
Technology Selection: Evaluate and recommend technologies, frameworks, and tools.
-
Performance Optimization: Optimize software performance through architectural design decisions and performance tuning.
Qualifications
-
Programming Languages: Proficiency in Python, Java, Scala, or SQL.
-
Data Modeling: Strong understanding of data modeling concepts and techniques.
-
Big Data Technologies: Experience with Databricks, Spark, Kafka, and Flink.
-
Modern Data Architectures: Experience with lakehouse architectures.
-
CI/CD and Version Control: Experience with CI/CD pipelines, Git, and containerization (e.g., Docker).
-
ETL Tools: Knowledge of tools like Apache Airflow, Informatica, or Talend.
-
Data Governance: Strong understanding of data governance and best practices.
-
Cloud Platforms: Experience with AWS, Azure, or GCP services.
-
SQL and Real-Time Data Streaming: Proficiency in SQL and real-time data streaming technologies like Apache Spark Streaming and Kafka.
-
Cloud Data Tools: Experience using data tools in at least one cloud service - AWS, Azure, or GCP (e.g., S3, EMR, Redshift, Glue, Azure Data Factory, Databricks, BigQuery, Dataflow, Dataproc).
Benefits
-
TBDContact: admin@globaljobservices.vn
0 comments