Must have experience with each of the following 1 Data Ingestion using PySpark; 2 Data processing using Hadoop HDFS; 3 Developing Big Data and Data Lake frameworks within AWS environment; 4 Machine learning algorithm development using Python; 5 Data ETL Pipelines using PySpark; 6 Orchestrating data pipelines using Active Batch. Remote work for 80 of the time is permitted.

Categories: eb3

0 Comments

Leave a Reply

Avatar placeholder

Your email address will not be published. Required fields are marked *