Must have experience with each of the following 1 Data Ingestion using PySpark; 2 Data processing using Hadoop HDFS; 3 Developing Big Data and Data Lake frameworks within AWS environment; 4 Machine learning algorithm development using Python; 5 Data ETL Pipelines using PySpark; 6 Orchestrating data pipelines using Active Batch. Remote work for 80 of the time is permitted.
Categories: eb3
0 Comments