Requires a Bachelors degree in Computer Engineering, Electronic Engineering, or a related field and 7 years of IT experience with large scale distributed framework. Experience must include building enterprise data lake, Datawarehouse and DataMart; strong analytic skills related to working with structured and unstructured datasets; building processes supporting data transformation, data structures, metadata, dependency, and workload management; manipulating, processing, and extracting value from large disconnected datasets; message queuing, stream processing, and highly scalable big data data stores; designing and implementing realtime data pipelines to create data assets; developing and designing dashboards using Qlik, Spotfire, Tableau, QuickSight, and Superset; querying massive datasets using Spark, Presto, Hive, and Impala; using one or more opensource Big Data technologies such as Apache Nifi, Hadoop, YARN, Spark, Hive, Presto, Kafka, Impala, Parquet, or HDFS; working with multi cloud or onprem Big DataMPP analytics platform such as Netezza, Teradata, AWS Redshift, Google BigQuery, or AWS Redshift Data Warehouse; using one or more opensource stream processing technologies such as Kenisis, Spark Streaming, or Kafka Streams; using workflow management engines such as Airflow, Google Cloud Composer, or AWS Step Functions; API management tools and working with SOAPREST APIs; complex query on relational SQL and NoSQL databases, including Postgres and Cassandra; programming experience manipulating and analyzing data using Python, R, or Scala on big data stores; developing and analyzing the MLAI Data workflows; and knowledge of container and orchestration frameworks including Docker and Kubernetes. Must be certified Qlik Designer, Qlik Developer, and Cloudera Developer for Apache Hadoop.br br Per Executive Order, Regeneron requires all U.S. employees to be fully vaccinated against COVID19, unless they are legally entitled to an accommodation for medical or religious reasons.
Categories: eb3
0 Comments