Apache Arrow, Dremio, Ray * Miscellaneous: Git, Jupyterhub, Apache Superset, Plotly Dash ... MapReduce, Spark, Storm). * Proficiency in Python, AWS services, and ETL/ELT pipelines
Schedule the jobs using Apache Nifi or Air flow * Analyze, recommend and implement improvements to ... Must be Cloudera Certified Hadoop and/or Spark Administrator * Hands-on experience on Cloudera ...
Experience working with large data sets and pipelines, ideally using the Apache software stack (e.g. Spark) * Experience with continuous integration and continuous development solutions (e.g. Jenkins ...
Strong technical background with experience in open-source technologies such as Apache, Hadoop, Spark and/or Kafka * Experience in development/engineering/administration of large scale, customer ...
Apache * Hadoop * Spark * Kafka * AWS/GCP/Azure * EC2/S3/VPC * Prometheus * Grafana * ELK Stack * PostgreSQL * Ansible * Istio * Datadog * CI/CD * Docker * Kubernetes * Jenkins * GitLab * Terraform
Proficient in data analysis using SQL and Spark * Experience implementing and maintaining ... AWS (Lambda, API Gateway, Managed Apache Flink, Kinesis Streams, DynamoDB, Firehose) * Confluent ...
... Spark, Hadoop or Flink, and the file types they deal with. * Experience with ETL and Data pipeline orchestration tools like Apache Airflow, dbt, etc. * Excellent coding skills in Java or Scala ...
Experience in ETL orchestration and workflow management tools with a strong preference for Apache Airflow * Experience in Spark or other distributed computing frameworks * SQL and Python * Advanced ...
Experience with large-scale data analysis systems, such as Databricks, Hadoop, Pig, Scala, Spark or MPP databases * Experience working with Apache Parquet and/or Delta Lake formatted data
... Apache Nifi, Spark, Kafka, HBase, Hadoop/HDFS, Hive, Drill, Pig, etc.) or commercial open source Big Data technology stacks (Hortonworks, Clouder, etc.) * 3+ years with document databases (e.g
Expert knowledge in Apache technologies such as Kafka, Airflow, and Spark to build scalable and efficient data pipelines. * Ability to design, build, and deploy data solutions that capture, explore ...
Experience building and optimizing reliable, idempotent data pipelines (streaming or batch) using ETL tools such as Matillion, Apache Airflow, FiveTran, Kafka and Spark required. * A successful ...
Minimum 2 years of experience with large-scale data processing platforms such as Spark, EMR, and/or HPC computing experience with e.g. Apache Aurora, Slurm. * Minimum 4 years of solutioning ...
Extensive experience with ETL tools (e.g., Apache Airflow, Talend, Informatica). * In-depth ... Experience with big data technologies (e.g., Hadoop, Spark) is a plus. Preferred Qualifications
Must have hands-on experience with big-data technologies, such as Hadoop, Spark, Apache Beam * Must have experience with comprehensive testing (e.g. unit, integration, etc.) * Experience working with ...
... Spark Foundry, Starcom and Zenith. A key business solution of Publicis Groupe ([Euronext Paris ... Web hosting with IIS, NGINX, or APACHE * Basic networking: IP Addressing, Subnets, Routing, and ...
... as Apache Hadoop, Hive or Spark, and/or Databricks * Demonstrated on-the-job experience with large-scale data migration from Oracle on prem to AWS Cloud environment. Demonstrated on-the-job ...
Design and build data solutions using Databricks, SQL, Python, Spark, and Delta Lake in the Azure ... Apache Airflow, Unity Catalog). * Collaborate with engineers, product, and business leaders to ...
Java, Linux, Apache, Perl/Python/PHP, Chef) · Hands-on experience with ETL tools (e.g Informatica ... Snowflake(Snowpark), SPARK, MapReduce, Hadoop, Sqoop, Pig, HBase, Hive, Flume) Benefits
Experience large-scale data analysis systems, such as Databricks, Hadoop, Pig, Scala, Spark or MPP databases * Experience working with Apache Parquet and/or Delta Lake formatted data * Experience ...
Experience with Apache Hadoop, HUE, Hive, Pig, Spark, Elasticsearch, Kibana, or Tableau * Knowledge of programming languages, including Java, Python, R, or SQL * Experience with using JEMA
Excellent understanding of typical multi-tier architectures: web servers (Apache, nginx, IIS ... Exposure to Big Data Infrastructure such as Spark, Storm, and Hadoop * AWS Solutions Architect ...