Eviden, part of the Atos Group, with an annual revenue of circa € 5 billion is a global leader in data-driven, trusted and sustainable digital transformation. As a next generation digital business with worldwide leading positions in digital, cloud, data, advanced computing and security, it brings deep expertise for all industries in more than 47 countries. By uniting unique high-end technologies across the full digital continuum with 47,000 world-class talents, Eviden expands the possibilities of data and technology, now and for generations to come.
Responsibilities:
Experience in Big Data technologies.
Mandatory - Hands-on experience in Python and PySpark.
Build PySpark applications using Spark DataFrames in Python using Jupyter notebook and PyCharm (IDE).
Optimize Spark jobs that process huge volumes of data.
Hands-on experience in version control tools like Git.
Experience with Amazon’s Analytics services like Amazon EMR, Amazon Athena, AWS Glue.
Experience with Amazon’s Compute services like Amazon Lambda, Amazon EC2, and Amazon’s Storage service like S3.
Experience/knowledge of bash/shell scripting will be a plus.
Build ETL processes to take data, copy it, and structurally transform it involving a wide variety of formats like CSV, TSV, XML, and JSON.
Experience in working with fixed width, delimited, multi-record file formats.
Good knowledge of data warehousing concepts – dimensions, facts, schemas - snowflake, star, etc.
Experience with columnar storage formats - Parquet, Avro, ORC, and compression techniques – Snappy, Gzip.
Good knowledge of AWS databases (at least one) such as Aurora, RDS, Redshift, Elasticache, DynamoDB.
Hands-on experience in tools like Jenkins to build, test, and deploy applications.
Awareness of DevOps concepts and ability to work in an automated release pipeline environment.
Willingness to work in shift timings of 2:00 PM - 11:00 PM.