Big Data with Python

5 - 10 years

0.5 - 0.9 Lacs P.A.

Hyderabad, Pune, Bengaluru

Posted:2 weeks ago| Platform: Naukri logo

Apply Now

Skills Required

PythonPysparkBig DataSpark

Work Mode

Hybrid

Job Type

Full Time

Job Description

Role & responsibilities Looking for experience in Big Data technologies. • Mandatory - Hands on experience in Python and PySpark. • Build pySpark applications using Spark Dataframes in Python using Jupyter notebook and PyCharm(IDE). • Worked on optimizing spark jobs that processes huge volumes of data. • Hands on experience in version control tools like Git. • Worked on Amazons Analytics services like Amazon EMR, Amazon Athena, AWS Glue. • Worked on Amazon’s Compute services like Amazon Lambda, Amazon EC2 and Amazon’s Storage service like S3 and few other services like SNS. • Experience/knowledge of bash/shell scripting will be a plus. • Has built ETL processes to take data, copy it, structurally transform it etc. involving a wide variety of formats like CSV, TSV, XML and JSON. • Experience in working with fixed width, delimited , multi record file formats etc. • Good to have knowledge of datawarehousing concepts – dimensions, facts, schemas- snowflake, star etc. • Have worked with columnar storage formats- Parquet,Avro,ORC etc. Well versed with compression techniques – Snappy, Gzip. • Good to have knowledge of AWS databases (atleast one) Aurora, RDS, Redshift, ElastiCache, DynamoDB. • Hands on experience in tools like Jenkins to build, test and deploy the applications • Awareness of Devops concepts and be able to work in an automated release pipeline environment. • Excellent debugging skills. Role & responsibilities

RecommendedJobs for You

Noida, Bengaluru, Thiruvananthapuram