Search by job, company or skills

Exl

Senior Data Engineer

Save
new job description bg glownew job description bg glownew job description bg svg
  • Posted 3 months ago
  • Be among the first 30 applicants
Early Applicant

Job Description

Job Description:

We are looking for a Scala & PySpark, solutions developer and data engineer that is able to design and build solutions for one of our Fortune 500 Client programs, which aims towards building data standardized and curation needs on Hadoop cluster. This is high visibility, fast-paced key initiative will integrate data across internal and external sources, provide analytical insights, and integrate with the customer's critical systems.

Key Responsibilities:

  • Ability to design, build and unit test applications on Spark framework on Scala and Python.
  • Build Spark based applications for both batch and streaming requirements, which will require in-depth knowledge on majority of Hadoop and NoSQL databases as well.
  • Develop and execute data pipeline testing processes and validate business rules and policies
  • Optimize performance of the built Spark applications in Hadoop using configurations around Spark Context, Spark-SQL, Data Frame, and Pair RDD's.
  • Optimize performance for data access requirements by choosing the appropriate native Hadoop file formats (Avro, Parquet, ORC etc) and compression codec respectively.
  • Ability to design & build real-time applications using Apache Kafka & Spark Streaming
  • Build integrated solutions leveraging Unix shell scripting, RDBMS, Hive, HDFS File System, HDFS File Types, HDFS compression codec.
  • Build data tokenization libraries and integrate with Hive & Spark for column-level obfuscation
  • Experience in processing large amounts of structured and unstructured data, including integrating data from multiple sources.
  • Create and maintain integration and regression testing framework on Jenkins integrated with Bitbucket and/or GIT repositories
  • Participate in the agile development process, and document and communicate issues and bugs relative to data standards in scrum meetings
  • Work collaboratively with onsite and offshore team.
  • Develop & review technical documentation for artifacts delivered.
  • Ability to solve complex data-driven scenarios and triage towards defects and production issues
  • Ability to learn-unlearn-relearn concepts with an open and analytical mindset
  • Participate in code release and production deployment.
  • Challenge and inspire team members to achieve business results in a fast paced and quickly changing environment.

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 141932911

Similar Jobs

Noida, India

Skills:

snowflake JavaSQL ServerTableauSqlJenkinsData QualityGitDockerData MigrationShell scriptingSparkVerticaMongoDBRest ApisOracleAzurePythonAWSEtlmicroservices architecture

Gurugram, Gurugram, India

Skills:

PysparkData ModelingAzureBuilding data pipelines

Gurugram, India

Skills:

snowflake Data QualityPysparkAWS GlueData WarehousingData GovernancePythonEtlELTData Catalog

Gurugram, Gurugram, India

Skills:

snowflake Python ScriptingData ModelingPysparkDatabricksAdvanced SqldbtAI coding assistants

Noida, India

Skills:

Spark SQLHadoopScalaApache SparkKafkaHBaseRedshiftSpark StreamingHiveSparkHDFSKafka Connect