Big Data Platform Support Engineer

Cynet Systems Logo
  • Healthcare

We are looking forBig Data Platform Support Engineer for our client in Reston, VA
Job Title: Big Data Platform Support Engineer
Job Location: Reston, VA
Job Type: Contract
Job Description:

  • The Big Data / Cloudera Lead Engineer will lead the design, development, and support of large-scale data platforms, focusing on Cloudera, Hadoop, Kafka, and related technologies.
  • This role involves ensuring high performance, scalability, and reliability of data pipelines and services while mentoring junior engineers and collaborating with cross-functional teams to optimize system performance and implement new solutions.

Responsibilities:

  • Lead a team of software engineers in the design, development, and implementation of multiple applications.
  • Perform detailed technical design, coding, code reviews, and testing of new and existing applications.
  • Serve as a subject matter expert for specified systems and ensure compliance with DevSecOps practices and policies.
  • Identify gaps in requirements and develop solutions to improve system performance and availability.
  • Lead evaluations of risk, opportunities, and technology integration across projects.
  • Support the Project Manager as the technical lead, estimating technical costs and defining system requirements.
  • Participate in architectural planning, roadmap discussions, and transformation initiatives to new software architectures.
  • Extend open-source libraries and systems to meet team requirements.
  • Ensure adherence to code quality, deployment standards, and practices.
  • Mentor junior developers and foster a culture of continuous improvement.

Experience:

  • 8 years of software design, development, testing, and requirements management.
  • Strong experience supporting Cloudera applications running in cloud environments.
  • Hands-on experience with Cloudera Manager for monitoring and maintaining production clusters.
  • Expertise in writing shell scripts for Hadoop daemon service health checks and automation tasks.
  • Experience working with Hive, Impala, HBase, and SOLR, including query optimization and database management.
  • Proficient in coordinating with application development, security, and platform teams for configuration and performance enhancements.

Skills Must Have:

  • Cloudera administration and optimization (CDP, CDSW, Hive, Spark, NiFi).
  • Cloud application support with MongoDB.
  • Experience in cloud disaster recovery strategies.
  • Strong Kafka administration and troubleshooting, including Streams API, integration with MQ, topic/offset management, and broker management.
  • Apache NiFi administration, registry management, and integration with Kafka, HBase, and Solr.
  • HBase administration, troubleshooting, and database management.
  • Solr administration, managing shards, collections, logging levels, and query optimization.
  • Flume integration with Kafka, NiFi, and IBM MQ.

Additional Skills:

  • Perform critical data migrations from CDH to CDP.
  • Design and implement big data pipelines and automated data flows using Python/R and NiFi.
  • Perform incremental updates and upgrades to Cloudera environments.
  • Manage job workflows using Oozie and Hue.
  • Implement security policies across Hadoop clusters using Ranger.
  • Configure and manage Cloudera Data Science Workbench (CDSW).
  • Troubleshoot Kerberos, TLS/SSL, and model-related issues.
  • Support Big Data/Hadoop databases through all development and production phases.
  • Optimize database performance, troubleshoot replication, and manage connectivity/security issues.
  • Create and implement automation scripts for maintenance and monitoring.
  • Experience with Kafka ecosystem (Brokers, Connect, Zookeeper) in production environments.
  • Support streaming technologies including Kafka, Spark, and Kudu.

Qualification And Education:

  • Bachelor s degree in Information Technology or Computer Science.
  • In lieu of a degree, an additional 4 years of relevant work experience is required.