Responsibilities

  • Define technical scope and objectives through research and participation in requirements-gathering and definition of processes.
  • Gather and process raw, structured, semi-structured, and unstructured data at scale, including writing scripts and developing programmatic interfaces against web APIs.
  • Design, review, implement and optimize data transformation processes in the Hadoop ecosystem.
  • Test and prototype new data integration tools, techniques and methodologies.
  • Participate in functional test planning and testing for the assigned application integrations, functional areas and projects.
  • Work with the team in an Agile environment to ensure a quality product is delivered.
  • Rapid response and cross-functional work to deliver appropriate resolution of technical, procedural, and operational issues.

Qualifications

  • MS/BS degree in Computer Science, related technical field, or equivalent work experience.
  • Minimum of 5 years of experience with the following:
    • Experience architecting and integrating the Hadoop platform with traditional RDBMS data warehouses.
    • Experience with major Hadoop distributions such as Cloudera (preferred), HortonWorks, MapR, BigInsights, or Amazon EMR is essential.
    • Experience developing within the Hadoop platform including Java, MapReduce, Hive, Spark, Python.
    • Experience with scheduling technologies such as Azkaban, Oozie, or Control-M.
  • Excellent oral and written communication skills
  • Excellent customer service skills
  • Excellent analytical and problem solving skills
  • Working knowledge of Linux O/S and Solaris environments 

Desirable Skills & Experience

  • Experience with logical, 3NF or Dimensional data models.
  • Experience with Netezza, Oracle, SQL Server.
  • Experience with Hadoop ecosystem technologies like Flume, Sqoop, NiFi, Spark Streaming.
  • Knowledge of Java SE, Java EE, JMS, XML, XSL, Web Services and other application integration-related technologies
  • Experience with NoSQL databases such as HBase, Cassandra, Redis or MongoDB.
  • Experience with Pig and Pig UDF development.
  • Certifications from Cloudera, HortonWorks and/or MapR.
  • Familiarity with Business Analytics tools and platforms such as Tableau, Jaspersoft, Business Objects, MicroStrategy, Platfora, a plus.
  • Experience in working in an Agile.

Location: Dallas, TX.