Job Details
            
              Location:
              
                
                  Austin, Travis County, Texas, USA
                
              
              Posted:
              
                Jun 04, 2020
              
            
           
          
            
Job Description
             * 3-5 years experience building large-scale distributed systems
* Experience in designing and developing ETL data pipelines. Should be proficient in writing Advanced SQLs, Expertise in performance tuning of SQLs
* We are looking for experience in designing and building dimensional data models to improve accessibility, efficiency, and quality of data
* Database development experience with Relational or MPP/distributed systems such as Oracle/Teradata/Vertica/Hadoop
* Knowledge with at least two of the following: Spark, MapReduce, HDFS, Cassandra, Kafka
* We are seeking programming experience in building high quality software in Java, Python or Scala preferred
* You will demonstrate excellent understanding of development processes and agile methodologies
* Strong analytical and interpersonal skills
* Enthusiastic, highly motivated and ability to learn quick
* Experience with or advance courses on data science and machine learning is ideal
Work/project experience with Big Data and advanced programming languages is a plus
* Experience developing Big Data/Hadoop applications using java, Spark, Hive, Oozie, Kafka, and Map Reduce is a huge plus
* Exceptional analytical and programming skills   
          
         
        
          
            
            
              About GlobalLogic
              GlobalLogic is a full-lifecycle product development services leader.
              View Website  
             
           
          
  Get More Interviews for This and Many Other Jobs
  Huntr helps you instantly craft tailored resumes and cover letters, fill out application forms with a single click, effortlessly keep your job hunt organized, and much more.
  Sign Up for Free