Skip to main content

Careers Home > Job Search Results > Lead Data Engineer IRC207521

Lead Data Engineer IRC207521

Job ID: IRC207521
Date Posted: Jan 25, 2024
Segment: Digital System & Service
Business Unit: Hitachi Services & Platforms
Company Name: GlobalLogic
Profession (Job Category): Engineering
Job Type (Experience Level): Management
Job Schedule: Full-Time
Remote: Yes

Share: mail
Save Job Saved

Description:

Our Big Data capability team needs hands-on developers who can produce beautiful & functional code to solve complex analytics problems. If you are an exceptional developer with an aptitude to learn and implement using new technologies, and who loves to push the boundaries to solve complex business problems innovatively, then we would like to talk with you.

You would be responsible for evaluating, developing, maintaining and testing big data solutions for advanced analytics projects
The role would involve big data pre-processing & reporting workflows including collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into business insights
The role would also involve testing various machine learning models on Big Data, and deploying learned models for ongoing scoring and prediction. An appreciation of the mechanics of complex machine learning algorithms would be a strong advantage.

Qualification & Experience

5-9 years of demonstrable experience designing technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions.

Ideally, this would include work on the following technologies:

Expert-level proficiency in Scala/PySpark knowledge is a strong advantage. Exp in at least one of Java, Scala or Python (preferred)
Strong understanding and experience in distributed computing frameworks, particularly Apache Hadoop (YARN, MR, HDFS) and associated technologies - one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, Impala, etc.
Hands-on experience with Apache Spark and its components (Streaming, SQL, MLLib) is a strong advantage.
Operating knowledge of cloud computing platforms (AWS/Azure/GCP)
Experience working within a Linux computing environment, and use of command line tools including knowledge of Shell/Python scripting for automating common tasks
Ability to work in a team in an agile setting, familiarity with JIRA and clear understanding of how Git works or any version control tools

In addition, the ideal candidate would have great problem-solving skills, and the ability & confidence to hack their way out of tight corners.

Experience:

Must Have (hands-on) experience:

Scala or Python/PySpark expertise
Distributed computing frameworks (Hadoop Ecosystem & Spark components)
Cloud computing platforms - AWS
Linux environment, SQL and Shell scripting

Nice to have : DevOps knowledge

Requirements:

Our Big Data capability team needs hands-on developers who can produce beautiful & functional code to solve complex analytics problems. If you are an exceptional developer with an aptitude to learn and implement using new technologies, and who loves to push the boundaries to solve complex business problems innovatively, then we would like to talk with you.

You would be responsible for evaluating, developing, maintaining and testing big data solutions for advanced analytics projects
The role would involve big data pre-processing & reporting workflows including collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into business insights
The role would also involve testing various machine learning models on Big Data, and deploying learned models for ongoing scoring and prediction. An appreciation of the mechanics of complex machine learning algorithms would be a strong advantage.

Qualification & Experience

5-9 years of demonstrable experience designing technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions.

Ideally, this would include work on the following technologies:

Expert-level proficiency in Scala/PySpark knowledge is a strong advantage. Exp in at least one of Java, Scala or Python (preferred)
Strong understanding and experience in distributed computing frameworks, particularly Apache Hadoop (YARN, MR, HDFS) and associated technologies - one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, Impala, etc.
Hands-on experience with Apache Spark and its components (Streaming, SQL, MLLib) is a strong advantage.
Operating knowledge of cloud computing platforms (AWS/Azure/GCP)
Experience working within a Linux computing environment, and use of command line tools including knowledge of Shell/Python scripting for automating common tasks
Ability to work in a team in an agile setting, familiarity with JIRA and clear understanding of how Git works or any version control tools

In addition, the ideal candidate would have great problem-solving skills, and the ability & confidence to hack their way out of tight corners.

Experience:

Must Have (hands-on) experience:

Scala or Python/PySpark expertise
Distributed computing frameworks (Hadoop Ecosystem & Spark components)
Cloud computing platforms (AWS/Azure/GCP)
Linux environment, SQL and Shell scripting

Job Responsibilities:

Our Big Data capability team needs hands-on developers who can produce beautiful & functional code to solve complex analytics problems. If you are an exceptional developer with an aptitude to learn and implement using new technologies, and who loves to push the boundaries to solve complex business problems innovatively, then we would like to talk with you.

You would be responsible for evaluating, developing, maintaining and testing big data solutions for advanced analytics projects
The role would involve big data pre-processing & reporting workflows including collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into business insights
The role would also involve testing various machine learning models on Big Data, and deploying learned models for ongoing scoring and prediction. An appreciation of the mechanics of complex machine learning algorithms would be a strong advantage.

Qualification & Experience

5-9 years of demonstrable experience designing technological solutions to complex data problems, developing & testing modular, reusable, efficient and scalable code to implement those solutions.

Ideally, this would include work on the following technologies:

Expert-level proficiency in Scala/PySpark knowledge is a strong advantage. Exp in at least one of Java, Scala or Python (preferred)
Strong understanding and experience in distributed computing frameworks, particularly Apache Hadoop (YARN, MR, HDFS) and associated technologies - one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, Impala, etc.
Hands-on experience with Apache Spark and its components (Streaming, SQL, MLLib) is a strong advantage.
Operating knowledge of cloud computing platforms (AWS/Azure/GCP)
Experience working within a Linux computing environment, and use of command line tools including knowledge of Shell/Python scripting for automating common tasks
Ability to work in a team in an agile setting, familiarity with JIRA and clear understanding of how Git works or any version control tools

In addition, the ideal candidate would have great problem-solving skills, and the ability & confidence to hack their way out of tight corners.

Experience:

Must Have (hands-on) experience:

Scala or Python/PySpark expertise
Distributed computing frameworks (Hadoop Ecosystem & Spark components)
Cloud computing platforms - AWS
Linux environment, SQL and Shell scripting

Nice to have : DevOps knowledge

What We Offer

Exciting Projects: We focus on industries like High-Tech, communication, media, healthcare, retail and telecom. Our customer list is full of fantastic global brands and leaders who love what we build for them.

Collaborative Environment: You Can expand your skills by collaborating with a diverse team of highly talented people in an open, laidback environment - or even abroad in one of our global centers or client facilities!

Work-Life Balance: GlobalLogic prioritizes work-life balance, which is why we offer flexible work schedules, opportunities to work from home, and paid time off and holidays.

Professional Development: Our dedicated Learning & Development team regularly organizes Communication skills training(GL Vantage, Toast Master),Stress Management program, professional certifications, and technical and soft skill trainings.

Excellent Benefits: We provide our employees with competitive salaries, family medical insurance, Group Term Life Insurance, Group Personal Accident Insurance , NPS(National Pension Scheme ), Periodic health awareness program, extended maternity leave, annual performance bonuses, and referral bonuses.

Fun Perks: We want you to love where you work, which is why we host sports events, cultural activities, offer food on subsidies rates, Corporate parties. Our vibrant offices also include dedicated GL Zones, rooftop decks and GL Club where you can drink coffee or tea with your colleagues over a game of table and offer discounts for popular stores and restaurants!

About GlobalLogic

GlobalLogic is a leader in digital engineering. We help brands across the globe design and build innovative products, platforms, and digital experiences for the modern world.

By integrating experience design, complex engineering, and data expertise-we help our clients imagine what's possible, and accelerate their transition into tomorrow's digital businesses.

Headquartered in Silicon Valley, GlobalLogic operates design studios and engineering centers around the world, extending our deep expertise to customers in the automotive, communications, financial services, healthcare and life sciences, manufacturing, media and entertainment, semiconductor, and technology industries.

GlobalLogic is a Hitachi Group Company operating under Hitachi, Ltd. (TSE: 6501) which contributes to a sustainable society with a higher quality of life by driving innovation through data and technology as the Social Innovation Business.
Share: mail

Similar Jobs

DevOps engineer IRC212246

Hitachi Services & Platforms

Angular Lead Engineer IRC210801

Hitachi Services & Platforms

Angular Lead Engineer IRC210785

Hitachi Services & Platforms

Scrum Master for Zipari IRC210516

Hitachi Services & Platforms

SOC Analyst IRC208228

Hitachi Services & Platforms

Senior iOS developer IRC212850

Hitachi Services & Platforms

Tech Lead (React+ NodeJS) IRC201335

Hitachi Services & Platforms

Junior React Native Developer IRC213143

Hitachi Services & Platforms

Android Software Engineer IRC211912

Hitachi Services & Platforms

Senior SAP Signavio Consultant IRC212983

Hitachi Services & Platforms

Senior Automotive Embedded Engineer IRC213037

Hitachi Services & Platforms

Trainee AQA Engineer (Python) IRC212242

Hitachi Services & Platforms

QA Lead Engineer IRC211007

Hitachi Services & Platforms

Angular Sr Engineer IRC211013

Hitachi Services & Platforms

QA Lead Engineer IRC211004

Hitachi Services & Platforms

QA Lead Engineer IRC211006

Hitachi Services & Platforms

.Net Engineer IRC212573

Hitachi Services & Platforms

Senior .NET Developer (Azure) IRC212651

Hitachi Services & Platforms

Vb.NET Developer IRC211739

Hitachi Services & Platforms

DevOps QA Engineer || 5-7 years IRC198859

Hitachi Services & Platforms

Cloud Sr Engineer IRC211167

Hitachi Services & Platforms

Compliance tool IRC206471

Hitachi Services & Platforms

RPG/DB2 Developer IRC208743

Hitachi Services & Platforms

RPG/DB2 Developer IRC208744

Hitachi Services & Platforms