The Data Engineer is responsible for building data ingestion processes, ensuring data quality, monitoring performance, and collaborating with the data science team to support analytics and machine learning initiatives.
Description and Requirements
Position Summary
Role Value Proposition:
Work and collaborate with a nimble, autonomous, cross-functional team of makers, breakers, doers, and disruptors who love to solve real problems and meet real customer needs.
You will be using cutting-edge technologies and frameworks to analyze data, to help create the pipeline, and collaborate with the data science team to enable the innovative work in machine learning and AI.
Eagerness to learn new technologies on the fly and ship to production
Knowledge in data science is a plus
More than just a job we hire people who love what they do!
Job Responsibilities
Building and Implementing data ingestion and curation process developed using Big data tools such as Spark (Scala/python), Data bricks, Delta lake, Hive, Pig, Spark, HDFS, Oozie, Sqoop, Flume, Zookeeper, Kerberos, Sentry, Impala etc. Ingesting huge volumes data from various platforms for Analytics needs and writing high-performance, reliable, and maintainable ETL code. Monitoring performance and advising any necessary infrastructure changes. Defining data security principals and policies using Ranger and Kerberos. Assisting application developers and advising on efficient big data application development using cutting edge technologies.
Knowledge, Skills and Abilities
Education
Experience
Knowledge and skills (general and technical)
Required:
Preferred:
About MetLife
Recognized on Fortune magazine's list of the "World's Most Admired Companies" and Fortune World's 25 Best Workplaces™, MetLife, through its subsidiaries and affiliates, is one of the world's leading financial services companies; providing insurance, annuities, employee benefits and asset management to individual and institutional customers. With operations in more than 40 markets, we hold leading positions in the United States, Latin America, Asia, Europe, and the Middle East.
Our purpose is simple - to help our colleagues, customers, communities, and the world at large create a more confident future. United by purpose and guided by our core values - Win Together, Do the Right Thing, Deliver Impact Over Activity, and Think Ahead - we're inspired to transform the next century in financial services. At MetLife, it's #AllTogetherPossible . Join us!
#BI-Hybrid
Position Summary
Role Value Proposition:
Work and collaborate with a nimble, autonomous, cross-functional team of makers, breakers, doers, and disruptors who love to solve real problems and meet real customer needs.
You will be using cutting-edge technologies and frameworks to analyze data, to help create the pipeline, and collaborate with the data science team to enable the innovative work in machine learning and AI.
Eagerness to learn new technologies on the fly and ship to production
Knowledge in data science is a plus
More than just a job we hire people who love what they do!
Job Responsibilities
Knowledge, Skills and Abilities
Education
- Bachelor's degree in Computer Science, Engineering, or related discipline
Experience
- 4+ years of solutions development experience
- Proficiency and extensive Experience with Spark & Scala, Python and performance tuning is a MUST
- Hive database management and Performance tuning is a MUST. (Partitioning / Bucketing)
- Strong SQL knowledge and data analysis skills for data anomaly detection and data quality assurance.
- Strong analytic skills related to working with unstructured datasets.
- Experience with building stream-processing systems, using solutions such as Storm or Spark-Streaming
- Experience in any model management methodologies.
Knowledge and skills (general and technical)
Required:
- Proficiency and extensive experience in HDFS, Hive, Spark, Scala, Python, Databricks/Delta Lake, Flume, Kafka etc.
- Analytical skills to analyze situations and come to optimal and efficient solution based on requirements.
- Performance tuning and problem-solving skills is a must
- Hive database management and Performance tuning is a MUST. (Partitioning / Bucketing)
- Hands on development experience and high proficiency in Java or, Python, Scala, and SQL
- Experience designing multi-tenant, containerized Hadoop architecture for memory/CPU management/sharing across different LOBs
Preferred:
- Proficiency and extensive Experience with Spark & Scala, Python and performance tuning is a MUST
- Hive database management and Performance tuning is a MUST. (Partitioning / Bucketing)
- Strong SQL knowledge and data analysis skills for data anomaly detection and data quality assurance.
- Knowledge in data science is a plus
- Experience with Informatica PC/BDM 10 and implemented push down processing into Hadoop platform, is a huge plus.
- Proficiency is using tools Git, Bamboo and other continuous integration and deployment tools
- Exposure to data governance principles such as Metadata, Lineage (Colibra /Atlas) etc.
About MetLife
Recognized on Fortune magazine's list of the "World's Most Admired Companies" and Fortune World's 25 Best Workplaces™, MetLife, through its subsidiaries and affiliates, is one of the world's leading financial services companies; providing insurance, annuities, employee benefits and asset management to individual and institutional customers. With operations in more than 40 markets, we hold leading positions in the United States, Latin America, Asia, Europe, and the Middle East.
Our purpose is simple - to help our colleagues, customers, communities, and the world at large create a more confident future. United by purpose and guided by our core values - Win Together, Do the Right Thing, Deliver Impact Over Activity, and Think Ahead - we're inspired to transform the next century in financial services. At MetLife, it's #AllTogetherPossible . Join us!
#BI-Hybrid
Top Skills
Bamboo
Bdm
Data Bricks
Delta Lake
Flume
Git
Hdfs
Hive
Impala
Informatica Pc
Java
Kerberos
Oozie
Pig
Python
Scala
Sentry
Spark
SQL
Sqoop
Zookeeper
Similar Jobs at MetLife
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
The role involves delivering digital solutions, managing cloud infrastructure, developing CI/CD pipelines, and engaging with stakeholders in an Agile environment.
Top Skills:
Api ManagementAzureAzure DevopsDatabricksGitJavaPythonReactSparkSQL
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
The role involves designing, implementing, and managing Azure cloud infrastructure, automating deployments with Terraform, and ensuring security and compliance across cloud environments, collaborating with internal and external stakeholders.
Top Skills:
AgileAzureCi/CdDevOpsPythonTerraform
Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
The Software Platform Engineer II will provide L2 support for AEM and Java issues, enhance performance, troubleshoot problems, and collaborate on system security and stability.
Top Skills:
AemJava
What you need to know about the Kolkata Tech Scene
When considering the industries shaping India's tech scene, gaming might not immediately come to mind. However, in the last decade, increased internet usage and greater access to mobile devices have catapulted the industry to new heights, with Kolkata-based companies like Virtualinfocom, Red Apple Technologies and Digitoonz, at the forefront, driving the design and animation of new gaming titles for players.