Data Engineer with Talend & AWS
Apply NowCompany: CapB InfoteK
Location: Hartford, CT 06106
Description:
CapB is a global leader on IT Solutions and Managed Services. Our R&D is focused on providing cutting edge products and solutions across Digital Transformations from Cloud, AI/ML, IOT, Blockchain to MDM/PIM, Supply chain, ERP, CRM, HRMS and Integration solutions. For our growing needs we need consultants who can work with us on salaried or contract basis. We provide industry standard benefits, and an environment for LEARNING & Growth.
For one of our ongoing projects we are looking for a Data Engineer with Talend & Spark. The position is based out of Hartford but can be done remote as of now.
Responsibilities:
Continuously optimize, enhance, monitor, support and maintain all Talend data integration processes and should be an expert in Talend Big Data jobs.
Use APIs or source XML Type columns to dynamically extract, integrate and load data in target schema using Talend.
Should be responsible for building extraction and mapping rules for loading data from multiple sources for greenfield data warehouse implementation based on Talend, AWS and Snowflake.
Should have experience working with multiple file formats especially with Parquet and Avro.
Should have experience with moving data into S3 folder structures and working with Talend Spark jobs on AWS EMR.
Should contribute to logical data model for data warehousing and making data available for downstream consumption
Maintain documentation, manage source code and deployments, and implement best-practices
Requirements:
5+ years of total experience in IT and seeking previous 3+ years with Talend ETL / Data Integration / Big Data / Cloud (AWS) Technologies experience
Knowledge of Oracle, Unix/Linux Shell scripting, Autosys scheduling tool and version control Tools
Must have hands on experience on AWS EMR/S3/Lambda/Kinesis/Batch
Hybrid/Multi Cloud implementation experience in Data Warehouse
Experience on Snowflake
CI/CD pipeline setup using Jenkins
Nice to have experience with P&C Insurance
AWS/ Talend certification preferred
For one of our ongoing projects we are looking for a Data Engineer with Talend & Spark. The position is based out of Hartford but can be done remote as of now.
Responsibilities:
Continuously optimize, enhance, monitor, support and maintain all Talend data integration processes and should be an expert in Talend Big Data jobs.
Use APIs or source XML Type columns to dynamically extract, integrate and load data in target schema using Talend.
Should be responsible for building extraction and mapping rules for loading data from multiple sources for greenfield data warehouse implementation based on Talend, AWS and Snowflake.
Should have experience working with multiple file formats especially with Parquet and Avro.
Should have experience with moving data into S3 folder structures and working with Talend Spark jobs on AWS EMR.
Should contribute to logical data model for data warehousing and making data available for downstream consumption
Maintain documentation, manage source code and deployments, and implement best-practices
Requirements:
5+ years of total experience in IT and seeking previous 3+ years with Talend ETL / Data Integration / Big Data / Cloud (AWS) Technologies experience
Knowledge of Oracle, Unix/Linux Shell scripting, Autosys scheduling tool and version control Tools
Must have hands on experience on AWS EMR/S3/Lambda/Kinesis/Batch
Hybrid/Multi Cloud implementation experience in Data Warehouse
Experience on Snowflake
CI/CD pipeline setup using Jenkins
Nice to have experience with P&C Insurance
AWS/ Talend certification preferred