Big Data Engineer

Careerbuilder-US
Careerbuilder-US

Job Overview

Title:Big Data Engineer

Mandatory skills :

MapReduce development, datasets
Hadoop, HDFS, YARN
Spark, Hive, Kafka, airflow
Apache HBase, AWS big data
ETL pipelines, collect data
data scientists, machine learning, big data infrastructure, cloud
data retention, data performance
data manipulation, ingestion tools, sqoop, flume
machine learning development process

Job Description:

The client network generates a large sum of data each day form of communications data, network device data, log files, customer interaction data, etc.
Resource will be architecting a “big data” platform to store, process these data to be used by data scientists and machine learning engineers.
You will have the opportunity to work with a small team of data scientists and machine learning engineers to build products and services to improve the state of the client network and elevate the customer experience.
You will design, develop, test, and maintain big data infrastructure in the cloud and on-prem locations.
You will develop ETL pipelines to collect data from various sources, transform and store them, and enable stakeholders to consume it.
You will be developing pipelines to support machine learning application development processes.
You will be working with different parts of a large organization to locate, understand, and extract data from a diverse variety of systems and transform them into a big data platform.
Monitoring data performance and modifying infrastructure as
Define data retention policies
Qualifications
Minimum:
Computer Science degree or relevant experience.
5+ years of industry experience in Data Engineering not necessarily Telecom
Strong experience with MapReduce development for large datasets (Hadoop, HDFS, YARN).
Strong background in Linux
5+ years of experience in python
Industry experience in developing ETL pipelines to manage large datasets.
Working knowledge of machine learning development process
Preferred:
Masters degree in computer science
Experience in terabyte scale data manipulation.
Experience in ingestion tools such as sqoop, and flume is a plus.
Experience in Kafka and airflow is a plus.
Processing framework such as Spark and Hive is a plus
Experience in Apache HBase is a plus
AWS big data certification is a plus
Note :
Hours:8:00am to 5:00pm
REMOTE POSITION BUT CANDIDATE MUST BE NEAR A CLIENT OFFICE LOCATION.

Job Requirements:

Mandatory skills:

MapReduce development, datasets
Hadoop, HDFS, YARN
Spark, Hive, Kafka, airflow
Apache HBase, AWS big data
ETL pipelines, collect data
data scientists, machine learning, big data infrastructure, cloud
data retention, data performance
data manipulation, ingestion tools, sqoop, flume
machine learning development process

VIVA USA is an equal opportunity employer and is committed to maintaining a professional working environment that is free from discrimination and unlawful harassment. The Management, contractors, and staff of VIVA USA shall respect others without regard to race, sex, religion, age, color, creed, national or ethnic origin, physical, mental or sensory disability, marital status, sexual orientation, or status as a Vietnam-era, recently separated veteran, Active war time or campaign badge veteran, Armed forces service medal veteran, or disabled veteran. Please contact us at [email protected] for any complaints, comments and suggestions.

Please send your resumes to “[email protected]” and one of our recruiter will get in touch with you.

Contact Details:

Account co-ordinator: Dinesh Kumar U, Phone No: (847) 461-3640 , Email: [email protected]

VIVA USA INC.
3601 Algonquin Road, Suite 425
Rolling Meadows, IL 60008
[email protected] | http://www.viva-it.com

View More
Job Detail
Shortlist Never pay anyone for job application test or interview.