We have two Data Engineering roles available one in Bangalore , India and one in Carlsbad, CA , United States local candidates are preferred for the Carlsbad, CA , United States position.
Position : Data Engineer
Experience : 3-5 years
Location : Bangalore
Duration : Full Time
Experience : 3-5 years
Location : Bangalore
Duration : Full Time
Job Description:
- Responsible to Ingest data from files, streams and databases.
- Process the data with #Python and #Pyspark and its storage into a time series database.
- Develop programs in #Python as part of data extraction, data cleaning, transformation and processing.
- Develop and maintain scalable data pipelines.
- Rest API’s development.
- Advanced working knowledge on #SQL to create complex queries
- Hands on experience with visualization tools like #Grafana & #PowerBI
- Experience working on #Azure cloud services #IaaS , #PaaS
- Responsible to Ingest data from files, streams and databases.
- Process the data with #Python and #Pyspark and its storage into a time series database.
- Develop programs in #Python as part of data extraction, data cleaning, transformation and processing.
- Develop and maintain scalable data pipelines.
- Rest API’s development.
- Advanced working knowledge on #SQL to create complex queries
- Hands on experience with visualization tools like #Grafana & #PowerBI
- Experience working on #Azure cloud services #IaaS , #PaaS
Nice-to-Have Skills:
Certification on any cloud platform. Hands-on experience on Docker Containers and microservices architecture using #Kubernetes. Hand on experience in #Golang programming.
Position : Data Engineer
Experience : 4-8 yearsLocation : Carlsbad, CA
Duration : Full Time
Note: We need resumes of the candidates who have worked on data integration/building pipelines using #Python and/or #Pyspark as programming language ideally on Databircks platform.
Job Description:
- Design, develop, test, deploy, support, enhance data integration solutions seamlessly to connect and integrate enterprise systems in our Enterprise Data Platform.
- Innovate for data integration in Apache Spark-based Platform to ensure the technology solutions leverage cutting edge integration capabilities.
- Experience with ETL, data pipeline creation to load data from multiple data sources.
Primary Skills:
- 4+ years working experience in data integration and pipeline development with BS degree in CS, CE or EE.
- 2+ years of Experience with AWS Cloud on data integration with Apache Spark, EMR, Glue, Kafka, Kinesis, and Lambda in S3, Redshift, RDS, MongoDB/DynamoDB ecosystems. - Strong real-life experience in python development especially in PySpark in AWS Cloud environment.
- Design, develop test, deploy, maintain and improve data integration pipeline. Experience in Python and common python libraries.
- Strong analytical experience with database in writing complex queries, query optimization, debugging, user defined functions, views, indexes etc.
- Strong experience with source control systems such as Git, Bitbucket, and Jenkins build and continuous integration tools. Databricks, Redshift Experience is a plus.
Please send your resumes to abhinav.galaxyinc@gmail.com