Data Engineer Profiles
Data Engineer Profiles
Data Engineer Profiles
SQL
Hands on Experience using AWS core services: EC2, S3, VPC, ELB, lambda.
Experience in building ETL routines in AWS (including Ingress and Egress)
Experience in data profiling and source system data analysis
Understand the concept and principles of data modeling
Good to have:
Experience working with scripting languages like Python, Bash, Pyspark
Experience in redshift and other PAAS databases in AWS
Experience in building Customer MDM solutions
ust Have : Spark with underlying technology Python” and "Hadoop ecosystem, architecture
of Hadoop, YARN, Sqoop, Hive and PySpark.
Nice to have :
5+ years demonstratable work experience as a Data Architect using data modeling tools such as
Erwin, Toad or ER/Studio
2-3 years T-SQL coding experience on large projects working with newer versions of SQL Server (2014
and newer)
Must possess expert T-SQL programming, performance tuning and troubleshooting skills
Create database objects and T-SQL scripts
Experience designing data warehousing solutions
In-depth understanding of database structure principles
Seamlessly able to translate business requirements to technology requirements
Familiarity with Microsoft Azure, Azure SQL, Azure Data Flow & Azure Data Factory
Expertise with No-SQL databases such as Azure Cosmos D
Overall experience of 6-8 years with 4-6 years of strong experience building ETLs and Data pipelines
and optimizing using SQL, Hive, Python and Pyspark. 2+ years of experience on AWS Bigdata
solutions like EMR (or Hadoop-Hive), Glue and PySpark and Redshift. 1+ years of experience in AWS
using services like S3, EC2, RDS, EMR, Redshift. Experience building and optimizing RDBMS/BIGDATA
data pipelines, architecture and data sets.
Looking for a tech savvy Data Engineer to Design, Develop and Support ETL interfaces
of a big data marketing technology platform built on AWS. Understand the existing
landscape, document and optimize the pipelines for best performance. Interact with Business
and Marketing users, Data Scientists and other developers.
Prefer candidates having exposure to Data science (Model creation and execution
Good to have knowledge in ETL tools such as Glue, Spark etc.
Strong communication and interpersonal skills
Minimum 4 years of relevant experience in data engineering, including strong experience with SQL,
Python,PySpark.
Strong Experience in any ETL tool like Informatica, DataStage, Alteryx, Talend, etc.
Expertise in writing complex SQL and PL SQL queries.
Proficiency in at least one of the cloud technologies: Azure, AWS, or GCP
3. Hive or spark
4. AWS/GCP
5. Airflow
Good to have:
Additional Skills: