What are the responsibilities and job description for the Data Architect position at Inherent Technologies?
Position: Data Architect
Location: Scottsdale, AZ ***Day 1 Onsite***
Duration: 1 Years
1 Must have minimum 710 years of handson experience on building data ingestion highly robust and scalable pipelines using AWS native services Cloud Formation AWS Glue Crawlers Catalog Step Functions Event Bridge Lambda S3 SNS Kinesis other data related services
2 Must be proficient in Python Java
3 Must be proficient with SQL and HQL
4 Must have handled structured and unstructured data ingestion from different file formats TXT CSV JSON XML Parquet etc
5 Must have handled data streaming from Kafka topics using Apache Flink etc
6 Must have prior experience with open table formats like Apache Iceberg Hudi etc
7 Must have exposure to Cloudera data lake environment should have exposure to Snowflake environment
8 Must be proficient with Github versioning activities
9 Prior experience on CICD using Jenkins would be an added advantage
10 Should be able to envision the ingestion end to end while keeping up with all the data privacy lineage the cost optimization requirements
11 Must possess clear written verbal communication skills along with interpersonal skill to collaborate with the team is very important
12 Should be willing to coordinate own the delivery with the team located in India
13 Must have a positive attitude should be open for discussions stay hungry to learn try out something new
14 Must have prior handson working experience in Cloudera Hadoop implementation Should be able to reverse engineer the current processes running on it
15 Prior Apache Flink experience would be ideal
Location: Scottsdale, AZ ***Day 1 Onsite***
Duration: 1 Years
1 Must have minimum 710 years of handson experience on building data ingestion highly robust and scalable pipelines using AWS native services Cloud Formation AWS Glue Crawlers Catalog Step Functions Event Bridge Lambda S3 SNS Kinesis other data related services
2 Must be proficient in Python Java
3 Must be proficient with SQL and HQL
4 Must have handled structured and unstructured data ingestion from different file formats TXT CSV JSON XML Parquet etc
5 Must have handled data streaming from Kafka topics using Apache Flink etc
6 Must have prior experience with open table formats like Apache Iceberg Hudi etc
7 Must have exposure to Cloudera data lake environment should have exposure to Snowflake environment
8 Must be proficient with Github versioning activities
9 Prior experience on CICD using Jenkins would be an added advantage
10 Should be able to envision the ingestion end to end while keeping up with all the data privacy lineage the cost optimization requirements
11 Must possess clear written verbal communication skills along with interpersonal skill to collaborate with the team is very important
12 Should be willing to coordinate own the delivery with the team located in India
13 Must have a positive attitude should be open for discussions stay hungry to learn try out something new
14 Must have prior handson working experience in Cloudera Hadoop implementation Should be able to reverse engineer the current processes running on it
15 Prior Apache Flink experience would be ideal