What are the responsibilities and job description for the Data Architect - Generative AI-W2 position at Vaiticka Solution?
Job Title: Data Architect - Generative AI & Digital Automation
Location: Tarry Town, NY
Employment Type: Full-Time
Mode of work: Hybrid (3 days in office in Tarry Town, NY)
Overview
Seeking a Data Architect to design, develop, and implement cutting-edge solutions leveraging generative AI technology, integration and Digital automation. The ideal candidate will play a pivotal role in architecting scalable, secure, and efficient applications that transform business processes, enhance decision-making, and deliver measurable value. This role requires a good understanding of architecture, AI/ML frameworks, Cloud, and modern automation tools, combined with a passion for solving complex problems.
Experience/Knowledge:
- 10 years of experience in data architecture, software engineering, or a related role.
- Proven expertise in designing and deploying applications with generative AI technologies (e.g., TensorFlow, PyTorch, Hugging Face Transformers).
- Hands-on experience with intelligent document processing.
- Strong background in automation technologies.
- Experience with cloud platforms (AWS)
- Experience with Streaming Data Integration / Event-Driven Integration
- Expertise in processing large amounts of diverse data types, including structured, semi-structured, and unstructured data.
- Familiarity with Data Federation & Data Virtualization
- Hands-on experience in ensuring data privacy, security (e.g., encryption, access control), and compliance with regulations, particularly when dealing with potentially sensitive data used in AI models.
- Experience in architecting and optimizing Retrieval-Augmented Generation (RAG) pipelines, including, integrating vector databases and complex data retrieval logic.
Technical Skills:
- Proficiency in Python, Databricks, NoSQL, Vector & Graph databases
- Familiarity with Python
- Familiarity with AI/ML frameworks, APIs, and model deployment
- Familiarity with data visualization tools (e.g., Tableau, Power BI)
- Knowledge of data pipeline tools (e.g., Apache Kafka, Airflow).
- Understanding of DevOps practices as in relates to Data (e.g. Data Pipelines)
- Proficiency in analyzing data requirements for new software and applications.