Demo

VDOT Cloud ETL Developer

Cyber Resource Provider LLC
Richmond, VA Full Time
POSTED ON 1/26/2025
AVAILABLE BEFORE 3/24/2025

Job Details

Engagement Type

Contract

Short Description

VDOT Data Analyst / SR Cloud ETL Developer.


Complete Description

*Local Richmond, VA candidates ONLY required due to onsite requirement

**This position requires onsite 3 days a week with 2 remote

**Contractor will be responsible for purchasing parking through VDOT's Parking Management Office or procuring their own parking





Job Description: SR ETL Developer

The Virginia Department of Transportation (VDOT)Information Technology Division (ITD) is seeking a Master Data Analyst with demonstrated experience in data analytics to work as a key member of Enterprise Data Asset team. This analyst will support teams working in Agile (Sprint) to analyze datasets to be made available in a cloud-based data management platform that will support the agency to produce master data with data governance.

Responsibilities include analyzing source systems which contain a spatial component for candidate datasets; documenting business processes and data lifecycle; developing data requirements, user stories and acceptance criteria; and testing strategies. Develop ETL to extract business data and spatial data and load it into a data warehousing environment. Design and test the performance of the system. Consult with various teams to understand the company's data storage needs and develop data warehousing options. Deep knowledge of coding languages, such as python, Java, XML, and SQL. Well-versed in warehousing architecture techniques such as MOLAP, ROLAP, ODS, DM, and EDW.

VDOT is a fast-paced organization with very high standards for work quality and efficiency. This position is expected to handle multiple projects, and remain flexible and productive, despite changing priorities and processes. Ongoing improvement and efficiency are a part of our culture, and each team member is expected to proactively contribute to process improvements.



Responsibilities:



Work with the Project team members and business stakeholders to understand business processes and pain points

Develop expertise in source system datasets and data lifecycle

Profile source data which may contain a spatial component; review source data and compare content and structure to dataset requirements; identify conflicts and determine recommendations for resolution

Conduct entity resolution to identify matching and merging and semantic conflicts

Elicit, record, and manage metadata

Diagram current processes and proposed modifications using process flows, context diagrams and data flow diagrams

Decompose requirements into Epics and Features and create clear and concise user stories that are easy to understand and implement by technical staff

Utilize progressive elaboration; map stories to data models and architectures to be used by internal staff to facilitate master data management

Identify and group related user stories into themes, document dependencies and associated business processes

Discover and document requirements and user stories with a focus on improving both business and technical processing

Assist Product Owner in maintaining the product backlog

Create conceptual prototypes and mock-ups

Collaborate with staff, vendors, consultants, and contractors as they are engaged on tasks to formulate, detail and test potential and implemented solutions

Perform Quality Analyst functions such as defining test objectives, test plans and test cases, and executing test cases

Coordinate and Facilitate User Acceptance Testing with Business and ensure Project Managers/Scrum Masters are informed of the progress

Designs and develops systems for the maintenance of the Data Asset Program(Data Hub), ETL processes, ETL processes for spatial data, and business intelligence.

Develop a new data engineering process that leverage a new cloud architecture and will extend or migrate our existing data pipelines to this architecture as needed.

Design and supports the DW database and table schemas for new and existent data sources for the data hub and warehouse. Design and development of Data Marts.

Work closely with data analysts, data scientists, and other data consumers within the business in an attempt together and populate data hub and data warehouse table structure, which is optimized for reporting.

The Data developers partners with Data modeler and Data architect in an attempt to refine the business's data requirements, which must be met for building and maintaining Data Assets.





Qualifications



Required:

The candidate must have a minimum of 10 years of experience delivering business data analysis artifacts

5 years of experience as an Agile Business Analyst; strong understanding of Scrum concepts and methodology

Experience organizing and maintaining Product and Sprint backlogs

Experience translating client and product strategy requirements into dataset requirements and user stories

Proficient with defining acceptance criteria and managing acceptance process

Exceptional experience writing complex sql queries for Sql Server and Oracle

Experience with Azure Databricks, Azure data factory, snowflake

Experience with ESRI ArcGIS

Experience with enterprise data management

Expertise with Microsoft Office products (Word, Excel, Access, Outlook, Visio, PowerPoint, Project Server)

Experience with reporting systems operational data stores, data warehouses, data lakes, data marts

The candidate must have exceptional written and oral communications skills and have the proven ability to work well with a diverse set of peers and customers


Preferred Skills:

Advanced understanding of data integrations.

Strong knowledge of database architectures

Strong analytical and problem solving skills

Ability to build strong relationships both internally and externally

Ability to negotiate and resolve conflicts

Ability to effectively prioritize and handle multiple tasks and projects

Strong written and verbal communication skills

Desire to learn, innovate and evolve technology

Computer Skills/MS Office/Software:

Excellent computer skills and be highly proficient in the use of MS Word, PowerPoint, MS Excel, MS Project, MS Visio, and MS Team Foundation Server, which will all be necessary in the creation of visually and verbally engaging ETL, data designs and tables as well as the communication of documentation and reporting.

Deep passion for data analytics technologies as well as analytical and dimensional modeling. The candidate must be extensively familiar with ETL(Extraction, Transformation & Load), data warehousing, and business intelligence tools such as business objects, PowerBI and Tableau.

The candidate must also have vast knowledge of database design and modeling in the context of data warehousing.

Experience with key data warehousing architectures including Kimball and Inmon, and has a broad experience designing solutions using a broad set of data stores (e.g., HDFS, Azure Data Lake Store, Azure Blob Storage, Azure SQL Data Warehouse, Azure Cosmos DB

Technologies Required:

Data Factory v2,Data Lake Store, Data Lake Analytics, Azure Analysis Services, AZURE Synapse

IBM Datastage, Erwin, SQL Server (SSIS, SSRS, SSAS), ORACLE, T-SQL, Azure SQL Database, Azure SQL Datawarehouse.

Operating System Environments (Windows, Unix, etc.).

Scripting experience with Windows and/or Python, Linux Shell scripting

Required/Desired Skills

Skill Required/Desired Amount of Experience
Designs and develops systems for the maintenance of the Data Asset Program, ETL processes, and business intelligence. Required 10 Years
Design and supports the DW database and table schemas for new and existent data sources for the data hub and warehouse. Design and development of Data Required 10 Years
Work closely with data analysts, data scientists, and other data consumers within the business in an attempt to gather and populate data hub and data Required 10 Years
Advanced understanding of data integrations. Strong knowledge of database architectures, strong understanding of ingesting spatial data Required 10 Years
Ability to negotiate and resolve conflicts, Ability to effectively prioritize and handle multiple tasks and projects Required 10 Years
Excellent computer skills and be highly proficient in the use of Ms Word, PowerPoint, Ms Excel, MS Project, MS Visio, and MS Team Foundation Server Required 10 Years
Experience with key data warehousing architectures including Kimball and Inmon, and has a broad experience designing solutions using a broad set of da Required 10 Years
expertise in Data Factory v2,Data Lake Store, Data Lake Analytics, Azure Analysis Services, Azure Synapse Required 10 Years
IBM Datastage, Erwin, SQL Server (SSIS, SSRS, SSAS), ORACLE, T-SQL, Azure SQL Database, Azure SQL Datawarehouse. Required 10 Years
Operating System Environments (Windows, Unix, etc.). Scripting experience with Windows and/or Python, Linux Shell scripting Required 10 Years
Experience in AZURE Cloud engineering Required 10 Years
Questions

Description
Question 1 Commonwealth of Virginia security policies prohibit the use of offshore IT contractors. Do you attest to the fact that your candidate will physically reside within the US for the duration of the assignment?
Question 2 Please list candidate's email address.
Question 3 In what city and state does your candidate PERMANENTLY reside?
Question 4 Is the candidate available to work in Richmond, VA at least 3 days/wk? This is REQUIRED.
Question 5 How soon after an offer can your candidate start?
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.

If your compensation planning software is too rigid to deploy winning incentive strategies, it’s time to find an adaptable solution. Compensation Planning
Enhance your organization's compensation strategy with salary data sets that HR and team managers can use to pay your staff right. Surveys & Data Sets

What is the career path for a VDOT Cloud ETL Developer?

Sign up to receive alerts about other jobs on the VDOT Cloud ETL Developer career path by checking the boxes next to the positions that interest you.
Income Estimation: 
$116,726 - $151,072
Income Estimation: 
$147,901 - $186,323
Income Estimation: 
$88,359 - $121,264
Income Estimation: 
$93,716 - $124,745
Income Estimation: 
$118,976 - $146,289
Income Estimation: 
$112,672 - $149,113
Income Estimation: 
$98,475 - $115,895
Income Estimation: 
$89,551 - $118,439
Income Estimation: 
$116,726 - $151,072
Income Estimation: 
$124,724 - $161,246
Income Estimation: 
$112,672 - $149,113
Income Estimation: 
$115,719 - $153,093
Income Estimation: 
$137,343 - $165,639
Income Estimation: 
$135,811 - $184,429
Income Estimation: 
$120,390 - $162,969
View Core, Job Family, and Industry Job Skills and Competency Data for more than 15,000 Job Titles Skills Library

Job openings at Cyber Resource Provider LLC

Cyber Resource Provider LLC
Hired Organization Address Richmond, VA Full Time
Job Details Engagement Type Contract Short Description VITA is seeking a Business Automation/AI & LCAP Team Lead This po...
Cyber Resource Provider LLC
Hired Organization Address Richmond, VA Full Time
Job Details Engagement Type Contract Short Description VITA - Technical PM - IN PERSON IVs ONLY! ON SITE REQUIRED: 4 day...
Cyber Resource Provider LLC
Hired Organization Address Richmond, VA Full Time
Job Details Engagement Type Contract Short Description VDOT Lead Azure with Dot Net and Angular Experience local candida...
Cyber Resource Provider LLC
Hired Organization Address Richmond, VA Full Time
Job Details Engagement Type Contract Short Description DBHDS - MS Power Automate RPA P/A 100% Remote / Must be able to p...

Not the job you're looking for? Here are some other VDOT Cloud ETL Developer jobs in the Richmond, VA area that may be a better fit.

Cloud ETL Developer

Careers Integrated Resources Inc, Richmond, VA

Cloud ETL Developer

LingaTech, Richmond, VA

AI Assistant is available now!

Feel free to start your new journey!