Job Location: REMOTE with Vocational Onsite Visits at Waterbury, VT (Travel Expenses will be reimbursed)
Duration: 3+ Years Contract
Job Description:
Background:
Client is seeking to improve statewide law enforcement data access. The purpose is to design and implement a state-controlled system to access de-identified, aggregated law enforcement and related data currently housed in a record management system.
Client in collaboration with the Agency of Digital (ADS) Services, is seeking to procure Amazon Web Service (AWS) professional services to work with the ADS Technical Lead to build out the Public Services Lakehouse environment.
Existing Technology Environment
Client currently uses a Computer Aided Dispatch Records Management System (CAD RMS) system that is currently running in a MySQL instance that will be used in the first phase of the Data Lake build. The supplemental data ranges from SQL server instances and flat file sources that are currently housed internally either in SharePoint or internal file stores.
Requirements:
Design and implement the Clients' Data Lake in the AWS Environment
Store Data in AWS CJIS compliant environment.
Implement and design the lake house technologies with the IT Tech Lead
assigned to the project.
Design and implement using latest AWS Lake House standards.
Data Security Layer
Implement and design security Identity and Access Management roles and processes.
Design and Create security IAM templates.
Data Ingestion Layer
Ingestion Design for variety of sources
Operational Database Sources
MySQL, SQL Server
SaaS Applications
File Shares (SharePoint and one drive)
Stream Data Sources
System Templates of ingestion processes
Data Storage Layer
AWS S3 CJIS compliant S3 buckets
AWS Redshift Infrastructure
Power BI Connector process
Data Processing Layer
Data Extract, Load and Transform (ELT) process for loading from source to S3 and Transforming data from S3 to Redshift for reporting and analytics.
Templates for creating ELT processes for future processes.
Data Catalog Layer
Design and implement solution to solve data schema drift in AWS Glue for use with reporting and analytical needs.
Design and build crawlers for schema and build catalog that stores schema information.
Meta data store in catalog for consumption in data warehouse
Create Catalog crawler templates for data sources.
Create Templates for expansion of the data lake for future agencies.
Base data warehouse implemented environment for reporting/dashboard use.
Template designs for future data Lakehouse implementations.
Lake House design needs to be able to be used by Power BI
Design each layer for scaling based on usage.
Professional Service Requirements:
AWS Lakehouse Certification
AWS Design and Implementation Lake House Technologies
AWS IAM
AWS Lakehouse Technology
AWS Glue Knowledge
Data Catalog and Crawler
AWS Redshift
AWS Athena
CJIS certified data storage
AWS CJIS Data location
CJIS Security Background Check (See Additional Attachments / Assurances)