
Onsite: Principal Data Architect
Hybrid in Houston 4x onsite (it is mandatory to be onsite in downtown Houston Monday- Thursday. Friday is remote option).
List of programs, software, tools, or systems this resource will use:
(We are aware someone may not have every item listed below, however the more items on this list your candidate has-- the better the fit-- HUGE plus!)
Database: Snowflake / aws postrges/Oracle/, sql server/ aurora
ETL tool: DBT/Matillion/Informatica, AWS glue
Language: Python /javaScript /SQl/NoSQL,spark, pyspark
Event Streaming platforms: Amazon Kinesis/ Apache Kafka/ Apache Spark / Qlik
AWS: S3 / SNS/SQS/ Lambda / glue/ec2/iam/cloud formation**
Data Modelling: Erwin
**Please note all items listed by "AWS" is a MUST have (i.e. S3 / SNS/SQS/ Lambda / glue/ec2/iam/cloud formation)!**
Job Description:
The Principal Data Architect will lead the design, implementation, and optimization of enterprise-wide data architectures ensuring they align with business objectives and technological advances. Working closely with cross-functional teams, data engineers, analysts, and business stakeholders, the Principal Data Architect will analyze, design and implement scalable, reliable and efficient data architectures and solutions to meet business requirements and translate business needs into appropriate data deliverables.
This role requires a deep understanding of data architecture, data modeling, and data management practices, so the ideal candidate will have a strong background in data engineering, data warehousing, cloud technologies, along with staying current with emerging technologies and industry best practices in data management and architecture.
Responsibilities:
Collaborate with business stakeholders to understand data needs and translate them into technical solutions
Requirements gathering, defining project scope, and structuring it into actionable phases
Build detailed architecture diagrams such as data models, data flow diagrams, data mappings, and data dictionaries for the EDW
Consult, architect and design data integrations and pipelines for data migration projects
Integrate disparate data models into coherent enterprise data models.
Develop data strategies and communicate future strategy and vision
Demonstrate advanced data strategy and architecture through execution of proofs of concept
Develop an understanding of data processes and data dependencies; create data dictionaries and business glossaries to document data lineages, data definitions and metadata for all business-critical data domains.
Promote use of modern technologies and frameworks
Develop and own strategic roadmaps and ensure they remain in sync with business strategy goals
Assist in the development, execution, and supervision of plans, policies, programs, and practices
Assist in issue resolution, measuring and tracking data quality issues
Creating and maintaining data management frameworks, data management SOPs, standards and guidelines, and control processes
Perform other duties as assigned
Required:
5+ years of experience with information modeling, information visualization, data modeling, data management technologies, data integration, data sourcing, data analytics, data warehousing, metadata, semantics and ontologies.
Experience architecting data products in Snowflake or other cloud data platforms
3+ years experience developing AWS data pipelines (Glue, Kinesis, S3, etc.)
5+ years in ELT or data warehouse data loading design and development. Experience with DBT a major plus.
Familiarity with modern data management concepts including Cloud-based Data warehousing, Big Data stack, NoSQL, in-memory DBMS and data caching; graph databases.
Familiarity and experience in data protection and information security.
Strong understanding of large scale system, information and technology architectures, including application integration patterns, messaging, service-oriented architecture, information models, and data lineage.
Hands-on experience with data and process modeling tools (e.g., ERWin,ER Studio, Lucidchart), and working knowledge of programming languages such as: Python, etc.
Demonstrated ability to work individually and as a part of the team in a collaborative manner.
Experience with Agile/Scrum and SDLC.
Knowledge of data governance practices, business and technology issues related to management of enterprise information assets and approaches related to data protection.
Education and Experience
Bachelor's degree (or foreign equivalent) in Computer Science, Data Science, or a related field.