I
Data Engineer Google Cloud Platform BigQuery Specialist - TX/FL/NJ - Hybrid
Salary undisclosed
Checking job availability...
Original
Simplified
Technical Skills
- Expertise in SQL optimization for BigQuery and relational databases.
- Proficiency in ETL/ELT development using Cloud Composer (Apache Airflow), Dataflow (Apache Beam), and BigQuery Transfer Service.
- Experience with Python, Shell scripting, or Java for automation and data transformation. Strong knowledge of data partitioning, clustering, and materialized views in BigQuery.
- Familiarity with data orchestration tools like Apache Airflow, Cloud Composer, or dbt (Data Build Tool).
- Hands-on experience with Google Cloud Platform services such as Cloud Storage, Pub/Sub, Cloud Functions, and IAM. Understanding of data warehouse design, star schema modeling, and data normalization/denormalization.
- Experience with data quality frameworks and best practices for data validation and monitoring.
- Working knowledge of version control tools like GitHub, Bitbucket, or GitLab.
Key Responsibilities
- Data Pipeline Development and Optimization Design and build scalable ETL/ELT data pipelines to transfer data from relational databases (SQL Server, PostgreSQL, MySQL, Oracle, etc.) to Google Cloud Platform BigQuery. Implement best practices for efficient data ingestion using tools such as Apache Airflow, Dataflow, Cloud Composer, and BigQuery Data Transfer Service. Optimize SQL queries for high-performance data movement and processing in BigQuery.
- Performance Tuning and Optimization Write and optimize complex SQL queries to ensure fast and cost-efficient data transformations. Monitor and fine-tune BigQuery performance using clustering, partitioning, and query execution plans. Implement best practices for cost optimization on BigQuery, reducing query costs and improving efficiency.
- Data Architecture and Schema Design Design and implement schema models in BigQuery, including denormalization techniques, partitioning, and clustering strategies. Collaborate with Data Architects to ensure scalability, reliability, and security of data architectures. Work with business teams to define data models that align with analytics and reporting needs.
- Data Integration and Transformation Develop real-time and batch data processing solutions using Cloud Dataflow, Pub/Sub, and Cloud Functions. Implement data cleansing, enrichment, and transformation using SQL, Python, or Apache Beam. Integrate data from multiple sources (structured, semi-structured, unstructured) into BigQuery.
- Cloud Infrastructure & Security Manage Google Cloud Platform IAM roles and permissions to ensure secure access to data. Implement data governance and compliance policies for handling sensitive data in BigQuery. Utilize Cloud Logging and Monitoring (Stackdriver, Cloud Logging, Cloud Monitoring) for troubleshooting and performance monitoring.
- Collaboration in a Multi-Vendor, Dynamic Work Environment Work in a fast-paced environment with multiple vendors, cloud service providers, and technology partners. Collaborate with data analysts, data scientists, and business intelligence teams to understand requirements and provide efficient data solutions. Coordinate with cross-functional teams across different geographical locations and time zones. Required
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
Report this job Technical Skills
- Expertise in SQL optimization for BigQuery and relational databases.
- Proficiency in ETL/ELT development using Cloud Composer (Apache Airflow), Dataflow (Apache Beam), and BigQuery Transfer Service.
- Experience with Python, Shell scripting, or Java for automation and data transformation. Strong knowledge of data partitioning, clustering, and materialized views in BigQuery.
- Familiarity with data orchestration tools like Apache Airflow, Cloud Composer, or dbt (Data Build Tool).
- Hands-on experience with Google Cloud Platform services such as Cloud Storage, Pub/Sub, Cloud Functions, and IAM. Understanding of data warehouse design, star schema modeling, and data normalization/denormalization.
- Experience with data quality frameworks and best practices for data validation and monitoring.
- Working knowledge of version control tools like GitHub, Bitbucket, or GitLab.
Key Responsibilities
- Data Pipeline Development and Optimization Design and build scalable ETL/ELT data pipelines to transfer data from relational databases (SQL Server, PostgreSQL, MySQL, Oracle, etc.) to Google Cloud Platform BigQuery. Implement best practices for efficient data ingestion using tools such as Apache Airflow, Dataflow, Cloud Composer, and BigQuery Data Transfer Service. Optimize SQL queries for high-performance data movement and processing in BigQuery.
- Performance Tuning and Optimization Write and optimize complex SQL queries to ensure fast and cost-efficient data transformations. Monitor and fine-tune BigQuery performance using clustering, partitioning, and query execution plans. Implement best practices for cost optimization on BigQuery, reducing query costs and improving efficiency.
- Data Architecture and Schema Design Design and implement schema models in BigQuery, including denormalization techniques, partitioning, and clustering strategies. Collaborate with Data Architects to ensure scalability, reliability, and security of data architectures. Work with business teams to define data models that align with analytics and reporting needs.
- Data Integration and Transformation Develop real-time and batch data processing solutions using Cloud Dataflow, Pub/Sub, and Cloud Functions. Implement data cleansing, enrichment, and transformation using SQL, Python, or Apache Beam. Integrate data from multiple sources (structured, semi-structured, unstructured) into BigQuery.
- Cloud Infrastructure & Security Manage Google Cloud Platform IAM roles and permissions to ensure secure access to data. Implement data governance and compliance policies for handling sensitive data in BigQuery. Utilize Cloud Logging and Monitoring (Stackdriver, Cloud Logging, Cloud Monitoring) for troubleshooting and performance monitoring.
- Collaboration in a Multi-Vendor, Dynamic Work Environment Work in a fast-paced environment with multiple vendors, cloud service providers, and technology partners. Collaborate with data analysts, data scientists, and business intelligence teams to understand requirements and provide efficient data solutions. Coordinate with cross-functional teams across different geographical locations and time zones. Required
Employers have access to artificial intelligence language tools (“AI”) that help generate and enhance job descriptions and AI may have been used to create this description. The position description has been reviewed for accuracy and Dice believes it to correctly reflect the job opportunity.
Report this job