GCP Data Analyst - BigQuery (USA) at Rackspace
Texas, Texas, USA -
Full Time


Start Date

Immediate

Expiry Date

09 Oct, 25

Salary

245520.0

Posted On

10 Jul, 25

Experience

5 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Python, Airflow, Sql, Data Science, Materialized Views, Computer Science, Time Series Analysis, Analytical Skills, Google Cloud Platform

Industry

Information Technology/IT

Description

QUALIFICATIONS:-



    • Bachelor’s degree in Computer Science, Data Science, Engineering, or a related field.

    • 5+ years of experience in data analyst or analytics engineering roles with strong BigQuery, SQL, and Python skills.
    • 5+ years of experience building and operating solutions on Google Cloud Platform (GCP).
    • Strong ability to write and optimize SQL queries to validate data, analyze trends, and detect inconsistencies.
    • Proficient in Python, including use of data frames and common analytical libraries.
    • Experience with advanced BigQuery features such as authorized views, materialized views, UDFs, partitions, and time series analysis.
    • Strong analytical skills and experience validating data across systems during migrations and ongoing operations.
    • Basic ability to read and understand Java or Scala code to support engineering collaboration.
    • Familiarity with Airflow (Cloud Composer) to interpret and trace data pipeline workflows.
    Responsibilities

    ABOUT THE ROLE

    We’re seeking a GCP Data Analyst with deep expertise in BigQuery, strong SQL and Python skills, and a sharp analytical mindset to support both data validation initiatives and ongoing analytics work. This role is ideal for someone who can navigate large datasets, build robust queries, and identify inconsistencies with precision and insight.
    The analyst will work across a variety of data workflows, from validating metrics during system migrations to supporting day-to-day data analysis and reporting needs. You’ll leverage advanced BigQuery features—such as authorized views, materialized views, UDFs, partitioning strategies, and time series analysis—to ensure data integrity and surface meaningful insights. Comfort working in Python with data frames and relevant packages is also essential, particularly for tasks involving data manipulation, anomaly detection, or prototyping workflows.
    A solid understanding of data engineering fundamentals and GCP infrastructure is important, as is the ability to read and interpret code in Java or Scala when collaborating with engineering teammates. Familiarity with Airflow (Composer) will help you understand orchestration logic, though this won’t be a core responsibility. Experience with BigQuery ML, anomaly detection frameworks, or Vertex AI is a plus.

    KEY RESPONSIBILITIES:-



      • Write, optimize, and execute complex SQL queries in BigQuery to validate data accuracy, identify inconsistencies, and support analytics and reporting.

      • Analyze large datasets to assess data quality, compare trends across systems, and surface anomalies or unexpected behaviors.
      • Utilize advanced BigQuery features such as authorized views, materialized views, UDFs, partitioned tables, and joins to support scalable, high-performance analysis.
      • Use Python (including data frames and relevant libraries) for exploratory analysis, data manipulation, and supporting validation workflows.
      • Support time series analysis and, where applicable, anomaly detection using SQL or Python-based approaches.
      • Assist with load/transform validation to ensure reliability and accuracy in data pipelines.
      • Collaborate with engineering teams to understand data pipelines, with basic ability to read and interpret Java or Scala code when needed.
      • Perform side-by-side comparisons of data across systems to ensure consistency during and after migrations.
      • Maintain basic familiarity with orchestration tools such as Airflow (Composer) to follow pipeline logic and collaborate effectively with engineering.
      • Work within the GCP environment, leveraging cloud tools and services to support analysis, troubleshoot issues, and navigate cloud-based workflows.
      • Clearly communicate analytical findings and data quality issues to cross-functional stakeholders to support decision-making.
      Loading...