Expoint - all jobs in one place

The point where experts and best companies meet

Limitless High-tech career opportunities - Expoint

Microsoft Senior Data Engineer 
United States, Washington 
194410970

25.06.2024

Required/Minimum Qualifications

  • Bachelor's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 4+ years experience in business analytics, data science, software development, data modeling or data engineering
    • OR Master's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 3+ year(s) experience in business analytics, data science, software development, or data engineering
    • OR equivalent experience
  • 4+ years of experience in data visualization using PowerBI.
  • 2+ years of experience in developing new products for companies or individuals to use, providing maintenance to current products to enhance use.

Other Requirements:

  • Ability to meet Microsoft, customer and/or government security screening requirements are required for this role. These requirements include, but are not limited to the following specialized security screenings:Microsoft Cloud Background Check: This position will be required to pass the Microsoft Cloud Background Check upon hire/transfer and every two years thereafter.


Additional or Preferred Qualifications

  • Bachelor's Degree in Computer Science, Math, Software Engineering, Computer Engineering , or related field AND 8+ years experience in business analytics, data science, software development, data modeling or data engineering
    • OR Master's Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 6+ years of business analytics, data science, software development, data modeling or data engineering work experience
    • OR equivalent experience.
  • 6+ years of experience in data visualization using PowerBI.

Certain roles may be eligible for benefits and other compensation. Find additional benefits and pay information here:

Responsibilities
  • Collaborates with appropriate stakeholders across teams and escalates concerns around data requirements by assessing and conducting feature estimation. Informs clients on feasibility of data needs and suggests transformations or strategies to acquire data if requirements cannot be met. Negotiates agreements with partners and system owners to align on project delivery, data ownership between both parties, and the shape and cadence of data extraction for one or more features. Proposes new data metrics or measures to assess data across varied service lines.
  • Leads the design of a data model that is appropriate for the project and prepares design specification documents to model the flow and storage of data for a data pipeline. Designs assigned components of the data model for a functional area of a project. Partners with stakeholders (e.g., Data Science Specialists) to make iterative improvements to design specifications, data visualizations, data models, or data schemas. Considers tradeoffs between analytical requirements with compute/storage consumption for data and anticipates cost that could be influenced by the cadence of data extraction, transformation, and loading into moderately complex data products or datasets in cloud and local environments. Demonstrates an advanced understanding of costs associated with data that are used to assess the total cost of ownership (TOC).
  • Identifies data sources and builds code to extract raw data from identified upstream sources using query languages while ensuring accuracy, validity, and reliability of the data across several pipeline components. Contributes to the code review process by providing feedback and suggestions for implementation.
  • Leverages reduction techniques, and aggregation approaches to validate the quality of extracted data across a data pipeline, consistent with the Service Level Agreement. Offers feedback on methods and tools used to track and maintain data source control and versioning. Applies deep knowledge of data to validate that the correct data is ingested and that the data is applied accurately across the pipeline.
  • Plans and creates efficient techniques and operations (e.g., inserting, aggregating, joining) to transform raw data into a form that is compatible with downstream data sources, databases, and visualizations. Independently uses software, query languages, and computing tools to transform raw data across end-to-end pipelines. Evaluates data to ensure data quality and completeness using queries, data wrangling, and statistical techniques. Merges data into distributed systems, products, or tools for further processing.
  • Writes code to implement performance monitoring protocols across data pipelines. Builds visualizations and smart aggregations. Develops and updates troubleshooting guides (TSGs) and operating procedures for reviewing, addressing, and/or fixing advanced problems/anomalies. Supports and monitors platforms.
  • Performs root cause analysis in response to detected problems/anomalies to identify the reason for alerts or customer escalations and implement solutions that minimize points of failure. Implements and monitors self-healing processes across multiple product features to prevent issues from recurring in the future and retain data quality and optimal performance (e.g., latency, cost) throughout the data lifecycle.
  • Documents the problem and solutions through postmortem reports and shares insights with team and the customer. Provides data-based insights into the health of data products owned by the team according to service level agreements (SLAs) across multiple features.
  • Anticipates the need for data governance and designs data modeling and data handling procedures, with direct support and partnership with Corporate, External, and Legal Affairs (CELA), to ensure compliance with applicable laws and policies across all aspects of the data pipeline. Tags data based on categorization (e.g., personally identifiable information [PII], pseudo-anonymized, financial). Documents data type, classifications, and lineage to ensure traceability. Governs accessibility of data within assigned data pipelines. Provides guidance on contributions to the data glossary to document the origin, usage, and format of data for each program.
  • Participate in 24/7 on-call rotation to support 1st party customers.
  • Embody our