Data Integrity Engineer
Amazon.com, Inc.
Seattle, United States of America
7 days ago
Role details
Contract type
Temporary contract Employment type
Full-time (> 32 hours) Working hours
Regular working hours Languages
English Experience level
Senior Compensation
$ 160KJob location
Seattle, United States of America
Tech stack
Microsoft Excel
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Data analysis
Spreadsheets
Cloud Computing
Data as a Services
Data Auditing
Data Centers
Data Validation
Data Cleansing
Data Governance
Data Integrity
ETL
Data Security
Data Visualization
Database Queries
Document-Oriented Databases
Excel Formulas
Machine Learning
Power BI
Salesforce
Tableau
Smartsheet
Data Processing
Scripting (Bash/Python/Go/Ruby)
AWS Lambda
Pandas
Data Lineage
Amazon Web Services (AWS)
Data Analytics
Data Management
Cloudwatch
Data Pipelines
Redshift
Web Api
Job description
- Perform data management tasks, ensuring accuracy, consistency, and integrity of data across various systems and platforms
- Conduct routine data quality checks, identify and resolve data discrepancies, anomalies, and out-of-range values
- Implement and maintain data governance policies and procedures, collaborating with team members to document data processes and maintain data lineage
- Provide support in the creation of data reports, dashboards, and visualizations, ensuring alignment with data visualization best practices
- Participate in data audits and assist in the preparation of data-related documentation
- Communicate with stakeholders to understand data requirements and provide updates on data quality improvements
- Maintain a strong understanding of data security, privacy, and compliance requirements
- Leverage Python scripting and SQL querying skills to automate data cleansing and correction processes
- Trace and isolate sources of data issues, including cell links, automations, scripts, and other dependencies
- Work with the central data analytics team to identify and implement necessary modifications to data pipelines and configurations
- Assess when to apply manual intervention versus automation for efficient data correction and enhancement
- Perform other ad-hoc data-related tasks as assigned by the team lead or manager, * Team takes requests for teams who want to change some element from data centers, vet those from a technical perspective, ensure change makes sense from business perspective, integrated technically across disciples in the design, push it out to active data center design portfolio that they are managing.
- Work in APAC, EMEA, ETC
- Basis of design for data center, site adapt it at.
- TIDUS machine learning, liquid cooling, etc. The number of changes has overtaxed the team.
Key Projects:
- Going to assign specific global programmatic design changes depending on when they start.
- Some will also depend on skillset being brought to the table
Day to Day:
- Data Reporting and Visualization (40% of time)
- Create data reports, dashboards, and visualizations based on stakeholder/Team/Project requirements using tools like Quicksight, Tableau, Smartsheet, Excel Sheet.
- Ensure alignment of data visualizations with the project and teams goal.
- Provide support and guidance to stakeholders/cross functional teams on interpreting data insights/queries.
- Data Automation and Scripting (20% of time)
- Leverage Python scripting and SQL querying skills to automate data cleansing and correction processes
- Develop scripts and workflows to improve data processing efficiency
- Assess opportunities to apply automation versus manual intervention
- Data Governance and Documentation (15% of time)
- Maintain and update data governance policies and procedures
- Document data processes and data lineage
- Collaborate with team members to ensure consistent data management practices
- Data Quality Assurance (15% of time)
- Conduct routine data quality checks across various systems and platforms like Salesforce, Cloud Forge, REV and internal Project Management tools like ( SIM, Smartsheet)
- Identify and resolve data discrepancies, anomalies, and out-of-range values
- Implement data validation rules and data cleansing processes to handle the data integrity, data availability issues.
- Stakeholder Collaboration and Troubleshooting (10% of time)
- Communicate with stakeholders to understand data requirements and provide updates
- Troubleshoot and isolate sources of data issues, including dependencies
- Participate in data audits and assist in the preparation of data-related documentation
Role interesting:
- If you want to solve hard problems with talented people, this is a great team to join.
- Team is very uniquely positioned within design engineering
- Team interacts very broadly with multiple stakeholder groups
- Real time innovating with infrastructure projects to keep AWS at the forefront
- Unique, challenging, fast paced but fun.
Leadership principle:
- Customer obsession
- Invent and simplify
- Learn and be curious
- Dive deep
- Insist on higher standards
Requirements
- Bachelor's degree or equivalent
- 3+ years of experience in a data management or data analytics related role
- Proficient in using spreadsheet software (e.g., Microsoft Excel, Smartsheet) and writing advanced functions, automations, and data meshes
- Experience with relational databases and SQL querying, including the use of window functions and complex joins
- 3+ years of experience with data visualization platforms, including creating calculated measures and working with denormalized data (preferred: Amazon Quicksight)
Preferred Qualifications:
- Python scripting experience, including libraries such as Pandas (preferred: experience with API calls and the Smartsheet SDK)
- Understanding of AWS data services such as AWS Glue, Amazon Redshift, or AWS Lambda for data processing and ETL
- Ability to independently troubleshoot and logically infer causes of data issues by isolating variables and testing
- Experience using IDEs such as DataGrip, VSCode
- Knowledge of effective data visualization design principles, * 3+ years of experience in a data management or data analytics related role
- Proficient in using spreadsheet software (e.g., Microsoft Excel, Smartsheet) and writing advanced functions, automations, and data meshes
- Experience with relational databases and SQL querying, including the use of window functions and complex joins
- 3+ years of experience with data visualization platforms, including creating calculated measures and working with denormalized data (preferred: Amazon Quicksight)
Degree/certs:
- Bachelor's degree or equivalent
Resume stand out:
- Advanced Data Visualization and Dashboard Design
- Cloud-based BI Platform Expertise (Amazon Quicksight, Amazon S3, Amazon Redshift, AWS Glue, AWS Lambda, Amazon CloudWatch)
- Expertise in ETL Process and Data Management and Analysis
- Former employee experience
Top must-have hard skills:
- Proficiency in using spreadsheet software (e.g., Microsoft Excel, Smartsheet) and writing advanced functions (e.g., Pivot, Index), automations, and data meshes
- Experience with data visualization platforms, including creating calculated measures and working with denormalized data (preferred: Amazon Quicksight, Tableau)
- Understanding a of Python scripting (preferred: experience with API calls and the Smartsheet SDK)
- Strong experience with relational databases and SQL querying, including the use of window functions and complex joins.
About the company
TekWissen is a global workforce management provider headquartered in Ann Arbor, Michigan that offers strategic talent solutions to our clients world-wide. Our client is a global technology leader that is transforming the way people shop, consume content, and interact with digital services. With a strong presence across e-commerce, cloud computing, digital streaming, and artificial intelligence, the company is known for its customer-centric approach, innovation-driven culture, and large-scale global impact. It continues to invest in building products and services that improve everyday life while supporting businesses, communities, and economies worldwide., © 2026 Careerjet All rights reserved