<Back to Search
Data Engineer
Buffalo, NYMarch 27th, 2026
Job Title: Data/Software EngineerLocation: RemoteIndustry: Pharmaceutical***NO C2C***Job Description: Theoris Services is assisting our client in their search for a Data/Software Engineer to add to their growing team. Our client is seeking someone with data visualization experience and software engineering (create reusable libraries, best practices, troubleshooting).Responsibilities:Data Pipeline & Backend Development Design, build, and optimize scalable data pipelines and ETL/ELT processes to integrate and harmonize scientific data (compounds, assays, experiments) from 30+ heterogeneous sources.Implement and maintain lakehouse architectures on AWS (S3, Glue, Athena, Iceberg) to support multibillion-record datasets.Develop federated query capabilities using Trino (or similar distributed engines) for unified access across platforms like PostgreSQL, Snowflake, and others.Build robust backend services, RESTful APIs, and data services using Python (FastAPI, Flask preferred) to enable seamless data flow and integration with scientific tools (e.g., Benchling, computational chemistry systems, AI/ML endpoints).Performance Optimization & Troubleshooting Optimize query and database performance for complex analytical workloads across PostgreSQL, Iceberg, Trino, and other platforms.Implement caching, indexing, and query tuning techniques to improve response times and scalability as data volumes and user bases grow.Apply reverse engineering and advanced troubleshooting skills to debug complex data issues, pipeline bottlenecks, application failures, and performance problems proactively.Monitor systems, identify root causes, and implement fixes for data and application reliability.Data Visualization & User-Facing Analytics Design and develop interactive dashboards, visual analytics, and scientific data visualizations using Power BI and Spotfire (or equivalent tools).Create reusable visualization components and data-rich UIs (React/TypeScript preferred) to enable scientists to search, filter, explore, and interpret complex datasets—including dose-response curves, chemical structures, and analytical results.Translate scientific and engineering data into clear, actionable visual insights for researchers and stakeholders.Software Engineering & Quality Practices Apply best software engineering practices: modular/reusable design, clean code principles, code reviews, comprehensive documentation, and creation of maintainable libraries/services.Write high-quality unit, integration, and end-to-end tests; use mock data effectively to create reliable automated test cases and ensure code stability.Implement CI/CD pipelines for automated testing, deployment, and monitoring on AWS (EC2, ECS, Lambda, S3).Collaborate on full-stack features from database to frontend, ensuring end-to-end functionality, security (SSO/LDAP), and performance.Collaboration & Governance Partner with scientists, UX designers, and cross-functional teams to gather requirements, conduct user testing, and iterate on usability.Implement data validation, quality checks, metadata management, and governance to ensure compliance and accuracy.Contribute to engineering best practices and foster a culture of quality and scalability.Requirements:Education & Experience Bachelor's degree in Computer Science, Data Engineering, Software Engineering, Information Systems, or a related technical field.3+ years of professional experience in data engineering, full-stack development, or closely related roles.Proven track record of building and delivering production-grade data pipelines, platforms, and/or user-facing scientific applications.Technical Skills Programming: Intermediate to strong proficiency in Python (core for pipelines, backend, and data manipulation with pandas/PySpark); familiarity with JavaScript/TypeScript for frontend.Data Engineering: Hands-on experience creating scalable pipelines, ETL/ELT processes, and distributed processing (Spark, Trino/Presto).Databases & Querying: Deep expertise in relational databases (PostgreSQL), modern warehouses (Snowflake, Redshift), and query engines; strong focus on query performance improvement and optimization.Cloud Platforms: Practical experience with AWS services (S3, Glue, Athena, Lambda, RDS, EC2/ECS).Data Visualization: Proven experience with Power BI and Spotfire (or similar) for scientific and analytical dashboards/visualizations.Frontend (preferred): Modern JavaScript/TypeScript frameworks (React preferred), responsive UI development, and component libraries.Testing & Quality: Strong unit testing skills; experience writing automated tests with mock data for robust coverage.Tools & Practices: Git for version control; API design (RESTful); CI/CD; clean code and reusable library development.Core Competencies Excellent reverse engineering and troubleshooting capabilities for complex data and system issues.Strong problem-solving skills with attention to detail and commitment to data quality/accuracy.Ability to work independently and collaboratively in cross-functional, scientific teams.Excellent communication skills to bridge technical concepts with non-technical stakeholders (scientists, researchers).Best-In-Class-Benefits:We are in the people business; treating people right is our ONLY priority. Theoris Services consultants are full-time employees with full benefits, including:Robust Health Insurance401(k) planAbout Theoris: Our goal is to Fuel Your Career! As a Theoris team member, you join a culture based on people-centered values and an environment that fosters both personal and professional growth. We build long-term relationships with our clients and our consultants. With over 30 years of building strong relationships in the industry, we’re uniquely positioned to make the right connections. This knowledge is used to find the right job placement. Our recruiting teams are experts dedicated to the information technology and engineering staffing space and are highly respected by our client base.
Showing 650 of 11,883 matching similar jobs in Springbrook, ND
- Lead Data Architect (Google Spanner)
- AVP, Metadata & Data Lineage Architect
- Senior BI/DW Engineer: AI-Driven Analytics & Data Flows
- Senior Data Architect - Financial Data & Cloud-First
- Senior Staff Software Engineer, Data Platform
- Director, Data Architecture - Azure & Oracle (Remote)
- Cloud SQL Developer: ETL & Data Pipelines
- Senior Data Quality Engineer - Azure & Databricks
- Adobe AEP Architect
- Remote Lead Data Architect: Snowflake + dbt Modernization
- Data Engineer / ETL Engineer for AI-Driven Pipelines
- Remote Azure Data Engineer - Pipelines & ETL
- Senior Power Platform BI & Apps Analyst
- Software Engineer - BigQuery Metadata & Data Warehouse
- AWS Cloud Platform Engineer (Remote) - Secure & Scalable
- Remote Azure Data Engineer - Pipelines & ETL
- Director Power BI & Python Developer (Microsoft Fabric Platform)
- Remote Data Engineer Consultant | Snowflake Pipelines
- Senior Software Engineer: BI, ETL & Data - TS/SCI
- Senior Automation Engineer - Data Warehouse (Remote)
- Senior Power BI Developer
- ERP/MRP Data Architect
- Senior Data Integration Engineer - AWS & API Pipelines
- Senior ETL Architect - Remote Data Integration
- CPU Performance Architect: Workload Analytics & SoC
- Data Migration Architect
- Data Engineer
- Data Engineer (Cloud Migration - AWS)
- ETL Architect
- Principal Data Architect & Data Modeler
- Senior Principal Enterprise Architect
- Senior Database Engineer & Cloud Solutions Consultant
- Sigma BI Developer: Dashboards & Data Analytics
- Sr. ETL Developer
- Big Data Solutions Architect
- Data Integration Specialist
- SF Datacloud Architect
- Ensemble Interface Developer/Analyst on w2 contract
- Senior Ab Initio ETL Lead - Data Integration & Architecture
- Remote Data Quality Consultant - Governance & Rules