Accepting Applications
Full-time
On-site
Posted 3 hours, 31 minutes ago
0 views
0 applications
Job Description
**About the Role**
We are looking for a strong Data Engineer/Scientist to help us build the systems that collect, clean, structure, sync, and maintain large volumes of data from many different sources.
This role is ideal for someone who is strong in backend engineering, data pipelines, system architecture, and data quality. The person should be able to take messy, inconsistent, high\-volume data and turn it into structured, reliable, usable datasets that can power internal tools, AI systems, search, analytics, and customer\-facing applications.
The ideal candidate is highly adaptable, communicates clearly, learns new tools quickly, and enjoys solving difficult data infrastructure problems.
**What You’ll Do**
* Design and build scalable data pipelines for collecting, syncing, processing, and structuring large volumes of data.
* Work with messy, incomplete, duplicated, or inconsistent data from many different external sources.
* Build systems for data cleaning, normalization, deduplication, validation, enrichment, and quality checks.
* Create reliable sync processes that keep data fresh, accurate, and consistent over time.
* Design backend services, databases, and data models that support fast querying, searching, and downstream usage.
* Improve speed, accuracy, efficiency, and reliability of data ingestion and processing workflows.
* Build monitoring, alerting, logging, and error\-handling systems for data pipelines.
* Work with engineering, product, and AI teams to understand how data should be structured and used.
* Evaluate and integrate new tools for scraping, data processing, workflow orchestration, and AI\-assisted data extraction.
* Help design systems that can support agentic workflows, automation, and AI\-powered applications.
**What We’re Looking For**
* Strong experience building production\-grade data pipelines or backend data systems.
* Strong understanding of databases, APIs, ETL/ELT workflows, data modeling, and distributed systems.
* Experience handling large, messy, real\-world datasets from multiple sources.
* Experience with deduplication, entity matching, normalization, data validation, and data quality processes.
* Strong programming skills in Python, TypeScript, Go, Java, or similar backend/data\-focused languages.
* Experience with PostgreSQL, MySQL, MongoDB, Elasticsearch/OpenSearch, Redis, or similar data stores.
* Experience with queues, workers, batch processing, streaming, or workflow orchestration tools.
* Ability to design systems for reliability, speed, accuracy, and maintainability.
* Strong debugging skills and the ability to trace problems across complex data flows.
* Clear communication skills and the ability to explain tradeoffs between accuracy, speed, cost, and complexity.
* High ownership, adaptability, and willingness to learn new tools quickly.
Login to Apply
Don't have an account? Register