Staff Data Engineer

Company Profile

Logicbroker provides modern dropship and marketplace solutions that connect retailers and brands to Connect, Orchestrate, and Grow their commerce platform to take control of their customer experience. Through curated expanded assortment, flexible integration tools, and automated onboarding procedures, Logicbroker clients enjoy unmatched speed-to-market capabilities. We work with mid-market and Enterprise manufacturers and retailers across a number of verticals including Health & Wellness, Home Improvement, Consumer Electronics, Toys & Babies, and Consumer Packaged Goods and service brands such as Samsung, Victoria’s Secret, Ace Hardware and more.

Position – Staff Data Engineer

As a Staff Data Engineer, you will set the technical vision and architecture for Logicbroker’s data infrastructure. You will build and optimize large-scale data pipelines, event streams, and data services that power analytics, AI-driven insights, and mission-critical commerce operations. You’ll lead initiatives to modernize our data platform, mentor engineers, and collaborate with cross-functional teams (Engineering, SRE, Data Science, Product) to deliver scalable, secure, and high-performance data systems.

Primary Duties and Responsibilities:

  • Define the strategy for our data pipelines, ETL/ELT workflows, and event streaming architectures to ensure scalability and reliability.
  • Design and implement real-time and batch data processing systems using frameworks like Apache Spark, Flink, or Beam.
  • Lead efforts to evolve our data lakehouse and warehouse solutions (e.g., Databricks, Redshift, BigQuery, or Delta Lake).
  • Establish best practices for data modeling, schema evolution, partitioning, and query optimization.
  • Champion data observability and quality, building automated validation, monitoring, and anomaly detection systems.
  • Build data ingestion services that handle high-volume e-commerce events (orders, inventory, shipments) in real time.
  • Implement and optimize data APIs and self-service data access for analytics, machine learning, and reporting teams.
  • Develop robust streaming pipelines (Kafka, Kinesis) and integrate with event-driven architectures across the platform.
  • Improve data tooling for developers: local testing environments, automated schema checks, and metadata management.
  • Partner with Product and Data Science teams to translate business needs into data products and pipelines.
  • Work closely with SRE to ensure data services meet strict SLAs, SLOs, and resilience goals.
  • Collaborate with stakeholders to deliver actionable insights, enabling real-time dashboards, operational metrics, and predictive analytics.
  • Represent data engineering in strategic planning sessions and technical deep-dives with leadership.
  • Mentor senior and mid-level engineers, guiding data architecture decisions and career development.
  • Advocate for data-driven engineering culture, setting standards for testing, documentation, and secure data handling.
  • Contribute to hiring efforts, ensuring we build a world-class data engineering team.

Essential Skills, Experience & Education

  • Bachelor’s degree in Computer Science, Data Engineering, or related field.
  • 10+ years of experience in software/data engineering, with at least 2 years in a senior or staff-level leadership capacity.
  • Deep expertise in cloud-native data engineering and distributed data processing.
  • Proficiency in Python, Java, Scala, Go, or other backend/data-focused languages.
  • Strong experience with streaming data frameworks (Kafka, Kinesis, or Pulsar) and data workflow orchestration (Airflow, Dagster, Prefect).
  • Advanced knowledge of SQL and NoSQL databases (Postgres, DynamoDB, MongoDB, Cassandra).
  • Proven ability to design and optimize large-scale data pipelines (batch and real-time).
  • Familiarity with containerization and orchestration (Docker, Kubernetes).
  • Expertise with at least one major cloud platform (AWS, GCP, or Azure) and their data services.
  • Experience with CI/CD pipelines, infrastructure-as-code, and automated testing in data environments.
  • Experience with lakehouse technologies (Delta Lake, Iceberg) and modern data warehouses (Snowflake, BigQuery).
  • Exposure to AI/ML data pipelines, feature stores, and model serving.
  • Strong knowledge of data security, compliance, and governance (e.g., SOC 2, GDPR).
  • Familiarity with observability and lineage tools (OpenLineage, Monte Carlo, or DataDog).

Compensation and Environment

Logicbroker offers a comprehensive retirement plan, workspace compensation, and more! Logicbroker embraces a truly remote work model. All Logicbroker employees will receive thorough training to kick-start their career with the organization, and we will provide the tools needed to build a successful home workspace.

Logicbroker is an equal opportunity employer.

Apply today

Send us your resume at [email protected]

Apply now