Unlimited Job Postings Subscription - $99/yr!

Job Details

Senior Data Engineer

  2025-06-04     Worth AI     all cities,AK  
Description:

Worth AI, a leader in the computer software industry, is looking for a talented and experienced Senior Data Engineer to join their innovative team. At Worth AI, we are on a mission to revolutionize decision-making with the power of artificial intelligence while fostering an environment of collaboration, and adaptability, aiming to make a meaningful impact in the tech landscape.. Our team values include extreme ownership, one team and creating reaving fans both for our employees and customers.

As a Senior Data Engineer, you will lead the design and development of data services and platforms that power our AI-driven products. You'll focus on creating well-structured, validated APIs and event-driven pipelines, enabling scalable, secure, and maintainable data workflows. This is a backend-heavy role ideal for engineers who thrive on clean architecture, automation, and cross-functional delivery.

Responsibilities
  • Design and build production-grade **FastAPI services** to serve, validate, and enrich data for ML and analytics use cases
  • Create and maintain **asynchronous event-driven pipelines** using
  • *Apache Kafka**, ensuring reliable and scalable communication across microservices
  • Define and enforce structured data contracts using **Pydantic** and OpenAPI standards
  • Develop robust, containerized data services with **Docker** and deploy them using modern cloud-native tooling
  • Build and optimize analytical models and data flows in **Amazon Redshift** for business-critical reporting and data science consumption
  • Collaborate with data scientists, ML engineers, and backend developers to streamline data sourcing, transformation, and model inference
  • Own the lifecycle of data services — including monitoring, observability, testing, and deployment pipelines
  • Maintain rigorous standards around data privacy, schema governance, and system performance
  • Design, build, code and maintain large-scale data processing systems and architectures that support AI initiatives.
  • Develop and implement data pipelines and ETL processes to ingest, transform, and load data from various sources.
  • Design and optimize databases and data storage solutions for high performance and scalability.
  • Collaborate with cross-functional teams to understand data requirements and ensure data quality and integrity.
  • Implement data governance and data security measures to protect sensitive data.
  • Monitor and troubleshoot data infrastructure and pipeline issues in a timely manner.
  • Stay up-to-date with the latest trends and technologies in data engineering and recommend improvements to enhance the company's data capabilities.

Requirements

    • 7+ years of professional experience in backend-focused data engineering or platform development
    • Strong proficiency in **Python**, with hands-on experience using **FastAPI**, **Pydantic**, and asynchronous programming patterns
    • Deep understanding of **event-driven architectures** and experience with **Kafka** (producers, consumers, schema evolution, retries, etc.)
    • Experience designing and deploying **containerized services** with **Docker** (Kubernetes or Fargate experience is a plus)
    • Proficiency in SQL and experience with modern cloud data warehouses, preferably **Amazon Redshift**
    • Familiarity with cloud services (preferably AWS), including CI/CD, infrastructure-as-code, and observability tooling
    • Experience integrating third-party APIs and working with versioned schema contracts
    • Strong communication and collaboration skills, especially in cross-functional and agile teams
    • Experience working with ML engineers to operationalize models (e.g., batch scoring, online inference, data validation at model boundaries)
    • In-depth knowledge of data modeling, data warehousing, and database design principles.
    • Strong programming skills in Python, SQL, and other relevant languages.
    • Experience with relational and NoSQL databases, such as PostgreSQL, MySQL, MongoDB
    • Proficiency in data integration and ETL tools, such as Apache Kafka, Apache Airflow, or Informatica.
    • Familiarity with big data processing frameworks, such as Hadoop, Spark, or Flink.
    • Knowledge of cloud platforms, such as AWS, Azure, or GCP, and experience with data storage and processing services in the cloud.
    • Understanding of data governance, data privacy, and data security best practices.
    • Strong problem-solving and troubleshooting skills, with a focus on data quality and system performance.
    • Excellent communication and collaboration skills to work effectively with cross-functional teams.
    • Prior collaborative work with data scientists or machine learning professionals with respect to sourcing, processing and scaling both input and output data
    • Comfortable going through documentation of third-party API's and identifying best procedures for integrating data from API's into broader ETL processes

Benefits

    • Health Care Plan (Medical, Dental & Vision)
    • Retirement Plan (401k, IRA)
    • Life Insurance
    • Unlimited Paid Time Off
    • 9 paid Holidays
    • Family Leave
    • Work From Home
    • Free Food & Snacks (Access to Industrious Co-working Membership!)
    • Wellness Resources


Apply for this Job

Please use the APPLY HERE link below to view additional details and application instructions.

Apply Here

Back to Search