Company Description
Modofy specializes in designing and engineering data platforms, AI solutions, and business intelligence tools that transform complex enterprise data into actionable insights. Partnering with Fortune 500 companies and high-growth organizations, Modofy delivers scalable solutions in data engineering, AI, and analytics, including real-time streaming systems, machine learning applications, and self-service analytics. With a tailored approach, the company adapts its expertise to unique project needs, whether it's a long-term transformation or an embedded team model. At Modofy, the mission is clear: to engineer data that drives performance and measurable business outcomes.
Responsibilities
- Design and build real-time data pipelines using Kafka, ClickHouse, and MongoDB
- Implement and maintain CDC pipelines from source systems (MySQL, Oracle, PostgreSQL) using tools like Streamkap or Debezium
- Develop and optimize ETL workflows — ingestion, transformation, and loading into ClickHouse medallion layers (bronze/silver/gold)
- Write and maintain DBMS_SCHEDULER jobs or equivalent for scheduled Oracle extracts to S3
- Build and manage ClickPipes, materialized views, and ReplacingMergeTree/Deduplication patterns in ClickHouse
- Collaborate with QA on data quality checks, row count reconciliation, and SLA monitoring
- Maintain Jira tickets, technical docs, and async Slack-based delivery cadence
Qualifications
- 3+ years hands-on data engineering experience with real-time or streaming pipelines
- Strong SQL — ClickHouse, Oracle, or PostgreSQL; comfortable writing complex analytical queries
- Experience with Kafka or equivalent message brokers (consumer groups, offsets, partitioning)
- Python proficiency for pipeline scripting, automation, and data validation
- Familiarity with CDC concepts — at-least-once delivery, idempotency, offset management
- MongoDB experience — change streams, document modeling, schema evolution
- Cloud-comfortable — S3, IAM, RDS, basic AWS networking
- Remote-ready; async-first, low-meeting work style
Nice to have
- Direct experience with ClickHouse — MergeTree engines, ClickPipes, materialized views
- Streamkap, Fivetran, or Airbyte connector experience
- Oracle DBMS_SCHEDULER or PL/SQL for extract jobs
- Exposure to Metabase or BI layer configuration
- Familiarity with Jira-based delivery workflows