
Job Overview
Location
Bengaluru
Job Type
Full-time
Category
Software Engineering
Date Posted
October 27, 2025
Full Job Description
đź“‹ Description
- • Architect & ship high-impact ingestion services that sustain 100M+ requests per minute, ensuring zero-downtime upgrades and sub-second latency for our global streaming audience.
- • Design and extend real-time event pipelines that consume billions of Kafka messages daily, transforming raw click-streams, video QoS metrics, and ad-impression logs into clean, governed datasets ready for AI training.
- • Build the data backbone that powers next-generation AI Agents and our proprietary MCP server—curating petabyte-scale feature stores, low-latency online inference tables, and versioned training snapshots that accelerate experimentation cycles.
- • Own end-to-end delivery of ETL/ELT workflows using Apache Spark/Flink on EMR, orchestrated through Airflow/Prefect, with rigorous unit, integration, and data-quality tests that guarantee 99.9 % accuracy.
- • Continuously refactor and harden existing data lake components (S3, Delta Lake, Iceberg) to cut storage cost by double-digit percentages while doubling query throughput for analysts and data scientists.
- • Instrument comprehensive observability—metrics, logs, traces, and data lineage—using Datadog, Prometheus, and Grafana so anomalies are caught and remediated before viewers feel any impact.
- • Collaborate with senior engineers to evolve our Infrastructure-as-Code stack (Terraform, CloudFormation), enabling one-click deployment of new environments and enforcing security/compliance guardrails.
- • Partner with product, ML, and content teams to translate ambitious AI roadmaps into concrete data requirements, ensuring every new model has the high-quality, privacy-compliant data it needs.
- • Participate in blameless post-mortems and design reviews, leveling up team standards for code quality, performance tuning, and operational excellence.
- • Mentor junior developers through pair-programming and crisp technical documentation, multiplying the team’s velocity and knowledge base.
🎯 Requirements
- • 2–4 years of software engineering with at least 2 years in data engineering or distributed systems roles
- • Hands-on expertise with Apache Spark or Flink for both batch and streaming workloads
- • Production experience with Kafka (or Kinesis/Pulsar) at multi-million-events-per-minute scale
- • Strong SQL and working knowledge of AWS data services (S3, EMR, Kinesis, Redshift, Glue)
- • Proficiency in Python, Java, Scala, or Go with a commitment to clean, testable code
- • (Nice-to-have) Exposure to ML feature pipelines, Airflow/Prefect orchestration, or Infrastructure-as-Code
🏖️ Benefits
- • Work on petabyte-scale systems that reach 750 million weekly viewers and shape the future of AI-powered entertainment
- • Hybrid on-site model in Bengaluru with a vibrant, inclusive culture that celebrates diversity and authenticity
- • Continuous learning budget and dedicated time for exploring cutting-edge data & AI technologies
- • Competitive compensation, comprehensive health coverage, and flexible leave policies
Skills & Technologies
Python
Java
Go
Scala
PostgreSQL
Onsite
About Jiostar Technologies Private Limited
Jiostar Technologies Private Limited is an Indian technology company delivering enterprise-grade digital transformation solutions. The firm specializes in cloud infrastructure, data analytics, cybersecurity, and AI-driven business applications for telecom, finance, and retail sectors. Its unified platform integrates IoT, edge computing, and API management to streamline operations and enhance customer engagement. Operating from Mumbai and Bengaluru, Jiostar serves large corporations seeking scalable, secure, and compliant technology stacks across India and emerging markets.



