
Job Overview
Location
Bengaluru
Job Type
Full-time
Category
Software Engineering
Date Posted
October 27, 2025
Full Job Description
đź“‹ Description
- • Architect and deliver the next-generation data backbone that powers 750 million weekly viewers across JioStar’s television and streaming ecosystem. You will own the end-to-end lifecycle of systems that ingest hundreds of millions of requests per minute, stream billions of daily events through Kafka, and store petabytes of data in S3, ensuring 99.99 % availability and sub-second latency for both batch and real-time workloads.
- • Design and build POCs, blueprints, and migration strategies for new services while rigorously evaluating trade-offs in performance, cost, and maintainability. Your architectural decisions will directly influence the scalability of AI-powered experiences and autonomous agents that redefine how users discover and interact with content.
- • Create resilient, real-time data pipelines that guarantee clean, accurate, and complete event collection across every touchpoint—mobile apps, web, set-top boxes, smart TVs, and ad servers. You will implement exactly-once semantics, idempotent writes, and schema-evolution strategies that future-proof the platform against exploding data volume and variety.
- • Establish the data foundation for cutting-edge AI initiatives, including the in-house MCP (Model Context Protocol) server and the infrastructure that trains and serves autonomous AI Agents. You will curate high-quality datasets, feature stores, and offline/online inference pipelines that enable personalized learning paths, dynamic ad insertion, and real-time content recommendations.
- • Institute a culture of data quality by building automated anomaly detection, data-contract validation, and lineage-tracking systems that surface issues before stakeholders notice them. You will define SLIs/SLOs for freshness, completeness, and correctness, and create self-healing workflows that reduce mean-time-to-recovery from hours to minutes.
- • Drive continuous cost optimization across ingestion, storage, and processing layers by tuning Spark/Flink jobs, adopting Graviton/ARM instances, implementing intelligent tiering on S3, and leveraging spot-instance orchestration. You will model unit economics and present monthly savings reports to executive leadership.
- • Provide technical leadership by owning the complete development lifecycle—from ideation and design to deployment and post-production support. You will author architecture decision records (ADRs), sequence diagrams, and runbooks that become the single source of truth for your services.
- • Collaborate cross-functionally with product managers, data scientists, ML engineers, and business stakeholders to translate ambiguous requirements into scalable technical solutions. You will run design reviews, threat-modeling sessions, and blameless post-mortems that foster shared ownership and rapid iteration.
- • Champion operational excellence by building comprehensive monitoring, alerting, and auto-remediation frameworks using Prometheus, Grafana, PagerDuty, and custom Kubernetes operators. You will define golden signals and error budgets that keep the platform resilient during marquee events like IPL finals and global movie premieres.
- • Mentor and up-level engineers through pair programming, architecture deep-dives, and weekly tech talks. You will write internal blogs and open-source contributions that propagate best practices in distributed systems, data engineering, and cloud cost optimization across the wider technology community.
- • Lead data governance initiatives that catalog every dataset, enforce PII masking, and ensure compliance with GDPR, CCPA, and emerging privacy regulations. You will partner with legal and security teams to implement fine-grained access controls and audit trails that protect user trust.
- • Reduce time-to-insight for analysts and data scientists by building intuitive, self-service interfaces—ranging from SQL-based exploration layers to semantic catalogs—that make petabyte-scale data discoverable and queryable in seconds rather than hours.
🎯 Requirements
- • 8+ years of progressive experience designing, building, and operating large-scale, high-throughput data platforms in consumer-facing or streaming industries, with proven ownership of systems handling hundreds of millions of RPM, billions of daily events, and petabytes of storage.
- • Expert-level proficiency in Python, Java, Scala, or Go, coupled with deep hands-on experience in distributed data processing frameworks such as Apache Spark, Apache Flink, or equivalent for both batch and streaming workloads.
- • Demonstrated track record of building data pipelines for machine-learning model training, inference, and feature stores, including familiarity with data requirements for AI Agents or similar intelligent systems.
- • Extensive production experience with high-throughput messaging and stream-processing platforms like Apache Kafka, AWS Kinesis, or equivalent, including partition-rebalancing, exactly-once semantics, and schema registry management.
- • Hands-on expertise with AWS data services—S3, EMR, Kinesis, Glue, Redshift, Lambda—and modern data lake/warehouse architectures such as Delta Lake, Iceberg, Hudi, Snowflake, or BigQuery.
- • Advanced SQL skills, strong data-modeling expertise (dimensional, Kimball, data vault), and experience orchestrating complex ETL/ELT pipelines using Apache Airflow, Prefect, or similar workflow engines.
🏖️ Benefits
- • Work on a platform that reaches 750 million viewers weekly, giving your code immediate, tangible impact on global entertainment and sports consumption.
- • Shape the future of AI-powered experiences and autonomous agents in media, influencing product roadmaps that redefine content discovery and personalization.
- • Collaborate with world-class engineers, data scientists, and product leaders in a culture that values diversity, inclusion, and continuous learning.
- • Competitive compensation package, comprehensive health coverage, and generous learning stipends for conferences, courses, and certifications.
Skills & Technologies
Python
Java
Scala
PostgreSQL
MySQL
Senior
Onsite
About Jiostar Technologies Private Limited
Jiostar Technologies Private Limited is an Indian technology company delivering enterprise-grade digital transformation solutions. The firm specializes in cloud infrastructure, data analytics, cybersecurity, and AI-driven business applications for telecom, finance, and retail sectors. Its unified platform integrates IoT, edge computing, and API management to streamline operations and enhance customer engagement. Operating from Mumbai and Bengaluru, Jiostar serves large corporations seeking scalable, secure, and compliant technology stacks across India and emerging markets.



