The Ingestion team owns the entry point into the Treasure Data platform. We reliably ingest high-volume event data from external sources, apply safe and programmable transformations, buffer data durably, and deliver it to multiple downstream storage and processing systems.
These pipelines form the foundation for analytics, personalization, and AI-driven use cases across the platform. As event volumes scale into the billions per day , the team plays a critical role in enabling customer growth, maintaining platform reliability, and preparing the system for multi-fold increases in traffic in a predictable and cost-efficient way.
Responsibilities
-
Design, build, and operate large-scale backend services for real-time and near-real-time data ingestion
-
Own major components or services end-to-end, from design through production operations
-
Improve system reliability, scalability, and cost efficiency under high and unpredictable traffic
-
Collaborate with product, customer success, and other engineering teams to support customer onboarding and growth
-
Participate in on-call rotations and take ownership of production issues and incident follow-ups
-
Use metrics, monitoring, and data to guide engineering decisions
-
Contribute clean, well-tested code and actively participate in design and code reviews
-
Mentor junior engineers and raise the technical bar of the team through example
Requirements
-
Have 3+ years of professional software engineering experience (or equivalent practical experience)
-
Have built and maintained backend or distributed systems at scale
-
Understand reliability, availability, and failure modes in production environments
-
Are comfortable owning services with real customer and business impact
-
Write clear, maintainable, and well-tested code
-
Communicate effectively in a distributed, global team
-
Are comfortable working primarily with Kotlin and Java , and open to learning new tools and technologies
-
Have experience with cloud platforms such as AWS
-
Business-level English proficiency
Nice to haves
While not specifically required, tell us if you have any of the following.
-
Experience with streaming or real-time data systems (Kafka, Kinesis, Flink, Spark)
-
Experience scaling systems by 3–5× or more
-
Familiarity with infrastructure as code, observability, or CI/CD
-
Open-source contributions