Eventador announces $3.8MM seed round (including conversion of $1.3MM angel round) led by LiveOak Venture Partners.
Streaming data is everywhere. IoT, high tech manufacturing, national security, smart cities, web log analysis, systems telemetry, AI and ML workflows, and a myriad of other modern use cases are driving this trend skyward.
Core support for Simple Authentication and Security Layer (SASL) was added to Apache Kafka in the 0.10.2 release. This allows for simple username/password authentication to Kafka using SASL. We are excited to add this authentication mechanism to the Eventador service. Here is how it works.
Apache Flink offers two simple API’s for accessing streaming data with declarative semantics – The table and SQL API’s. In this post we dive in an build a simple processor in Java using these relatively new API’s.
Early this week I gave a talk at the Austin Kafka/Stream Processing Meetup. It was a great time and we had a fantastic turnout. I wanted to share the slides, examples, and a couple of thoughts on the Meetup itself.
When we started Eventador.io in 2016 we needed a simple data source to help us build the platform on. We needed something that exemplified streaming data, something massively dynamic, and something with a lot of data. Tweets were played out, we wanted something better.
With the addition of Apache Flink – Eventador.io has a true end-to-end enterprise grade stream processing platform. We run the complex infrastructure and provide support, you can focus on your streaming code.
One of the omnipresent challenges of building a product from scratch is you don’t initially know exactly how customers will want to use it. You build the product you would want to use and are passionate about, however, you must also listen to customers as you evolve your product to deliver exactly what they really […]
Since we first opened the doors at Eventador.io, customers have been building applications that make use of Apache Kafka for a wide variety of streaming data use cases. Over time, it became clear we were only solving for one part of the complete picture. With Kafka, our service had the data transport, durability, and scalability, […]
This release focuses on making the service even more robust, easier to use, and overall customer experience. Many of these features were inspired by direct feedback from you, our customers. Thank you for helping us build the best Apache Kafka™ managed service in existence.