Every Kafka deployment on Eventador has an associated access control list (ACL). The ACL defines what IP addresses are whitelisted and allowed access to produce and consume to and from your deployment. In fact, there are no entries at deploy time, thus access is completely denied. In order to use our service you must first […]
Getting connected and producing (publishing) and consuming (subscribing) messages is relatively easy in Apache Kafka. In this post we will go over connecting, producing a simple message, and consuming that message via one of a couple native python clients. Most languages are similar and there are a host of native drivers to choose from.
With the release of Eventador 0.5 we are introducing new plans with one-click provisioning. This allows you to deliver your projects in a more timely basis, save costs on valuable resources, leverage the cloud more effectively, get worry free 24×7 support, and have best in class data pipeline infrastructure simply by using Eventador.io. I thought […]
Apache Kafka was built by a team of Engineers at LinkedIn circa 2000. They were trying to solve the problem of how to pipeline data between various components of their micro-services based architecture. They decided they needed a single pub/sub messaging platform designed to move data efficiently. Kafka was born. Flash forward 2015. @erikbeebe and […]
Real-time data is only as good as your ability to analyze and use that data. We wanted a powerful yet simple interface to Eventador data pipelines. Notebooks have been very popular inside the data science community for some time, and they are a natural fit for Eventador. Today we are releasing Eventador Beta 0.3.0 which […]
We built Eventador to solve a pervasive and tricky problem; It’s exceptionally challenging to build real-time data systems using existing technology. To address this problem, in July, we launched our real-time data pipeline service based on Apache Kafka. Since launch we have been working feverishly on the platform. Iterating, adding features, and core functions. Today […]
Over the past two decades I have witnessed many massive shifts in the data technologies landscape. I have seen a maturity of the RDBMS market, the adoption of NoSQL technologies, and organizations of every size struggling with data management in one way or another. Throughout all of this change, one area is starting to become […]