Level Up Your Kafka Skills in Just 5 Days | Join Season of Streaming On-Demand
For the first 12 days of December, Confluent shared a daily tech tip related to managing Apache Kafka® in the cloud. These tips make it easier for you to get started with Kafka, ramp up on KSQL and move applications to the cloud. We’ve collected all the tips here, so if your New Year’s resolution is learning to use Kafka or its ecosystem—bookmark this page!
One of the easiest ways to get started with Apache Kafka is by…not installing Kafka at all. You can get started in three minutes and nine seconds by creating a cluster on Confluent Cloud.
Once you have access to your cluster, the next logical step for Java developers is to learn the APIs by writing a small application. Confluent Cloud is secured by default, so your first application will include a secured connector to the Kafka cluster…Let us show you how easy it is!
Once you’ve learned the basic APIs, perhaps it is time to create a data pipeline that transforms events and loads them to any external end system, which in this case, is Google BigQuery. With KSQL and a BigQuery connector, this doesn’t take long.
If you liked the pipelines example, you might wonder how to run a KSQL server with your Confluent Cloud cluster. This is quite easy when you know about a few simple parameters.
But why stop there? Apache Kafka and the Confluent Platform include many more components, and you probably want to run all of them—Confluent Schema Registry, connectors, Confluent REST Proxy and more. Our Terraform automation will help you spin all of them up.
And if you use Kafka Streams to build event-driven microservices, you will want to connect those to Confluent Cloud, too.
If all goes well, then you’ve developed an app or two and have run them against Confluent Cloud. This means that you are now working toward a production rollout. For production, you’ll want to know about your cluster capabilities and limitations, so make sure you read the list.
If your production rollout was successful, you want to keep an eye on your cloud usage—what’s your network throughput? What’s your storage usage? Are you close to hitting any limits? You don’t want to be surprised when you hit a limit, or when you look at your next bill.
Going to production sometimes means you need extra security. For that, you may decide to run your Confluent Cloud cluster within your corporate VPC. You can use VPC peering to merge the Confluent Cloud VPC with yours.
If your organization already uses Kafka, you may need to stream some data from your on-prem cluster to Confluent Cloud so you can use this data in your cloud applications.
You may decide to use your new data replication capabilities to assist in a larger-scale cloud migration effort—2019 is a great year to migrate applications and data to the cloud!
Bridge to cloud is more than just lifting and shifting data from on prem to the cloud. It is about integrating on-prem and cloud datastores and data processing. Learn how to use Kafka Connect and KSQL to bridge between two ecosystems of data and applications:
With that, enjoy the holidays and have a productive 2019! If you have additional tips to share with new Kafka users, please let us know in the comments below.
This blog announces the general availability of Confluent Platform 7.8 and its latest key features: Confluent Platform for Apache Flink® (GA), mTLS Identity for RBAC Authorization, and more.
We covered so much at Current 2024, from the 138 breakout sessions, lightning talks, and meetups on the expo floor to what happened on the main stage. If you heard any snippets or saw quotes from the Day 2 keynote, then you already know what I told the room: We are all data streaming engineers now.