Splunk and Pivotal Cloud Foundry: Get the New Nozzle

Splunk proudly welcomes the new Splunk Firehose Nozzle for Pivotal Cloud Foundry! This release builds on the previous efforts of our work with Pivotal, moving the release from a private beta to an official public release.

What is Pivotal Cloud Foundry?

Pivotal Cloud Foundry (PCF) is one of the world’s most powerful cloud-native platforms to rapidly develop and run modern and legacy applications at startup speeds. In addition, Pivotal Cloud Foundry is the only application development platform that runs on any cloud infrastructure—across public, private and managed clouds.

Cloud Foundry is used by over half of the Fortune 500 and a rapidly growing portion of the Fortune 2000. Pivotal’s customers have experienced a 2,000 percent increase in developer productivity, as well as a 50 percent reduction in IT costs due to platform automation. To learn more, visit the Pivotal Cloud Foundry website.

What is the Splunk Firehose Nozzle for Pivotal Cloud Foundry?

Pivotal Cloud Foundry consolidates application logs and platform components’ metrics to a PCF component known as Loggregator. To get events out of PCF and into your Splunk environment, you need a Nozzle that attaches to the Loggregator Firehose.

This is where the Splunk Firehose Nozzle for Pivotal Cloud Foundry comes in!

Splunk Firehose Nozzle connects to the Loggregator Firehose Endpoint and streams all available events into your Splunk environment via the HTTP Event Collector (HEC).

Use the Splunk Firehose Nozzle for PCF to select, buffer, and transform your events. Adding extra metadata to your events can be leveraged in Splunk correlation searches for data loss tracking and event filtering, among other interesting use cases.

It’s What's Inside that Counts

The Splunk Firehose Nozzle for PCF was developed as an application that runs on Pivotal Cloud Foundry. The Nozzle subscribes to the Loggregator endpoint and writes events to an external Splunk environment.

Figure 1 – High-Level System Integration Diagram (Splunk + Pivotal Cloud Foundry)

  1. The Splunk Firehose Nozzle for PCF collects events from the PCF Loggregator endpoint and streams them to Splunk via HTTP event collector. Nozzle has in-memory queue buffers to increase reliability, and has parallel client to scale out multiple ingestion channels to HEC.
  2. The Splunk Firehose Nozzle for PCF can be deployed natively within a PCF environment and is available as a free tile from the PCF Marketplace.
  3. The Splunk HTTP Event Collector clients run concurrently, consuming events from the queue to enrich PCF events by attaching metadata fields. For scaling out, add as many HECs and place a load balancer in front.
  4. After data has been ingested into Splunk, it can be explored using the Splunk Add-on for Cloud Foundry—an add-on to Splunk which parses data from any Cloud Foundry distribution—which includes pre-built panels and pre-configured search parameters*.

*Note: Many configurable environment parameters are included in this release, which can modify the features discussed above as well as many others. For a closer look at these configurations, see this page on the Splunk Firehose Nozzle.

Searching PCF Data in Splunk

After ingesting Pivotal Cloud Foundry events, use Splunk’s search language (SPL) to configure visualizations and alerts on important PCF logs and metrics.

The following SPL returns the percentage of events that your Splunk deployment sees and indexes. This search can identify any data loss within the Splunk Firehose Nozzle, which you can use to trigger an investigation into your environment.

subscription-id=splunk-firehose uuid=b28978ba-f83d-4d2f-99c3-c18b1a3f8ebf 
| stats count as total_events , max(nozzle-event-counter) as max_number 
| eval total=(total_events/max_number) * 100 
| table total

Figure 2 – SPL demonstrating percentage of successfully indexed events.

Version 1.0.0 Release Highlights - Fast, Scalable, Reliable

Highlights in this release include:

Hearing what our users want, we included three core capabilities into v1.0.0:

We tested the Splunk Firehose Nozzle running as a single deployed nozzle on the AWS instance type c4.4xlarge. This EC2 instance type has 8 CPU and 32 GB memory. Storage is EBS-only and has a dedicated EBS bandwidth of 2,000 Mbps.

Tests were performed with structured and unstructured data with two different event sizes—256 and 1,024 bytes. The following is a table showing average performance metrics:

Figure 3 – Test results

Note: These performance results are a guideline as different configurations and environments may vary results.

Want to Send Feedback or Enhancement Request?

For those interested in assisting in the next release, this project is hosted on github. Feel free to open Pull Requests and raise issues there. For technical feedback and questions, please reach out to Splunk directly here or raise any questions on Splunk Answers.

Don Tregonning
Solutions Architect
Splunk

Elias Haddad
Principal Product Manager – Splunk Data Collection
Splunk

----------------------------------------------------
Thanks!
Don Tregonning

Related Articles

Announcing the General Availability of Splunk POD: Unlock the Power of Your Data with Ease
Platform
2 Minute Read

Announcing the General Availability of Splunk POD: Unlock the Power of Your Data with Ease

Splunk POD is designed to simplify your on-premises data analytics, so you can focus on what really matters: making smarter, faster decisions that drive your business forward.
Introducing the New Workload Dashboard: Enhanced Visibility, Faster Troubleshooting, and Deeper Insights
Platform
3 Minute Read

Introducing the New Workload Dashboard: Enhanced Visibility, Faster Troubleshooting, and Deeper Insights

Announcing the general availability of the new workload dashboard – a modern and intuitive dashboard experience in the Cloud Monitoring Console app.
Leading the Agentic AI Era: The Splunk Platform at Cisco Live APJ
Platform
5 Minute Read

Leading the Agentic AI Era: The Splunk Platform at Cisco Live APJ

The heart of our momentum at Cisco Live APJ is our deeper integration with Cisco, culminating in the Splunk POD and new integrations, delivering unified, next-generation data operations for every organization.
Dashboard Studio: Token Eval and Conditional Panel Visibility
Platform
4 Minute Read

Dashboard Studio: Token Eval and Conditional Panel Visibility

Dashboard Studio in Splunk Cloud Platform can address more complex use cases with conditional panel visibility, token eval, and custom visualizations support.
Introducing Resource Metrics: Elevate Your Insights with the New Workload Dashboard
Platform
4 Minute Read

Introducing Resource Metrics: Elevate Your Insights with the New Workload Dashboard

Introducing Resource Metrics in Workload Dashboard (WLD) – a modern and intuitive monitoring experience in the Cloud Monitoring Console (CMC) app.
Powering AI Innovation with Splunk: Meet the Cisco Data Fabric
Platform
3 Minute Read

Powering AI Innovation with Splunk: Meet the Cisco Data Fabric

The Cisco Data Fabric brings AI-centric advancements to the Splunk Platform, seamlessly connecting knowledge, business, and machine data.
Remote Upgrader for Windows Is Here: Simplifying Fleet-Wide Forwarder Upgrades
Platform
3 Minute Read

Remote Upgrader for Windows Is Here: Simplifying Fleet-Wide Forwarder Upgrades

Simplify fleet-wide upgrades of Windows Universal Forwarders with Splunk Remote Upgrader—centralized, signed, secure updates with rollback, config preservation, and audit logs.
Dashboard Studio: Spec-TAB-ular Updates
Platform
3 Minute Read

Dashboard Studio: Spec-TAB-ular Updates

Splunk Cloud Platform 10.0.2503 includes a number of enhancements related to tabbed dashboards, trellis for more charts, and more!
Introducing Edge Processor for Splunk Enterprise: Data Management on Your Premises
Platform
2 Minute Read

Introducing Edge Processor for Splunk Enterprise: Data Management on Your Premises

Announcing the introduction of Edge Processor for Splunk Enterprise 10.0, designed to help customers achieve greater efficiencies in data transformation and improved visibility into data in motion.