Advanced Link Analysis: Part 2 - Implementing Link Analysis

Link analysis, which is a data analysis approach used to discover relationships and connections between data elements and entities, has many use cases including cybersecurity, fraud analytics, crime investigations, and finance. In my last post, "Advanced Link Analysis: Part 1 - Solving the Challenge of Information Density," I covered how advanced link analysis can be used to solve the challenge of information density. I also introduced the Sigbay Link Analysis app to help you accelerate uncover actionable insights faster in Splunk Enterprise and Splunk Cloud.

In this post, I'll show you a step-by-step process to building the dashboard with Sigbay Link Analysis visualization app from scratch.

There are few key differences between how Sigbay Link Analysis viz functions compare to other Splunk visualization applications:

1. Sigbay Link Analysis app does not use predefined SPL query like other Splunk visualization tools. In its current implementation, visualization is powered by the data model, defined within settings and optional timeframe defined within a custom filter.

2. It is powered by an Accelerated Data Model (ADM) defined within settings. It does not pull data from raw indexes.

3. When a user interacts with the visualization, such as clicking on the nodes, it dynamically generates custom TSTATS queries to ADM. ADM allows for quick evaluation of important metrics and statistics for every node. Dynamic query generation gives the user greater flexibility to focus on investigating the data and discovery of insights instead of struggling with custom queries and the need to constantly update the dashboard.

4. Visualization contains custom filter area where you can define the time range, such as:

earliest=$timeframe.earliest$ latest=$timeframe.latest$

This allows the user to pass time tokens and other tokens to define initial data coverage for visualization.

5. Visualization populates few tokens, such as:

$token_la_filter$

This token contains a fragment of SPL filter query representing nodes that user is currently interacting with. This is suitable for driving other panels within the same dashboard, depending on values clicked in Link Analysis.

To implement the first dashboard with link analysis visualization, we will follow this process:

  1. Create custom, basic Splunk app
  2. Ingest data
  3. Create an Accelerated Data Model
  4. Build an actual dashboard with Link Analysis

I've recorded the full demo video which you can follow to accomplish all steps here:

YouTube video player

Here is the copy of the anonymized dataset (web_traffic2.csv) used for this demo.

To follow the step-by-step guidance in the video above, make sure that you already have Splunk Cloud and the dataset example downloaded locally, as well as the Sigbay Link Analysis app installed (if you don’t, go to “Manage Apps” -> [Install App from File]).

Creating a custom basic Splunk app

  1. Navigate to Manage Apps
  2. Click [Create app] button. Fill in dialog like this:

  3. Click [Save] button.

This will create a blank app as a placeholder for us to build the dashboards.

Start Ingesting Data

  1. Make sure "Web Traffic" app is currently viewed.
  2. Navigate to: Settings -> [Add Data] -> [Upload files from my computer]
  3. Click on [Select file], Navigate to location of web_traffic2.csv file on your local computer
  4. Click [Next] button.
  5. You'll be presented with table with data.
  6. Click [Save As] to set the source type for your data. It is a good practice to assign customer source type to your data in case you want to customize it away from "csv" defaults later on.
  7. Follow conventions on the image below. Click [Save]
  8. Click [Next]

  1. On the next page "Input Settings" Click on "Create a new index" and for "Index Name" type: "webt"
  2. Press [Save]
  3. Press [Review] button. Then [Submit] button
  4. The file upload will continue. Once finished - you may click on [Start Searching] to see raw data

Creating an Accelerated Data Model

  1. Navigate to Settings -> Data models
  2. Click [New Data Model]
  3. Enter "Title" = "webt". Make sure App = "Web Traffic" is selected.
  4. Press [Create] button. Data model creation dialog will open.
  5. Click [Add Dataset] -> [Root Event] and fill in dialog as follows:

  1. Press [Save] to create initial data model
  2. Click [Add Field], [Auto-Extracted]
  3. The "Add Auto-Extracted Field" dialog will open.
  4. Check the following fields at least (you may select more for experimenting):
    Country, src_ip, username_tried, logged_in, accept_language, http_method, status, http_user_agent, uri_path

  1. Press [Save]
  2. Press [Edit], [Edit Permissions], fill settings as follows. Press [Save] when done.

This Data Model Acceleration process will begin and it may take up to an hour depending on the speed of your server. You can monitor this process until 100% completion by clicking "Update" link at data model manager panel here:

The easiest way to build a dashboard with Sigbay Link Analysis app is to start with a simple, dummy SPL search and then select an appropriate visualization and apply configuration:

Fill the configuration form as shown on the image.

For "Fields" input, enter:

Country|src_ip|username_tried|logged_in|accept_language|http_method|status|http_user_agent|uri_path

For "Aggregates" input, enter:

count as Events|sum(bytes_in) as BytesIn|dc(username_tried) as Usernames

Fields of interest are separated with a "|" character. Aggregate functions of interest can also be separated with a “|” character.

Once you filled in the dialog - click "Save As" -> "Dashboard Panel", give dashboard name - "Web Traffic Analysis" and press [Save]:

Once the dashboard is created - click [View Dashboard] button.

That's how the dashboard should look like:

At this point, you may slightly customize the dashboard by pressing the [Edit] button. I would prefer to make visualization slightly bigger (taller) and switch the dashboard theme to "dark" mode:

We may also start clicking on nodes or selecting multiple nodes (Ctrl+Click on node will select it) and executing more complex filters.

Here we selected Country=Russia and status=500 - to investigate all traffic originating from Russia and causing 500 errors:

Now that we build the fully functioning dashboard you may start doing investigations.

Full video is here:

Here are short examples:

Analysing all successful logins (where logged_in flag = 1):



Selecting all traffic that caused error codes >=400. Link analysis allows to use math expressions to select numerical nodes efficiently:

Analyzing all traffic accessing administrative accounts. Link analysis allows to dynamically select matching nodes by typing partial value in "global search" input within the sidebar:

Feel free to contact me with any questions at the link in my profile.

Related Articles

Predicting Cyber Fraud Through Real-World Events: Insights from Domain Registration Trends
Security
12 Minute Read

Predicting Cyber Fraud Through Real-World Events: Insights from Domain Registration Trends

By analyzing new domain registrations around major real-world events, researchers show how fraud campaigns take shape early, helping defenders spot threats before scams surface.
When Your Fraud Detection Tool Doubles as a Wellness Check: The Unexpected Intersection of Security and HR
Security
4 Minute Read

When Your Fraud Detection Tool Doubles as a Wellness Check: The Unexpected Intersection of Security and HR

Behavioral analytics can spot fraud and burnout. With UEBA built into Splunk ES Premier, one data set helps security and HR reduce risk, retain talent, faster.
Splunk Security Content for Threat Detection & Response: November Recap
Security
1 Minute Read

Splunk Security Content for Threat Detection & Response: November Recap

Discover Splunk's November security content updates, featuring enhanced Castle RAT threat detection, UAC bypass analytics, and deeper insights for validating detections on research.splunk.com.
Security Staff Picks To Read This Month, Handpicked by Splunk Experts
Security
2 Minute Read

Security Staff Picks To Read This Month, Handpicked by Splunk Experts

Our Splunk security experts share their favorite reads of the month so you can follow the most interesting, news-worthy, and innovative stories coming from the wide world of cybersecurity.
Behind the Walls: Techniques and Tactics in Castle RAT Client Malware
Security
10 Minute Read

Behind the Walls: Techniques and Tactics in Castle RAT Client Malware

Uncover CastleRAT malware's techniques (TTPs) and learn how to build Splunk detections using MITRE ATT&CK. Protect your network from this advanced RAT.
AI for Humans: A Beginner’s Field Guide
Security
12 Minute Read

AI for Humans: A Beginner’s Field Guide

Unlock AI with the our beginner's field guide. Demystify LLMs, Generative AI, and Agentic AI, exploring their evolution and critical cybersecurity applications.
Splunk Security Content for Threat Detection & Response: November 2025 Update
Security
5 Minute Read

Splunk Security Content for Threat Detection & Response: November 2025 Update

Learn about the latest security content from Splunk.
Operation Defend the North: What High-Pressure Cyber Exercises Teach Us About Resilience and How OneCisco Elevates It
Security
3 Minute Read

Operation Defend the North: What High-Pressure Cyber Exercises Teach Us About Resilience and How OneCisco Elevates It

The OneCisco approach is not about any single platform or toolset; it's about fusing visibility, analytics, and automation into a shared source of operational truth so that teams can act decisively, even in the fog of crisis.
Data Fit for a Sovereign: How to Consider Sovereignty in Your Digital Resilience Strategy
Security
5 Minute Read

Data Fit for a Sovereign: How to Consider Sovereignty in Your Digital Resilience Strategy

Explore how digital sovereignty shapes resilient strategies for European organisations. Learn how to balance control, compliance, and agility in your data infrastructure with Cisco and Splunk’s flexible, secure solutions for the AI era.