GCP Dataflow templatesedit

In this tutorial, you’ll learn how to ship logs directly from the Google Cloud Console with the Dataflow template for analyzing GCP Audit Logs in the Elastic Stack.

What you’ll learnedit

You’ll learn how to:

  • Export GCP audit logs through Pub/Sub topics and subscriptions.
  • Ingest logs using Google Dataflow and view those logs in Kibana.

Before you beginedit

Create a deployment using our hosted Elasticsearch Service on Elastic Cloud. The deployment includes an Elasticsearch cluster for storing and searching your data, and Kibana for visualizing and managing your data. For more information, see Spin up the Elastic Stack.

Step 1: Install the GCP integrationedit

You’ll start with installing the Elastic GCP integration to add prebuilt dashboards, ingest node configurations, and other assets that help you get the most of the GCP logs you ingest.

  1. Go to Integrations in Kibana and search for gcp.

    Kibana integrations
  2. Click the Elastic Google Cloud Platform (GCP) integration to see more details about it, then click Add Google Cloud Platform (GCP).

    GCP integration
  3. Click Save integration.

Step 2: Create a Pub/Sub topic and subscriptionedit

Before configuring the Dataflow template, create a Pub/Sub topic and subscription from your Google Cloud Console where you can send your logs from Google Operations Suite. There are three available filesets: audit, vpcflow, firewall. This tutorial covers the audit fileset.

  1. Go to the Logs Router page to configure GCP to export logs to a Pub/Sub topic. Use the search bar to find the page:

    Navigate to Logs Router page

    To set up the logs routing sink, click Create sink. Set sink name as monitor-gcp-audit-sink. Select the Cloud Pub/Sub topic as the sink service and Create new Cloud Pub/Sub topic named monitor-gcp-audit:

    Create Pub/Sub topic

    Finally, under Choose logs to include in sink, add logName:"cloudaudit.googleapis.com" (it includes all audit logs). Click create sink. It will look something like the following:

    Create logs routing sink
  2. Now go to the Pub/Sub page to add a subscription to the topic you just created. Use the search bar to find the page:

    GCP Pub/Sub

    To add a subscription to the monitor-gcp-audit topic click Create subscription:

    Create GCP Pub/Sub Subscription

    Set monitor-gcp-audit-sub as the Subscription ID and leave the Delivery type as pull:

    GCP Pub/Sub Subscription ID

    Finally, scroll down and click Create.

Step 3: Configure the Google Dataflow templateedit

After creating a Pub/Sub topic and subscription, go to the Dataflow Jobs page and configure your template to use them. Use the search bar to find the page:

GCP Dataflow Jobs

To create a job, click Create Job From Template. Set Job name as auditlogs-stream and select Pub/Sub to Elasticsearch from the Dataflow template dropdown menu:

GCP Dataflow Pub/Sub to Elasticsearch

Before running the job, fill in required parameters:

GCP Dataflow Required Parameters

For Cloud Pub/Sub subscription, use the subscription you created in the previous step. For Cloud ID and Base64-encoded API Key, use the values you got earlier. If you don’t have an Error output topic, create one like you did in the previous step.

After filling the required parameters, click Show Optinal Parameters and add audit as the log type parameter.

GCP Dataflow Optional Parameters

When you are all set, click Run Job and wait for Dataflow to execute the template, which takes a few minutes.

Finally, navigate to Kibana to see your logs parsed and visualized in the [Logs GCP] Audit dashboard.

GCP audit overview dashboard

Besides collecting audit logs from your Google Cloud Platform, you can also use Dataflow integrations to ingest data directly into Elastic from Google BigQuery and Google Cloud Storage.