<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/">
    <channel>
        <title>Elastic Security Labs - Articles by Melissa Alvarez</title>
        <link>https://www.elastic.co/security-labs</link>
        <description>Trusted security news &amp; research from the team at Elastic.</description>
        <lastBuildDate>Thu, 05 Mar 2026 22:21:01 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>https://github.com/jpmonette/feed</generator>
        <image>
            <title>Elastic Security Labs - Articles by Melissa Alvarez</title>
            <url>https://www.elastic.co/security-labs/assets/security-labs-thumbnail.png</url>
            <link>https://www.elastic.co/security-labs</link>
        </image>
        <copyright>© 2026. Elasticsearch B.V. All Rights Reserved</copyright>
        <item>
            <title><![CDATA[Detect domain generation algorithm (DGA) activity with new Kibana integration]]></title>
            <link>https://www.elastic.co/security-labs/detect-domain-generation-algorithm-activity-with-new-kibana-integration</link>
            <guid>detect-domain-generation-algorithm-activity-with-new-kibana-integration</guid>
            <pubDate>Wed, 17 May 2023 00:00:00 GMT</pubDate>
            <description><![CDATA[We have added a DGA detection package to the Integrations app in Kibana. In a single click, you can install and start using the DGA model and associated assets, including ingest pipeline configurations, anomaly detection jobs, and detection rules.]]></description>
            <content:encoded><![CDATA[<p>Searching for a way to help protect your network from potential domain generation algorithm (DGA) attacks? Look no further — a DGA detection package is now available in the Integrations app in Kibana.</p>
<p>In a single click, users can install and start using the DGA model and associated assets, including ingest pipeline configurations, anomaly detection jobs, and detection rules. Read on for step-by-step instructions on installing and fully enabling the DGA package.</p>
<p>[Related article: <a href="https://www.elastic.co/blog/automating-security-protections-rapid-response-to-malware">Automating the Security Protections rapid response to malware</a>]</p>
<h1>What is a DGA?</h1>
<p>A DGA is a technique employed by many malware authors to ensure that infection of a client machine evades defensive measures. The goal of this technique is to hide the communication between an infected client machine and the command and control (C &amp; C or C2) server by using hundreds or thousands of randomly generated domain names, of which one will ultimately resolve to the IP address of a C &amp; C server.</p>
<p>To more easily visualize what’s occurring in a DGA attack, imagine for a moment you’re a soldier on a battlefield. Like many soldiers, you have communication gear that uses radio frequencies for communication. Your enemy may try to disrupt your communications by jamming your radio frequencies. One way to devise a countermeasure for this is by frequency hopping — using a radio system that changes frequencies very quickly during the course of a transmission. To the enemy, the frequency changes appear to be random and unpredictable, so they are hard to jam.</p>
<p>DGAs are like a frequency-hopping communication channel for malware. They change domains so frequently that blocking the malware’s C2 communication channel becomes infeasible by means of DNS domain name blocking. There are simply too many randomly generated DNS names to successfully identify and block them.</p>
<p>This technique emerged in the world of malware with force in 2009, when the “Conficker” worm began using a very large number of randomly generated domain names for communication. The worm’s authors developed this countermeasure after a consortium of security researchers interrupted the worm’s C2 channel by shutting down the DNS domains it was using for communication. DNS mitigation was also performed in the case of the 2017 WannaCry ransomware global outbreak.</p>
<h1>Getting started</h1>
<p>We have released the model and the associated assets — including the pipelines, anomaly detection configurations, and detection rules — to the Integrations app in Kibana as of 8.0. We will be maintaining this format moving forward.</p>
<p>If you don’t have an Elastic Cloud cluster but would like to start experimenting with the released ProblemChild package, you can start a <a href="https://cloud.elastic.co/registration">free 14-day trial</a> of Elastic Cloud.</p>
<p>We will now look at the steps to get DGA up and running in your environment in a matter of minutes using the released DGA package.</p>
<h3>Step 1: Installing the package assets</h3>
<p>In Kibana, the Integrations app now includes the DGA detection package. To install the assets, click the <strong>Install DGA assets</strong> button under the <strong>Settings</strong> tab. This will install all of the artifacts necessary to use the DGA model to generate alerts when DGA activity is detected in your network data.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/blog-elastic-DGA-1.png" alt="" /></p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/blog-elastic-DGA-2.jpg" alt="" /></p>
<p>Once installation is complete, you can navigate to <strong>Stack Management &gt; Ingest Pipelines</strong> and see that the <strong><code>&lt;version-number&gt;-ml\_dga\_ingest\_pipeline</code></strong> has been installed and can now be used to enrich incoming ingest data. The ingest pipeline leverages the <strong><code>&lt;version-number&gt;-ml\_dga\_inference\_pipeline</code></strong> to do this.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/blog-elastic-DGA-3.png" alt="" /></p>
<p>Similarly, the installed DGA model can now be seen in <strong>Machine Learning &gt; Model Management &gt; Trained Models</strong>.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/blog-elastic-DGA-4.jpg" alt="" /></p>
<h3>Step 2: Enriching your data</h3>
<p>Now you are ready to ingest your data using the ingest pipeline. The supervised model will analyze and enrich incoming data containing DNS events with a DGA score.</p>
<p>This pipeline is designed to work with data containing DNS events — such as <a href="https://www.elastic.co/beats/packetbeat">packetbeat</a> data — which contain these ECS fields: dns.question.name and dns.question.registered_domain. You can add the installed ingest pipeline to an Elastic beat by adding a simple <a href="https://www.elastic.co/guide/en/elasticsearch/reference/current/ingest.html#pipelines-for-beats">configuration setting</a>.</p>
<p>If you already have an ingest pipeline associated with your indices, you can use a <a href="https://www.elastic.co/guide/en/elasticsearch/reference/current/pipeline-processor.html">pipeline processor</a> to integrate the DGA ingest pipeline into your existing pipeline.</p>
<p>You will also want to add the following mappings to the beat you chose:</p>
<pre><code>{
  &quot;properties&quot;: {
    &quot;ml_is_dga&quot;: {
      &quot;properties&quot;: {
        &quot;malicious_prediction&quot;: {
          &quot;type&quot;: &quot;long&quot;
        },
        &quot;malicious_probability&quot;: {
          &quot;type&quot;: &quot;float&quot;
        }
      }
    }
  }
}
</code></pre>
<p>You can do this under <strong>Stack Management &gt; Index Management &gt; Component Templates.</strong> Templates that can be edited to add custom components will be marked with a <em>@custom</em> suffix. Edit the <em>@custom</em> component template for your Elastic beat by pasting the above JSON blob in the <strong>Load JSON</strong> flyout.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/Screen_Shot_2022-07-29_at_8.37.43_AM.jpeg" alt="" /></p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/Screen_Shot_2022-07-29_at_8.38.11_AM.jpeg" alt="" /></p>
<p>You should now see that the model enriches incoming DNS events with the following fields:</p>
<ul>
<li>
<p><strong>Ml_is_dga.malicious_prediction:</strong> A value of “1” indicates the DNS domain is predicted to be the result of malicious DGA activity. A value of “0” indicates it is predicted to be benign.</p>
</li>
<li>
<p><strong>Ml_is_dga.malicious_probability:</strong> A probability score, between 0 and 1, that the DNS domain is the result of malicious DGA activity.</p>
</li>
</ul>
<p>If you want an immediate way to test that the ingest pipeline is working as expected with your data, you can use a few sample documents with the <a href="https://www.elastic.co/guide/en/elasticsearch/reference/current/simulate-pipeline-api.html">simulate pipeline API</a> and confirm you see the <strong>ml_is_dga</strong> fields.</p>
<h3>Step 3: Running anomaly detection</h3>
<p>The package includes a pre-configured anomaly detection job. This machine learning (ML) job examines the DGA scores produced by the supervised DGA model and looks for anomalous patterns of unusually high scores for a particular source IP address. These events are assigned an anomaly score.</p>
<p>To run this job on your enriched data, go to <strong>Machine Learning &gt; Anomaly Detection</strong>. When you create a job using the job wizard, you should see an option to Use preconfigured jobswith a card for DGA. After selecting the card, you will see the pre-configured anomaly detection job that can be run. Note this job is only useful for indices that have been enriched by the ingest pipeline.</p>
<h3>Step 4: Enabling the rules</h3>
<p>To maximize the benefit of the DGA framework, activate the installed detection rules. They are triggered when certain conditions for the supervised model or anomaly detection job are satisfied. The complete list of the installed rules can be found in the <strong>Overview</strong> page of the package itself or in the latest experimental detections <a href="https://github.com/elastic/detection-rules/releases/tag/ML-experimental-detections-20211130-7">release</a>.</p>
<p>To fully leverage the included preconfigured anomaly detection job, enable the complementary rule: <em>Potential DGA Activity.</em> This will create an anomaly-based alert in the detection page in the security app.</p>
<p>The preconfigured anomaly detection job and complementary rule are both available in the detection rules repo <a href="https://github.com/elastic/detection-rules/releases">releases</a>. To enable and use the installed rules, navigate to <strong>Security &gt; Rules</strong> and select <em>Load Elastic prebuild rules and timeline templates</em>.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/blog-elastic-DGA-5.jpg" alt="" /></p>
<h1>Get in touch</h1>
<p>We’d love for you to try out ProblemChild and give us feedback as we work on adding new capabilities to it. If you run into any issues during the process, please reach out to us on our <a href="https://ela.st/slack">community Slack channel</a>, <a href="https://discuss.elastic.co/c/security">discussion forums</a>, or even our <a href="https://github.com/elastic/detection-rules">open detections repository</a>.</p>
<p>You can always experience the latest version of <a href="https://www.elastic.co/elasticsearch/service">Elasticsearch Service</a> on Elastic Cloud and follow along with this blog to set up the ProblemChild framework in your environment for your Windows process event data. And take advantage of our <a href="https://www.elastic.co/training/elastic-security-quick-start">Quick Start training</a> to set yourself up for success. Start your <a href="https://cloud.elastic.co/registration">free trial of Elastic Cloud</a> today to get access to the platform. Happy experimenting!</p>
]]></content:encoded>
            <category>security-labs</category>
            <enclosure url="https://www.elastic.co/security-labs/assets/images/detect-domain-generation-algorithm-activity-with-new-kibana-integration/library-branding-elastic-stack-midnight-1680x980-no-logo.jpg" length="0" type="image/jpg"/>
        </item>
        <item>
            <title><![CDATA[Detecting Living-off-the-land attacks with new Elastic Integration]]></title>
            <link>https://www.elastic.co/security-labs/detecting-living-off-the-land-attacks-with-new-elastic-integration</link>
            <guid>detecting-living-off-the-land-attacks-with-new-elastic-integration</guid>
            <pubDate>Wed, 01 Mar 2023 00:00:00 GMT</pubDate>
            <description><![CDATA[We added a Living off the land (LotL) detection package to the Integrations app in Kibana. In a single click, you can install and start using the ProblemChild model and associated assets including anomaly detection configurations and detection rules.]]></description>
            <content:encoded><![CDATA[<p>It is becoming more common that adversary attacks consist of more than a standalone executable or script. Advanced attacker techniques, like “living off the land” (LotL) that appear normal in isolation become more suspicious when observed in a parent-child context. If you are running Windows in your environment, it is important to have a system for detecting these types of attacks. Traditional heuristic-based detections, though effective in detecting a single event, often fail to generalize across a multi-step attack. At Elastic we have trained a Living off the Land classifier, anomaly detection jobs and security detection rules to help our security professionals discover LotL attacks.</p>
<p>With the advent of <a href="https://www.elastic.co/integrations/">Integration packages</a> in the Elastic stack we can now deliver the full, customizable package that includes the LotL classification model, anomaly detection job configurations, detection rules, and inference pipelines to make it easier to install and get up and running the entire end-to-end data pipeline from collecting windows events to alerting on potential Lotl attacks. We will walk you through how we set it up so you can try it yourself.</p>
<h1>ProblemChild: Recap</h1>
<p>In an earlier blog post, we talked about how to use<a href="https://www.elastic.co/blog/problemchild-generate-alerts-to-detect-living-off-the-land-attacks">the detection rules repository command line interface (CLI), to set up the ProblemChild framework and get it up and running in your environment</a>. We have now added a <a href="https://docs.elastic.co/integrations/problemchild">Living off the land (LotL) detection package</a> to the Integrations app in Kibana. In a single click, you can install and start using the ProblemChild model and associated assets including anomaly detection configurations and detection rules.</p>
<p>As outlined in the <a href="https://www.elastic.co/blog/problemchild-generate-alerts-to-detect-living-off-the-land-attacks">previous blog</a>, ProblemChild is a framework built using the Elastic Stack to detect LotL activity. LotL attacks are generally tricky to detect, given that attackers leverage seemingly benign software already present in the target environment to fly under the radar. The lineage of processes spawned in your environment can provide a strong signal in the event of an ongoing attack.</p>
<p>The supervised machine learning (ML) component of ProblemChild leverages process lineage information present in your Windows process event metadata to classify events as malicious or benign using <a href="https://www.elastic.co/guide/en/machine-learning/current/ml-dfa-classification.html#ml-inference-class">Inference</a> at the time of ingest. Anomaly detection is then applied to detect rare processes among those detected as malicious by the supervised model. Finally, detection rules alert on rare parent-child process activity as an indication of LotL attacks.</p>
<p>The sheer volume and variety of events seen in organizations poses a challenge for detecting LotL attacks using rules and heuristics, making an ML-based framework such as ProblemChild a great solution.</p>
<h2>Getting Started</h2>
<p>We have released the model and the associated assets - including the pipelines, anomaly detection configurations, and detection rules - to the Integrations app in Kibana as of 8.0. We will be maintaining this format moving forward.</p>
<p>If you don’t have an Elastic Cloud cluster but would like to start experimenting with the released ProblemChild package, you can start a <a href="https://cloud.elastic.co/registration">free 14-day trial</a> of Elastic Cloud.</p>
<p>We will now look at the steps to get ProblemChild up and running in your environment in a matter of minutes using the released Living off the land (LotL) detection package.</p>
<h3>Step 1: Installing the package assets</h3>
<p>In Kibana, the Integrations app now includes the LotL Attack Detection package. To install the assets, click the <code>Install LotL Attack Detection assets</code> button under the <code>Settings</code> tab.</p>
<p>This will install all of the artifacts necessary to use the ProblemChild model to generate alerts when LotL activity is detected in your environment.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-living-off-the-land-attack-1.png" alt="To install the assets, click the Install LotL Attack Detection assets button under the Settings tab." /></p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-detecting-lotl-attacks-2.png" alt="To install the assets, click the Install LotL Attack Detection assets button under the Settings tab." /></p>
<p>Once installation is complete, you can navigate to <strong>Stack Management &gt; Ingest Pipelines</strong> and see that the <strong><code>&lt;version-number&gt;-problem\_child\_ingest\_pipeline</code></strong> has been installed and can now be used to enrich incoming ingest data. The ingest pipeline leverages the <strong><code>&lt;version-number&gt;-problem\_child\_inference\_pipeline</code></strong> in order to do this.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-detecting-lotl-attacks-3.png" alt="Once installation is complete, you can navigate to Stack Management &gt; Ingest Pipelines and see that the &lt;version-number&gt;-problem_child_ingest_pipeline has been installed and can now be used to enrich incoming ingest data." /></p>
<p>Similarly, the installed ProblemChild model can now be seen in <strong>Machine Learning &gt; Model Management &gt; Trained Models</strong></p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-detecting-lotl-attacks-4.jpg" alt="Similarly, the installed ProblemChild model can now be seen in Machine Learning &gt; Model Management &gt; Trained Models" /></p>
<h3>Step 2: Enriching your data</h3>
<p>Now you are ready to ingest your data using the ingest pipeline. This will enrich your incoming data with predictions from the machine learning model.</p>
<p>This pipeline is designed to work with Windows process event data such as <a href="https://www.elastic.co/downloads/beats/winlogbeat">Winlogbeat data</a>. You can add the installed ingest pipeline to an Elastic beat by adding a simple <a href="https://www.elastic.co/guide/en/elasticsearch/reference/current/ingest.html#pipelines-for-beats">configuration setting</a>.</p>
<p>If you already have an ingest pipeline associated with your indices, you can use a <a href="https://www.elastic.co/guide/en/elasticsearch/reference/current/pipeline-processor.html">pipeline processor</a> to integrate the ProblemChild ingest pipeline into your existing pipeline.</p>
<p>You will also want to add the following mappings to the Elastic beat you chose:</p>
<pre><code>{
  &quot;properties&quot;: {
    &quot;problemchild&quot;: {
      &quot;properties&quot;: {
        &quot;prediction&quot;: {
          &quot;type&quot;: &quot;long&quot;
        },
        &quot;prediction_probability&quot;: {
          &quot;type&quot;: &quot;float&quot;
        }
      }
    },
    &quot;blocklist_label&quot;: {
      &quot;type&quot;: &quot;long&quot;
    }
  }
}

</code></pre>
<p>You can do this under <strong>Stack Management &gt; Index Management &gt; Component Templates.</strong> Templates that can be edited to add custom components will be marked with a <em>@custom</em> suffix. Edit the <em>@custom</em> component template for your Elastic beat by pasting the above JSON blob in the <strong>Load JSON</strong> flyout.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/Screen_Shot_2022-07-29_at_8.13.52_AM.jpeg" alt="" /></p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/Screen_Shot_2022-07-29_at_8.14.10_AM.jpeg" alt="" /></p>
<p>You should now see that the model enriches incoming Windows process events with the following fields:</p>
<p><strong>problemchild.prediction</strong></p>
<ul>
<li>A value of 1 indicates that the event is predicted to be malicious and a value of “0” indicates that the event is predicted to be benign.</li>
</ul>
<p><strong>prediction_probability</strong></p>
<ul>
<li>A value between 0 and 1 indicating the confidence of the model in its prediction. The higher the value, the higher the confidence.</li>
</ul>
<p><strong>blocklist_label</strong></p>
<ul>
<li>A value of 1 indicates that the event is malicious because one or more terms in the command line arguments matched a blocklist.</li>
</ul>
<p>If you want an immediate way to test that the ingest pipeline is working as expected with your data, you can use a few sample documents with the <a href="https://www.elastic.co/guide/en/elasticsearch/reference/current/simulate-pipeline-api.html">simulate pipeline API</a> and confirm you see the <strong>problemchild</strong> fields.</p>
<h3>Step 3: Running anomaly detection</h3>
<p>The package includes several preconfigured anomaly detection jobs. These jobs enable you to find the rarest events among those detected as malicious by the supervised model in order to decide which events require immediate attention from your analysts.</p>
<p>To run these jobs on your enriched data, go to <strong>Machine Learning &gt; Anomaly Detection</strong>. When you create a job using the job wizard, you should see an option to Use preconfigured jobs with a card for LotL Attacks. After selecting the card, you will see several preconfigured anomaly detection jobs that can be run. Note these jobs are only useful for indices that have been enriched by the ingest pipeline.</p>
<h3>Step 4: Enabling the rules</h3>
<p>To maximize the benefit of the ProblemChild framework, activate the installed detection rules. They are triggered when certain conditions for the supervised model or anomaly detection jobs are satisfied. The complete list of the installed rules can be found in the <strong>Overview</strong> page of the package itself or in the latest experimental detections <a href="https://github.com/elastic/detection-rules/releases/tag/ML-experimental-detections-20211130-7">release</a>.</p>
<p>In order to enable and use the installed rules, you can navigate to <strong>Security &gt; Rules</strong> and select <code>_Load Elastic prebuild rules and timeline templates</code>_.</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-detecting-lotl-attacks-5.png" alt="In order to enable and use the installed rules, you can navigate to Security &gt; Rules and select Load Elastic prebuild rules and timeline templates." /></p>
<p>Note that there are search rules as well as ML job rules. The search rules are triggered by the supervised model, for example this rule:</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-detecting-lotl-attacks-6.jpg" alt="The above rule matches on any Windows process event for which the supervised model or its blocklist has a prediction value of 1 (malicious)." /></p>
<p>The above rule matches on any Windows process event for which the supervised model or its blocklist has a prediction value of 1 (malicious).</p>
<p>The ML job rules are triggered by anomalies found by the anomaly detection jobs that you set up in Step 3 — for example, this rule:</p>
<p><img src="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/blog-elastic-detecting-lotl-attacks-6.jpg" alt="The above rule is triggered each time the anomaly detection job problem_child_rare_process_by_host detects an anomaly with an anomaly score greater than or equal to 75." /></p>
<p>The above rule is triggered each time the anomaly detection job problem_child_rare_process_by_host detects an anomaly with an anomaly score greater than or equal to 75.</p>
<h1>Summary</h1>
<p>As mentioned in the first blog post, the supervised ML component of ProblemChild is trained to predict a value of 1 (malicious) on processes or command line arguments that can be used for LotL attacks. This does not mean that everything that the supervised model predicts with a value 1 indicates LotL activity. The prediction value of 1 should be interpreted more as “this could be potentially malicious,” instead of “this is definitely LotL activity.”</p>
<p>The real beauty of ProblemChild is in the anomaly detection, wherein it surfaces rare parent-child process relationships from among the events the supervised model marked as suspicious. This not only helps in reducing the number of false positives, but also helps security analysts focus on a smaller, more targeted list for triage.</p>
<p>You could of course start with the search rules, which will alert directly on the results of the supervised model. If the number of alerts from these rules is manageable and you have the time and resources to drill into these alerts, you might not need to enable the anomaly detection jobs. However, if you then notice that these rules are producing too many alerts (which is usually the case in most large organizations), you may benefit from enabling the anomaly detection jobs and their corresponding rules.</p>
<h1>Get in touch with us</h1>
<p>We’d love for you to try out ProblemChild and give us feedback as we work on adding new capabilities to it. If you run into any issues during the process, please reach out to us on our <a href="https://ela.st/slack">community Slack channel</a>, <a href="https://discuss.elastic.co/c/security">discussion forums</a> or even our <a href="https://github.com/elastic/detection-rules">open detections repository</a>.</p>
<p>You can always experience the latest version of <a href="https://www.elastic.co/elasticsearch/service">Elasticsearch Service</a> on Elastic Cloud and follow along with this blog to set up the ProblemChild framework in your environment for your Windows process event data. And take advantage of our <a href="https://www.elastic.co/training/elastic-security-quick-start">Quick Start training</a> to set yourself up for success. Happy experimenting!</p>
]]></content:encoded>
            <category>security-labs</category>
            <enclosure url="https://www.elastic.co/security-labs/assets/images/detecting-living-off-the-land-attacks-with-new-elastic-integration/security-threat-hunting-incidence-response-1200x628.jpg" length="0" type="image/jpg"/>
        </item>
    </channel>
</rss>