DTRH: Scripting Anything and Reaping Data

cancel
Showing results for 
Show  only  | Search instead for 
Did you mean: 
Announcements

DTRH: Scripting Anything and Reaping Data

L0 Member

JEbrahimi_0-1622052573092.png

 

 

DTRH: Scripting Anything and Reaping Data

 

Overview

Customers are always asking for additional capabilities in the product and often times these feature request may come during a POC where having that capability can be the deciding factor in winning the deal. The feature request to delivered into the product can be a long cycle, often time you may be able to achieve the desired functionality through the scripting capabilities that exist today.

In this DTRH, we are going to cover a real request from a customer in a POC and walk you through the basics of script creation and showing how we use the data from the script and make it usable in the Cortex Data Lake.

 

Let’s start with the customer use case.  The customer already had some monitoring capabilities on the endpoint that would dump any PowerShell commands to a log file.  The wanted the capability to ingest the contents of that log file into the Cortex Data Lake with the additional goal of being able to search through the content via XQL.  Now if this were going to be a feature request it would most likely take considerable engineering effort to get it all done but let’s see how we can accomplish this in 25 lines of python code.

 

 

Creating the Script

There are 2 main components for creating usable scripts in XDR.

  1. The Script – The set of instructions that will run on the endpoint
  2. The UI – Setup in the UI to name, get arguments and other configuration

We will start taking a look at the script itself but there are some areas related to the UI that will dictate how we develop this script. Here is a look at the script in full:

 

JEbrahimi_1-1622052573097.png

 

 

First the entry point into function of script is set with ‘def run’, the entry point will be chosen and set when we add it to the UI.  Then you can have an arbitrary number of variables that are populated in the UI and passed to the script. In this case here we have 3: apiurl, apikey, path.

 

JEbrahimi_2-1622052573099.png

 

 

Now its best practice to have some error checking in your code, possibly checking the values of the variables and/or checking for possible exceptions in running. In this case, since we are dumping the contents of the file, one of the checks is to make sure the file exists and throw an error if it does not.

 

JEbrahimi_3-1622052573100.png

 

 

A script will then have a number of lines of code to complete the desired action(s) on the endpoint. For this script we are going to read in the contents of the file that we want to eventually insert into the Cortex Data Lake

 

JEbrahimi_4-1622052573100.png

 

 

For many scripts you may not need to even return any data back to the data lake so we could potentially return from the script at this point. In this case we are going to use another feature in the XDR console, the HTTP Event Collector, to pass data over that is stored in the data lake.  This is fairly straight forward but will require some additional setup in the UI to enable the collector and requires massaging the data that we are going to pass in as well.

 

One of the main items to consider when using the HTTP Event Collector is that when we post data to it, every single newline is a break and will create separate events in the XQL. This could be the desired usage but in our case we are reading in an entire file that we would like to stay as a whole single event so we have to deal with these newlines. Here is the code for this:

 

JEbrahimi_5-1622052573101.png

 

 

What this does is replace every actual newline (\n) with the word ‘<newline>’. This will then basically turn the file into one long string with ‘<newline>’ as a separator. Not only does this create the separator and allow you to upload as a single event but through XQL you could potentially reverse the process in the UI and substitute ‘<newline>’ back to something else.

The python requests library is a great addition to the language and allows for much easier web based transaction. The following code is something that would be valuable/reusable in other scripts that want to store data from the script into the data lake.

 

Authorization is done through the API key which is created by the HTTP Collector and set through the UI. We will show both of these steps later but for now its just important to note that an API key (apikey) was created and it is passed in to the script. In addition, the URL for the API (apiurl) that we need to send the request to is handled the same way.

 

JEbrahimi_6-1622052573101.png

 

 

The body builds the actual content of data going into data lake as you can see in the post request which contains the API URL, the needed headers and the body. There is a check for a 200 response code from the server to validate the data was delivered without issue.  Refer to the Requests Python Library for more information but knowing these few lines of code can make any script push data into the Cortex Data Lake.  

 

Connecting the Script To The UI

Now that we have the completed script, we will configure the UI to use it and fill in some of the pieces we discussed previously in more detail.  Let’s walk through this step by step.

 

  1. Navigate to: Response -> Action Center and select Available Script Library in the left-hand side pane
  2. Select  JEbrahimi_7-1622052573102.png 

     

    in the top right corner
  3. Drag the script of select Browse to import the scriptJEbrahimi_8-1622052573104.png
  4.  

    Fill in the General script definitions
    1. The Script Name will be used to run on the end point, so use a name that makes sense.
    2. Define the supported OS’s. Since Python is cross platform, many scripts can run across all platforms if it makes sense.
    3. A timeout can be set for the maximum number of seconds a script should run
  5. Fill in the Input section
    1. Run by entry point will let you choose from a drop down the function you wish to called to start the script.
    2. The variables you defined in the script will show up here and you can select the expected type. Some values can be number or Booleans but, in our case, all 3 values will be String
  6. I left the Output as Auto Detect for this script. This will probably work in most cases and since we are sending our data for searching in the data lake, this output will most likely only be used in error cases.
  7. Once finished hit the  JEbrahimi_10-1622052573105.png 

     

    button in the bottom right-hand corner.

Your script is now created and will show up in the script library with the information populated

 

JEbrahimi_11-1622052573108.png

 

 

Custom Collector for The Data Lake

This next step is around setting up the server to accept data from the script through a Custom Collector. It should be noted that for this script the goal was to get the data into the data lake. This step would not be needed if the data did not need to be ingested into the data lake.

 

  1. JEbrahimi_12-1622052573110.png

     

    Navigate toJEbrahimi_13-1622052573110.png

     

    -> Configuration and select Data Collection->Customer Collectors in the left had side bar
  2. Click to add a new HTTP Collector
  3. JEbrahimi_14-1622052573113.pngTypically, the custom collector is used to support ingesting data from products from other vendors but we can take advantage of it but it will mean some of the field names do not match up exactly as we fill in the collection details
    1. Name – Collector Name Seen in UI
    2. Vendor – In this case PAN was used
    3. Product – Used the name of the script action
    4. Compression – our data is uncompressed
    5. Log Format – Sending Tex

 

! It is important to note that the vendor and product name will be the name of dataset used by XQL. The overall dataset name will look like:<vendor>_<product>_raw

 

  1. JEbrahimi_15-1622052573116.png

     

    Click Save & Generate Token
  2. Copy the Token

! The API Token will pop up and must be copied. Once this window is closed there is no way to access this token anymore and if not saved a new collector would have to be created.

 

       3. Once the collector is created there is a clickable link to copy the API URL that will be used to post the data to. Copy the link.

 

Usage

 

At this point everything has been setup and we can run the script. These steps will walk through a sample run. We have the API URL and API Key from the steps above, the two options that need to be added are what file we want to dump and on which endpoints should we dump this file from.

 

  1. JEbrahimi_17-1622052573119.png

     

    Navigate to Response->Action Center and select Available Scripts Library from the left hand side navigation
  2. Right click on the DumpFileContents script and select Run
  3. This will load the script and present you with the variables that you wish to pass into the script
    1. Path – This is the FULL PATH to the file you want the contents sent to the datalake.
    2. Apiurl – This is the full URL to the HTTP Collector we created in the Custom Collector steps
    3. Apikey – This is the key we copied when we created the HTTP Collector

JEbrahimi_19-1622052573123.png      4. Click next and then you will be presented the endpoint you would like to run the script on. You can select any N number of endpoints you would like this to run on but remember that the file would need to exist on the selected systems.

 

 

 

      5. Click JEbrahimi_20-1622052573123.png Run in the bottom right corned to execute the script on selected endpoints

      6. In the Action Center click on All Actions to see the progress of the execution of the script. Once it has run on all endpoints you will see the Status if updated as completed.

 

JEbrahimi_21-1622052573125.png

 

 

Results

Now that the script has run, the results will be stored in the Cortex Data Lake and we can query those results through XQL.

  1. Navigate to Investigation -> Query Builder and select XQL Search
  2. In the XQL query window we a going to select the dataset where the data was stored in.

JEbrahimi_22-1622052573126.png

! Remember the dataset name will follow the format <vendor>_<product>_raw. These values where set when we create the HTTP Collector

 

  1. The contents of the file are stored in the _raw field. We can limit this field and rename it for easier access
  2. The results how up in the query window, this is a sample 3 line file and you can see where the <newline> was substituted for the actual newline (\n) as previously discussed in this doc.

 

JEbrahimi_0-1622073721473.png

 

 

 

Tips/Tricks/Hacks

One thing I learned in running this script over and over is that it is tedious to have to enter the API URL and API Key everytime when these values do not change. Really only the Path variable is something that you should have to populate each run. To achieve this result, you can hard code the APIKEY and APIURL into the script. This does put these credentials in the clear on the file system but they are only useful for being able to post to the HTTP Collector, but it is still something you should consider.

By adding the apikey and apiurl here in the script, we also omit having those variables passed into the run function as seen below.

 

JEbrahimi_24-1622052573128.png

 

 

Conclusion

In only 25 lines of code, we were able to achieve new capabilities of the product and have the data stored in the Cortex Data Lake for searching. When a customer/prospect asks for something outside the current features it is worth looking at scripting to see if this can still be completed.

 

5 REPLIES 5

L4 Transporter

Very informative, thank you! 

L4 Transporter

Amazing how with just a few lines of code we can create new functionalities at Cortex XDR. 

Simple, functional, fast and effective.  

L3 Networker

This requires a Pro Pre TB license?

Hi @NathanBradley 

just excuting scripts per-se just requires a pro per endpoint license. Please check the following doc to see also other requirements as agent version, permissions on the hub, etc.

https://docs.paloaltonetworks.com/cortex/cortex-xdr/cortex-xdr-pro-admin/investigation-and-response/...

 

Another matter is what you want to do with the scripts, for example depending on the license you have you might be able to ingest 3rd party logs or other logs apart from the endpoint logs or even the enhanced data collection.... and for that you might need the Pro Per TB. So please keep in mind the whole picture of what you really want to achieve 

KR,
Luis

 

Yes i should have been more specific, creating, ingesting, and storing the data collected from the http event collector requires per tb?

  • 6988 Views
  • 5 replies
  • 9 Likes
Like what you see?

Show your appreciation!

Click Like if a post is helpful to you or if you just want to show your support.

Click Accept as Solution to acknowledge that the answer to your question has been provided.

The button appears next to the replies on topics you’ve started. The member who gave the solution and all future visitors to this topic will appreciate it!

These simple actions take just seconds of your time, but go a long way in showing appreciation for community members and the LIVEcommunity as a whole!

The LIVEcommunity thanks you for your participation!