Standard Fields
Panther's log analysis applies normalization fields to all log records
Panther's log analysis applies normalization fields (IPs, domains, etc) to all log records. These fields provide standard names for attributes across all data sources enabling fast and easy data correlation.
For example, each data source has a time that an event occurred, but each data source will likely not name the attribute the same, nor is it guaranteed that the associated time has a timezone consistent with other data sources.
The Panther attribute p_event_time
is mapped to each data source's corresponding event time and normalized to UTC. This way you can query over multiple data sources joining and ordering by p_event_time
to properly align and correlate the data despite the disparate schemas of each data source.
All appended standard fields begin with p_.
Required Fields
The fields below are appended to all log records:
Field Name | Type | Description |
|
| The type of log. |
|
| Unique id (UUID) for the row. |
|
| The associated event time for the log type is copied here and normalized to UTC.
Format: |
|
| The current time when the event was parsed, normalized to UTC.
Format: |
|
| The version of the schema used for this row. |
|
| The Panther generated internal id for the source integration. |
|
| The user supplied label for the source integration (may change if edited). |
|
| Available for S3 sources only, this field contains metadata of the file that this event originated from, including the bucket name and object key. |
If an event does not have a timestamp, then p_event_time
will be set to p_parse_time
, which is the time the event was parsed.
The p_source_id
and p_source_label
fields indicate where the data originated. For example, you might have multiple CloudTrail sources registered with Panther, each with a unique name (e.g., "Dev Accounts", "Production Accounts", "HR Accounts", etc.). These fields allow you to separate data based on the source, which is beneficial when configuring detections in Panther.
In addition, the fields below are appended to log records of all tables in the panther_rule_matches
database:
Field Name in panther_rule_matches | Type | Description |
|
| Id of alert related to row. |
|
| Time of alert creation related to row. |
| object | A JSON object returned from the rule's alert_context() function. |
|
| The severity level of the rule at the time of the alert. This could be different from the default severity as it can be dynamically set. |
|
| Time of last alert update related to row. |
|
| The id of the rule that generated the alert. |
|
| The error message if there was an error running the rule. |
|
| List of user defined rule reporting tags related to row. |
|
| The default severity of the rule. |
|
| List of user defined rule tags related to row. |
Core Fields
Panther Core Fields are in open beta starting with Panther version 1.86, and are available to all customers. Please share any bug reports and feature requests with your Panther support team.
Core Fields make up the Panther Unified Data Model (UDM). They normalize data from various sources into a consistent structure while maintaining its context. This makes Core Fields useful for searching and writing detections across log types.
The Panther UDM fields help define user and machine attributes. The user performing the action in the log (i.e. the actor) is represented as user
, while machines are represented as either source
or destination
.
Learn how to map fields in your custom log source schemas to Core Fields below, in Mapping Core Fields in custom log schemas. In certain cases, you may want to map one event field to both an Indicator (p_any) field, and a Core Field—learn more in Core Fields vs. Indicator Fields. Core Fields also differ from Data Models for detections; their differences are described below, in Core Fields vs. Data Models in Python detections.
The Panther-managed log types listed below have UDM field mappings configured:
Events ingested prior to the Panther UDM being enabled in your Panther instance will not contain Core Fields.
See the full list of Panther Core Fields (also known as UDM fields) below:
Field Name | Display Name | Description | Example |
---|---|---|---|
| Destination Address | Some event destination addresses are defined ambiguously. The event will sometimes list an IP, a domain, or a unix socket. You should always store the raw address in the .address field. Then it should be duplicated to | foo.acme.com, 1.1.1.1, 1.1.1.1:55001 |
| Destination ARNs | ARNs associated with the destination resource | [”arn:aws:iam::560769337183:role/PantherLogProcessingRole-panther-account-us-west-2”] |
| Destination Bytes | Bytes sent from the destination to the source | 192 |
| Destination Domain | The domain name of the destination system. This value may be a host name, a fully qualified domain name, or another host naming format. The value may derive from the original event or be added from enrichment. | foo.acme.com |
| Destination IP | IP address of the destination (IPv4 or IPv6) | 1.1.1.1 |
| Destination MAC | Mac address of the destination. The notation format from RFC 7042 is suggested: Each octet (that is, 8-bit byte) is represented by two [uppercase] hexadecimal digits giving the value of the octet as an unsigned integer. Successive octets are separated by a hyphen. | 00-00-5E-00-53-23 |
| Destination Port | Port of the destination | 80 |
| Source Address | Some event source addresses are defined ambiguously. The event will sometimes list an IP, a domain, or a unix socket. You should always store the raw address in the .address field. Then it should be duplicated to | foo.acme.com, 1.1.1.1, 1.1.1.1:55001 |
| Source ARNs | ARNs associated with the source resource | [”arn:aws:iam::560769337183:role/PantherLogProcessingRole-panther-account-us-west-2”] |
| Source Bytes | Bytes sent from the source to the destination | 192 |
| Source Domain | The domain name of the source system. This value may be a host name, a fully qualified domain name, or another host naming format. The value may derive from the original event or be added from enrichment. | foo.acme.com |
| Source IP | IP address of the source (IPv4 or IPv6) | 1.1.1.1 |
| Source MAC | Mac address of the source. The notation format from RFC 7042 is suggested: Each octet (that is, 8-bit byte) is represented by two [uppercase] hexadecimal digits giving the value of the octet as an unsigned integer. Successive octets are separated by a hyphen. | 00-00-5E-00-53-23 |
| Source Port | Port of the source | 80 |
| User ARNs | ARNs associated with the user | [”arn:aws:iam::560769337183:role/PantherLogProcessingRole-panther-account-us-west-2”] |
| User Email | User email address | |
| User Full Name | User’s full name, if available | Bob Sanders |
| User Display Name | Short name or login of the user | b.sanders |
| User Providers ID | The ID given by another provider for the user, if it exists | abc123 |
Core Fields vs. Indicator Fields
It may make sense to classify certain event fields as both Core (UDM) and Indicator (p_any) fields. For example, you might map one event field to the destination.ip
UDM field, another event field to the source.ip
UDM field, and include the ip
indicator on both fields, so that the value of each one may be included in p_any_ip_addresses
.
In general, when a field can be classified as both a UDM and p_any field, the UDM mapping maintains relationship information, while the p_any field records that the value was in the event at all (but does not indicate whether it came from the side performing the action or having the action performed on it).
When a field can be classified as both a UDM and p_any field, it is recommended to create both mappings. This will allow you to, in cross-log detections and searches, either include logic that asks if the value was present in the log at all (using the p_any field), or if the value came from a certain side of the relationship (the UDM field).
Core Fields vs. Data Models in Python detections
In addition to Core Fields, Panther supports Data Models in detections, which allow you to define common aliases for event fields across log types, which can then be referenced in Python detections. Below are key differences between Data Models for detections and Core Fields:
When to use each one:
Use Data Models for detections if your objective is only to write detections that can use one alias to reference differently named fields in various log types.
Use Core Fields if, in addition to the above ability in detections, you would also like the convenience of being able to use a single field to search your data lake for values in those differently named event fields across detection types.
How each one is defined:
Data Models for detections, while specific to a log type, are defined separately from the log schema: in the Data Models tab in the Panther Console, or in a
datamodel
file in the CLI workflow.Core Fields are defined in a log schema.
How each one transforms an incoming log:
Creating an alias in Data Models for detections does not alter the structure of incoming events.
When a Core Field is mapped, for each incoming event of that log type, the Core Field/value pair will be added within the event’s
p_udm
object.
How you can access each one in a Python detection:
To access a Data Model for detections field, use the
udm()
function on the event object.For example:
event.udm(...)
To access a Core Field in a Python detection, use either the
deep_get()
orudm()
function on the event object. Note that it is only possible to useudm()
to access a Core Field if there is not also a Data Model for detections mapping defined with the same key name. Learn more about howudm()
works here.For example:
event.deep_get("p_udm", ...)
orevent.udm(...)
If you are using both Core Fields and Data Models for detections, it is possible for a naming conflict to arise if p_udm.
is at the beginning of your Data Models for detections alias name. In these cases, the Core Field takes precedence.
For example, say you mapped an event field to the source.ip
Core Field and defined a Data Models for detections alias called p_udm.source.ip
. In your Python detection, when calling event.udm(p_udm.source.ip)
, the value for the event field mapped to the Core Field (not the Data Model for detections alias) will be used.
Mapping Core Fields in Custom Log Schemas
You can map fields in custom log schemas to Core Fields.
To map fields in a custom log schema to a Core Field:
In the left-hand navigation bar of your Panther Console, click Configure > Schemas.
In the code editor, under the existing schema definition, add a
udm
key.The
udm
field takes a list ofname
andpaths
pairs.The
name
key takes one of the values in the "Field Name" column of the Core Fields table above. The value is denoted with JSON dot notation.The
paths
key takes a list ofpath
keys, whose value is the path to the event key whose value you'd like to set for the UDM field indicated inname
. Thepaths
list will be evaluated chronologically and the first non-nullpath
value will be set to the UDM field. The value is denoted with JSON dot notation.
Example:
Indicator Fields
A common security question is, “Was some indicator
ever observed in any of our logs?” Panther's Search tool enables you to find the answer by searching across data from all of your various log sources.
As log events are ingested, the indicators
field in their corresponding schema identifies which fields should have their values extracted into p_any_
fields, which are appended to and stored with the event. The table below shows which p_any_
field(s) data is extracted into, by indicator
.
When constructing a custom schema, you can use the values in the Indicator Name column in the table below in your schema's indicators
field. Each of the rows (except for hostname
, net_addr
, and url
) corresponds to a "Panther Fields" option in Search. You may want to map certain event fields to Core (UDM) Fields, in addition to Indicator Fields. Learn more in Core Fields vs. Indicator Fields.
Note that field name/value pairs outside of the fields in the table below can be searched with Search's key/value filter expression functionality—though because those fields have not been mapped to corresponding ones (with different syntax) in different log sources, only matches from log sources containing the exact field name searched will be returned.
In order for a field to be designated as an indicator, it must be type string
.
Indicator Name | Extracted into fields | Description |
---|---|---|
actor_id | p_any_actor_ids | Append value to p_any_actor_ids. |
aws_account_id | p_any_aws_account_ids | If the value is a valid AWS account id, append to p_any_aws_account_ids. |
aws_arn | p_any_aws_arns, p_any_aws_instance_ids, p_any_aws_account_ids, p_any_emails | If value is a valid AWS ARN then append to p_any_aws_arns. If the ARN contains an AWS account id, extract and append to p_any_aws_account_ids. If the ARN contains an EC2 instance id, extract and append to p_any_aws_instance_ids. If the ARN references an AWS STS Assume Role and contains and email address, then extract email address into p_any_emails. |
aws_instance_id | p_any_aws_instance_ids | If the value is a valid AWS instance id then append to p_any_aws_instance_ids. |
aws_tag | p_any_aws_tags | Append value into p_any_aws_tags. |
domain | p_any_domain_names | Append value to p_any_domain_names. |
p_any_emails | If value is a valid email address then append value into p_any_emails.
The portion of the value that precedes | |
hostname | p_any_domain_names, p_any_ip_addresses | Append value to p_any_domain_names. If value is a valid ipv4 or ipv6 address then append to p_any_ip_addresses. |
ip | p_any_ip_addresses | If value is a valid ipv4 or ipv6 address then append to p_any_ip_addresses. |
mac | p_any_mac_addresses | If a value is a valid IEEE 802 MAC-48, EUI-48, EUI-64, or a 20-octet IP over InfiniBand link-layer address then append to p_any_mac_addresses. |
md5 | p_any_md5_hashes | If the value is a valid md5 then append value into p_any_md5_hashes. |
net_addr | p_any_domain_names, p_any_ip_addresses | Extracts from values of the form <host>:<port>. Append host portion to p_any_domain_names. If host portion is a valid ipv4 or ipv6 address then append to p_any_ip_addresses. |
serial_number | p_any_serial_numbers | Append value to p_any_serial_numbers. This feature is in closed beta starting with Panther version 1.69. To share any bug reports or feature requests, please contact your Panther support team. |
sha1 | p_any_sha1_hashes | If the value is a valid sha1 then append to p_any_sha1_hashes. |
sha256 | p_any_sha256_hashes | If the value is a valid sha256 then append to p_any_sha256_hashes. |
trace_id | p_any_trace_ids | Append value to p_any_trace_ids. Tag fields such as session ids and document ids that are used to associated elements with other logs in order to trace the full activity of a sequence of related events. |
url | p_any_domain_names, p_any_ip_addresses | Parse url, extract the host portion after "http://" or "https://". Append host portion to p_any_domain_names. If host portion is a valid ipv4 or ipv6 address then append to p_any_ip_addresses. |
username | p_any_usernames | Append value into p_any_usernames.
This field is also populated with values marked with the |
Enrichment Fields
The Panther rules engine will take the looked-up matches from Lookup Tables and append that data to the event using the key p_enrichment
in the following JSON structure:
Enrichment Field Name | Type | Description of Enrichment Field |
---|---|---|
| object | Dictionary of lookup results where matching rows were found. |
| string |
|
The "all_logs" view
Panther manages a view over all data sources with standard fields.
This allows you to answer questions such as, "Was there any activity from some-bad-ip
, and if so, where?"
The query below will show how many records, by log type, are associated with the IP address 95.123.145.92
:
From these results, you can pivot to the specific logs where activity is indicated.
Standard Fields in detections
The Panther standard fields can be used in detections.
For example, the Python rule below triggers when any GuardDuty alert is on a resource tagged as Critical
:
Last updated