Anvilogic on Snowflake Architecture

Anvilogic implementation on Snowflake (AWS, GCP, Azure).

Architecture Diagram

Below is the generic architecture digram for how Anvilogic works on top of Snowflake.

circle-info

This supports Snowflake on Azure, AWS, and GCP.

Overall Diagram:

Anvilogic on Snowflake (AWS, GCP, or Azure)

ETL Parsing & Normalization Process

ETL Process for moving data from Raw format to Gold.

PDF Download:

Frequently Asked Questions (FAQs)

chevron-rightDoes it matter which public cloud I own?hashtag

Snowflake will be configured in the IaaS environment that you have and is available across AWS, GCP, and Azure.

You can also have a separate Snowflake account per environment if you are a multi-IaaS organization.

Data that already originates in IaaS that can be sent to cloud storage does not require a streaming tool and can be onboarded to Snowflake directly.

chevron-rightHow do I onboard logs from data center assets?hashtag

Datasets that come from assets hosted within a data center or not in a public IaaS environment will require a solution to route that data to Snowflake.

Data streaming tools (ex. Cribl, Fluentbit, Apache NiFi) can be used to send on-prem. logs directly to Snowflake.

It is a requirement that you have a data transport/streaming tool to send data to IaaS storage or Anvilogic pipelines for ingestion.

Forwarding agents that are installed on endpoints also need to be re-configured to send to the streaming tools for ingestion into Snowflake.

Snowflake and/or Anvilogic does not provide any data streaming or endpoint agent technology.

chevron-rightHow do you get data that originates in public cloud into Snowflake?hashtag

Configure your security tools & appliances to log to cloud storage services like S3, Blob storage, or GCP storage - Snowpipe then picks it up and ingests into Snowflake.

chevron-rightHow does the Anvilogic SaaS platform communicate with the Snowflake database that gets installed in your Snowflake account?hashtag

All communication (search and detection use cases deployments) are done over REST API using HTTPS/443 with TLS v1.2+.

chevron-rightCan Anvilogic help with getting raw data into Snowflake?hashtag

Yes, if you have a streaming tool (ex. Cribl, Fluentbit, Apache NiFi) you can send custom data sources directly to Anvilogic’s ingestion pipeline. Anvilogic also has some out of the box support for raw data ingestion sources in our integrations armory.

This will send data to our S3 storage service, which temporarily stores data to process it into Snowflake.

chevron-rightDoes Anvilogic help with parsing and normalization of raw data into Snowflake?hashtag

Yes, Anvilogic helps with all of the parsing and normalization of security relevant data into the Anvilogic schema. We have onboarding templates and configs that will help ensure the data you are brining into Snowflake is properly formatted to execute detections and perform triage, hunting, and response.

chevron-rightCan Anvilogic help getting enrichment data into Snowflake?hashtag

Yes, if you have third party Intel or CMDB tools that are required to be used within detection enrichment, those can be called via REST API and transported into a Snowflake table.

Anvilogic detections can then leverage those enrichment tables to enrich detections before those detections are stored in the Alert lake (upstream of SOAR).

chevron-rightDoes Anvilogic have out-of-the-box integrations for specific vendors alert sources?hashtag

Yes, Anvilogic can provide out of the box integrations for common vendor alerts and data collection for specific SaaS Security tools (ex. Crowdstrike FDR).

Tools not listed in our integration marketplace can be sent through the Custom Data Integration pipeline as a self service option.

chevron-rightWhat is the difference between raw data vs Alert data?hashtag

Raw data sources are events/telemetry that is generated from endpoints/tools/appliances (ex. Windows Event logs, EDR logs).

Alerts data is curated signals from security tools (ex. Proofpoint alerts, Anti-virus alerts, etc.) that has already been identified to be suspicious or malicious by the vendor.

chevron-rightDo you use Snowflake Warehouses?hashtag

Yes, Anvilogic requires 2 warehouses to run.

  • Ad-hoc Warehouse - Compute for queries to assist search, hunt, and IR

  • Detect Warehouse - Run 24/7 executing scheduled tasks (detections) on a cron

chevron-rightDo you integrate with SOAR?hashtag

Yes, Anvilogic can integrate with most SOARs via REST API through either a push or a pull method.

chevron-rightDoes Anvilogic have a search user interface (UI) for Snowflake?hashtag

Yes, Anvilogic has a search user interface (UI) to make it easy to query data that is inside of a Snowflake database.

In addition, Anvilogic makes it easy to build repeatable detections that can execute on top of Snowflake using a low-code UI builder.

chevron-rightDoes Anvilogic have a data model? Does it work with OCSF?hashtag

Yes, Anvilogic has a data model and offers parsing and normalization code for any security data set that you want to use within the platform.

Yes, we can also work with OCSF data, and each data feed can be modified/controlled to customize to your needs.

chevron-rightDoes Anvilogic support IOC collection & searching?hashtag

Yes, Anvilogic can onboard IOCs from your third party threat intel tools (ex. Threat Connect) and use that data to create new detections, conduct ongoing exposure checks across your data feeds, or use it to enrich your alert output for triage analysts.

Last updated

Was this helpful?