Skip to main content

ip_cidr_contains

Applies to: check marked yes Databricks Runtime 18.2 and above

Beta

This feature is in Beta. Workspace admins can control access to this feature from the Previews page. See Manage Databricks previews.

Returns True if an IP address or CIDR block is contained within another CIDR block, False otherwise.

For the corresponding SQL function, see ip_cidr_contains function.

Syntax

Python
from pyspark.databricks.sql import functions as dbf

dbf.ip_cidr_contains(col1=<col1>, col2=<col2>)

Parameters

Parameter

Type

Description

col1

pyspark.sql.Column or str

A STRING or BINARY value representing a valid IPv4 or IPv6 CIDR block.

col2

pyspark.sql.Column or str

A STRING or BINARY value representing a valid IPv4 or IPv6 address or CIDR block.

Examples

Example 1: Check if an IP address is contained in a CIDR block.

Python
from pyspark.databricks.sql import functions as dbf
df = spark.createDataFrame([('192.168.1.0/24', '192.168.1.100')], ['cidr', 'ip'])
df.select(dbf.ip_cidr_contains('cidr', 'ip').alias('result')).collect()
Output
[Row(result=True)]

Example 2: Check if a smaller CIDR block is contained in a larger CIDR block.

Python
from pyspark.databricks.sql import functions as dbf
df = spark.createDataFrame([('192.168.0.0/16', '192.168.1.0/24')], ['cidr', 'needle'])
df.select(dbf.ip_cidr_contains('cidr', 'needle').alias('result')).collect()
Output
[Row(result=True)]

Example 3: None input returns None.

Python
from pyspark.databricks.sql import functions as dbf
df = spark.createDataFrame([(None, '192.168.1.1')], 'cidr: string, ip: string')
df.select(dbf.ip_cidr_contains('cidr', 'ip').alias('result')).collect()
Output
[Row(result=None)]