Databricks utils
Utilities for databricks operations.
DatabricksUtils
¶
Bases: object
Databricks utilities class.
Source code in mkdocs/lakehouse_engine/packages/utils/databricks_utils.py
15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 | |
get_databricks_job_information(spark)
staticmethod
¶
Get notebook context from running acon.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
spark
|
SparkSession
|
spark session. |
required |
Returns:
| Type | Description |
|---|---|
Tuple[str, str]
|
Dict containing databricks notebook context. |
Source code in mkdocs/lakehouse_engine/packages/utils/databricks_utils.py
get_db_utils(spark)
staticmethod
¶
Get db utils on databricks.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
spark
|
SparkSession
|
spark session. |
required |
Returns:
| Type | Description |
|---|---|
Any
|
Dbutils from databricks. |
Source code in mkdocs/lakehouse_engine/packages/utils/databricks_utils.py
get_spark_conf_values(usage_stats, spark_confs)
staticmethod
¶
Get information from spark session configurations.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
usage_stats
|
dict
|
usage_stats dictionary file. |
required |
spark_confs
|
dict
|
optional dictionary with the spark tags to be used when collecting the engine usage. |
required |
Source code in mkdocs/lakehouse_engine/packages/utils/databricks_utils.py
get_usage_context_for_serverless(usage_stats)
classmethod
¶
Get information from the execution environment for serverless scenarios.
Since in serverless environments we might not have access to all the spark confs we want to collect, we will try to get that information from the execution environment when possible.
Parameters:
| Name | Type | Description | Default |
|---|---|---|---|
usage_stats
|
dict
|
usage_stats dictionary file. |
required |
Source code in mkdocs/lakehouse_engine/packages/utils/databricks_utils.py
is_serverless_workload()
staticmethod
¶
Check if the current databricks workload is serverless.
Returns:
| Type | Description |
|---|---|
bool
|
True if the current databricks workload is serverless, False otherwise. |