Libraries failing with owner or network errors on Databricks Runtime 13.3 LTS - current (15.3)
Problem When migrating to Databricks Runtime 13.3 LTS to current (15.3), libraries start failing with owner or network related errors. Example Library installation attempted on the driver node of cluster XXXX-XXXXXX-XXXXXXXX and failed. Please refer to the following error message to fix the library or contact Databricks support. Error Code: DRIVER_L...
1 min reading timeDatabricks cannot access a notebook in GitHub
Problem Your Databricks job may fail to access a notebook in a GitHub repository after previously being able to. Unable to access the notebook 'resources/notebooks/examplename'. Either it does not exist, or the identity used to run this job, <identity-name> (<identity-reference>), lacks the required permissions. Cause There are two pos...
0 min reading timeRestricting sensitive data in the workspace
Problem You want to restrict workspace users from accessing specific data. For example, you have sensitive data that you do not want everyone to be able to access or modify. Cause The DBFS root is accessible to all users and does not support access control. You should not save sensitive data on DBFS. For more information, review the Recommendations...
0 min reading timeMaven Libraries Start Failing with Timed-Out Errors When Updating to Databricks Runtime 11.3 LTS - 15.3 (current)
Problem When updating Databricks Runtime from previous versions (9.x - 11.x) to any of 11.3 LTS - 15.3 (current), the Maven Libraries start failing with connection timed-out issues while connecting to the repository. Example Server access error at url https://repo1.maven.org/maven2/com/microsoft/azure/azure-eventhubs-spark_2.12/2.3.22/azure-eventhu...
0 min reading timeDatabricks-sql-python package fails with self-signed certificate errors and code _ssl.c:1006
Problem When trying to connect using the databricks-sql-python package from your local Python environment or using development tools like VSCode or PyCharm, you receive a certificate error. MaxRetryError (note: full exception trace is shown but execution 1is paused at: _run_module_as_main) HTTPSConnectionPool(host='my-workspace.cloud.databricks.com...
0 min reading timeGanglia metrics not appearing in historical metrics snapshots list
Problem When you navigate to your Metrics tab to access the Ganglia metrics for a cluster, you notice you have zero files in your Historical metrics snapshots list. Within the file list, you see a message, “No metrics found.” Cause Ganglia snapshots are taken every 15 minutes. If the cluster runs for fewer minutes, and outside the snapshot-taken ...
0 min reading time