About 21,500,000 results
Open links in new tab
  1. Printing secret value in Databricks - Stack Overflow

    Nov 11, 2021 · 2 Building on @camo's answer, since you're looking to use the secret value outside Databricks, you can use the Databricks Python SDK to fetch the bytes representation of the secret …

  2. Is there a way to use parameters in Databricks in SQL with parameter ...

    Sep 29, 2024 · Databricks demands the use of the IDENTIFIER () clause when using widgets to reference objects including tables, fields, etc., which is exactly what you're doing.

  3. Databricks: managed tables vs. external tables - Stack Overflow

    Jun 21, 2024 · While Databricks manages the metadata for external tables, the actual data remains in the specified external location, providing flexibility and control over the data storage lifecycle. This …

  4. Databricks shared access mode limitations - Stack Overflow

    Oct 2, 2023 · You're correct about listed limitations. But when you're using Unity Catalog, especially with shared clusters, you need to think a bit differently than before. UC + shared clusters provide very …

  5. Databricks Permissions Required to Create a Cluster

    Nov 9, 2023 · In Azure Databricks, if you want to create a cluster, you need to have the " Can Manage " permission. This permission basically lets you handle everything related to clusters, like making new …

  6. how to get databricks job id at the run time - Stack Overflow

    Jun 9, 2025 · 1 I am trying to get the job id and run id of a databricks job dynamically and keep it on in the table with below code

  7. Efficient SQL query with pandas using databricks-sql-python

    Nov 28, 2024 · 2 Databricks allows to make SQL queries via an API using the databricks-sql-python package. There are then two ways of creating a connection object that can be put into a …

  8. REST API to query Databricks table - Stack Overflow

    Jul 24, 2022 · Is databricks designed for such use cases or is a better approach to copy this table (gold layer) in an operational database such as azure sql db after the transformations are done in pyspark …

  9. Where does databricks store the managed tables? - Stack Overflow

    Nov 6, 2024 · Answering your two sub questions individually below: Does this mean that databricks is storing tables in the default Storage Account created during the creation of Databricks workspace ? …

  10. Databricks - Download a dbfs:/FileStore file to my Local Machine

    Method3: Using third-party tool named DBFS Explorer DBFS Explorer was created as a quick way to upload and download files to the Databricks filesystem (DBFS). This will work with both AWS and …