About 10,800,000 results
Open links in new tab
  1. Printing secret value in Databricks - Stack Overflow

    Nov 11, 2021 · First, install the Databricks Python SDK and configure authentication per the docs here. pip install databricks-sdk Then you can use the approach below to print out secret …

  2. Databricks: How do I get path of current notebook?

    Nov 29, 2018 · The issue is that Databricks does not have integration with VSTS. A workaround is to download the notebook locally using the CLI and then use git locally. I would, however, …

  3. Connecting C# Application to Azure Databricks - Stack Overflow

    The Datalake is hooked to Azure Databricks. The requirement asks that the Azure Databricks is to be connected to a C# application to be able to run queries and get the result all from the C# …

  4. Databricks - Download a dbfs:/FileStore file to my Local Machine

    In a Spark cluster you access DBFS objects using Databricks file system utilities, Spark APIs, or local file APIs. On a local computer you access DBFS objects using the Databricks CLI or …

  5. How to to trigger a Databricks job from another Databricks job?

    Jul 31, 2023 · Databricks is now rolling out the new functionality, called "Job as a Task" that allows to trigger another job as a task in a workflow. Documentation isn't updated yet, but you …

  6. Run a notebook from another notebook in a Repo Databricks

    Jul 6, 2021 · So I cloned the two files (function_notebook, processed_notebook) into a Repo in Databricks. When I try to copy the path where I just cloned it, onlt this option appears: Copy …

  7. Databricks: managed tables vs. external tables - Stack Overflow

    Jun 21, 2024 · While Databricks manages the metadata for external tables, the actual data remains in the specified external location, providing flexibility and control over the data storage …

  8. How to zip files (on Azure Blob Storage) with shutil in Databricks

    Jan 13, 2020 · Actually, without using shutil, I can compress files in Databricks dbfs to a zip file as a blob of Azure Blob Storage which had been mounted to dbfs. Here is my sample code using …

  9. Databricks: convert data frame and export to xls / xlsx

    Sep 30, 2019 · df.write \ .format("com.databricks.spark.csv") \ .option("header", "true") \ .save("myfile.csv") In this example, you can try changing the extension to xls before you run …

  10. How to do an INSERT with VALUES in Databricks into a Table

    This table is mapped via JDBC as a table in Databricks. I want to do insert like in SQL Server: INSERT INTO table_name (column1, column2, column3, ...) VALUES (value1, value2, value3, …

Refresh