Databricks

Reading Data from Cosmos DB in Databricks: A Comprehensive Guide

In today’s data-driven world, organizations leverage various data storage solutions to manage and analyze their data effectively. Cosmos DB, a globally distributed NoSQL database service from Microsoft Azure, is widely used for building highly scalable and responsive applications. In this blog post, we will explore how to read data from Cosmos DB in Databricks, a […]

Reading Data from Cosmos DB in Databricks: A Comprehensive Guide Read More »

Download a file from DBFS – Databricks to the local machine

In this post, we will learn how we can download a file from DBFS i.e. Databricks File System to the Local machine. DBFS is the File system that Databricks uses to store its files. It is a distributed file system mounted into a Databricks workspace and it is available on Databricks clusters. To demonstrate how

Download a file from DBFS – Databricks to the local machine Read More »

Use HDFS API to read Azure Blob files in Databricks

Databricks provides a wrapper file system API named DBFS (Databricks File System) to perform any file-level operation such as read, write, move, delete, rename, etc. However, sometimes we may need to read the underlying file system objects directly without using the DBFS wrapper APIs. To do so, we can use HDFS APIs available through py4j

Use HDFS API to read Azure Blob files in Databricks Read More »