I am new to azure databricks. I was trying to read data from datalake into databricks. I found that there are mainly two methods
- Mounting the file present in datalake into dbfs (Advantage being Authentication required just once)
- Using Service Principal and OAuth (Authentication required for each request)
I am interested to know if there is some significant memory consumption when we choose to mount folders in dbfs. I learnt that the data mounted is persisted . So I guessing that might lead to some memory consumption. I'll like if somebody can explain me what's going on the backend when we mount a file in dbfs