<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Reiji Otake</title>
    <description>The latest articles on DEV Community by Reiji Otake (@_d2a1ea24c442526a9777).</description>
    <link>https://dev.to/_d2a1ea24c442526a9777</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2869291%2F82d4546b-2c47-45b0-ad5f-8572666c2a2e.jpeg</url>
      <title>DEV Community: Reiji Otake</title>
      <link>https://dev.to/_d2a1ea24c442526a9777</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/_d2a1ea24c442526a9777"/>
    <language>en</language>
    <item>
      <title>Fabric &amp; Databricks Interoperability (4): Using Databricks Tables in Fabric for Viewing, Analysis, and Editing</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 10:52:27 +0000</pubDate>
      <link>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-4-using-databricks-tables-in-fabric-for-viewing-analysis-2n2m</link>
      <guid>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-4-using-databricks-tables-in-fabric-for-viewing-analysis-2n2m</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;Is it possible to seamlessly reference and edit tables created in Fabric within Databricks?&lt;br&gt;&lt;br&gt;
Many people may have this question.&lt;/p&gt;

&lt;p&gt;In this article, we will specifically explore the use case of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Utilizing tables created in Databricks within Fabric.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For prerequisite settings and configurations, please refer to previous articles.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;This article is part of a four-part series:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer"&gt;Overview and Purpose of Interoperability&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer"&gt;Detailed Configuration of Hub Storage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer"&gt;Using Tables Created in Fabric in Databricks&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Using Tables Created in Databricks in Fabric (this article)&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;h1&gt;
  
  
  Linking Tables Created in Databricks to Fabric
&lt;/h1&gt;
&lt;h3&gt;
  
  
  Creating a New Table in Databricks
&lt;/h3&gt;

&lt;p&gt;Create a new empty external table from Databricks.&lt;br&gt;&lt;br&gt;
Specify the folder path of the hub storage as the &lt;code&gt;Location&lt;/code&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9gw6t5m71lji56wl67td.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9gw6t5m71lji56wl67td.png" alt="image.png" width="800" height="182"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;CREATE&lt;/span&gt; &lt;span class="k"&gt;TABLE&lt;/span&gt; &lt;span class="n"&gt;create_from_Databricks_sales&lt;/span&gt;
&lt;span class="k"&gt;USING&lt;/span&gt; &lt;span class="n"&gt;DELTA&lt;/span&gt;
&lt;span class="k"&gt;LOCATION&lt;/span&gt; &lt;span class="s1"&gt;'abfss://&amp;lt;container_name&amp;gt;@&amp;lt;ADLS2_name&amp;gt;.dfs.core.windows.net/folder_name/create_from_Databricks_sales'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;h1&gt;
  
  
  Checking the Created Table
&lt;/h1&gt;

&lt;p&gt;You can verify that the external table created from the Catalog Explorer contains data.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F12p0agyv80gwm8ruy3oe.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F12p0agyv80gwm8ruy3oe.png" alt="image.png" width="800" height="362"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A folder named &lt;code&gt;create_from_Databricks_sales&lt;/code&gt; is created in the &lt;code&gt;ext&lt;/code&gt; folder of the hub storage.&lt;br&gt;
(This means that the newly created external table physically exists in the hub storage.)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi5mqmy10qlypxwbotd56.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi5mqmy10qlypxwbotd56.png" alt="image.png" width="800" height="200"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;It can also be confirmed that the table is in Delta format.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjkcb8byqyzrt10yh5m99.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjkcb8byqyzrt10yh5m99.png" alt="image.png" width="800" height="165"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;At this point, the &lt;code&gt;create_from_Databricks_sales&lt;/code&gt; table also becomes visible from Fabric's Lakehouse.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5vbjfzqstq9km8y11su5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5vbjfzqstq9km8y11su5.png" alt="image.png" width="800" height="297"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Viewing and Analyzing Tables Created in Databricks in Fabric (Creating BI)
&lt;/h1&gt;

&lt;p&gt;From the Semantic Model, select the &lt;code&gt;create_from_Databricks_sales&lt;/code&gt; table (created in Databricks) and click &lt;strong&gt;[Confirm]&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsi9a6ftlv5iw6fz8m0zv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsi9a6ftlv5iw6fz8m0zv.png" alt="image.png" width="800" height="477"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Now, the table created in Databricks can be analyzed in Fabric.&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6byru9bnamh8n4hepg4f.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6byru9bnamh8n4hepg4f.png" alt="image.png" width="800" height="374"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Editing (DML) Tables Created in Databricks from Fabric
&lt;/h1&gt;

&lt;p&gt;Execute an &lt;code&gt;UPDATE&lt;/code&gt; statement (DML statement) from Fabric's Notebook.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwml9wx811uvq2v5jb7a4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwml9wx811uvq2v5jb7a4.png" alt="image.png" width="800" height="320"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;UPDATE&lt;/span&gt; &lt;span class="n"&gt;Fabric_Lakehouse&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ext&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;create_from_Databricks_sales&lt;/span&gt;
&lt;span class="k"&gt;SET&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'No.1 Quantity Water Bottle - 30 oz.'&lt;/span&gt;
&lt;span class="k"&gt;WHERE&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'Water Bottle - 30 oz.'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Of course, it was confirmed that changes were reflected from Fabric.&lt;br&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3dmp5cfgs18eifpmgat5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3dmp5cfgs18eifpmgat5.png" alt="image.png" width="800" height="405"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Editing was performed from Fabric, and the changes were also reflected on the Databricks side.&lt;br&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmlpsnk1519v4v2xko504.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmlpsnk1519v4v2xko504.png" alt="image.png" width="800" height="347"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Therefore, it is possible to edit (DML statements) in Fabric for tables created in Databricks.&lt;/strong&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Issues and Specific Operational Methods
&lt;/h1&gt;

&lt;p&gt;The method introduced here has the advantage that both Fabric and Databricks can edit data. However, this can also be a weakness, as it makes table updates too easy.&lt;/p&gt;

&lt;p&gt;Additionally, in this case, an external table in Databricks was used.&lt;br&gt;&lt;br&gt;
However, predictive optimization is currently only available for managed tables, making it ideal to use managed tables rather than external ones.&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://docs.databricks.com/en/optimizations/predictive-optimization.html" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fwww.databricks.com%2Fwp-content%2Fuploads%2F2020%2F04%2Fog-databricks.png" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://docs.databricks.com/en/optimizations/predictive-optimization.html" rel="noopener noreferrer" class="c-link"&gt;
          Predictive optimization for Unity Catalog managed tables | Databricks on AWS
        &lt;/a&gt;
      &lt;/h2&gt;
        
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdocs.databricks.com%2Fen%2F_static%2Ffavicon.ico" width="32" height="32"&gt;
        docs.databricks.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;



&lt;p&gt;We will continue to examine these challenges and share specific operational methods in the future.&lt;/p&gt;

&lt;p&gt;I also think that table cloning in Databricks might provide some useful hints.&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/taka_yayoi/items/50c5a75caff8a6fb721d" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRmxoMy5nb29nbGV1c2VyY29udGVudC5jb20lMkZhLSUyRkFPaDE0R2lIRHVtdWNzM282Zm16ZnJFc1NjdjJ4UkNIbE9sVXBuNXpOcTV1JTNEczUwP2l4bGliPXJiLTQuMC4wJmFyPTElM0ExJmZpdD1jcm9wJm1hc2s9ZWxsaXBzZSZmbT1wbmczMiZzPTQzMzcwYmE0ZDk0OGQyZTg1YjQyZWI3MjI3Nzk0NjEy%2526blend-x%253D120%2526blend-y%253D462%2526blend-w%253D90%2526blend-h%253D90%2526blend-mode%253Dnormal%2526mark64%253DaHR0cHM6Ly9xaWl0YS1vcmdhbml6YXRpb24taW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1vcmdhbml6YXRpb24taW1hZ2UlMkZiODdjZTQ0N2RjZWRiZGFhM2UzOTFmOTFlYjIzZjZiMzE4ZjM4ZjAxJTJGb3JpZ2luYWwuanBnJTNGMTYzNDA5MzAxNT9peGxpYj1yYi00LjAuMCZ3PTQ0Jmg9NDQmZml0PWNyb3AmbWFzaz1jb3JuZXJzJmNvcm5lci1yYWRpdXM9OCZib3JkZXI9MiUyQ0ZGRkZGRiZmbT1wbmczMiZzPTg2NWY1ODMxODg3MTVhYmI4YWU0YjJiZTMxN2JiOTM4%2526mark-x%253D186%2526mark-y%253D515%2526mark-w%253D40%2526mark-h%253D40%2526s%253Daa176652e589296aa56b217cc954a4ae%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RGF0YWJyaWNrcyVFMyU4MSVBQiVFMyU4MSU4QSVFMyU4MSU5MSVFMyU4MiU4QiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MSVBRSVFMyU4MiVBRiVFMyU4MyVBRCVFMyU4MyVCQyVFMyU4MyVCMyZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPWQwYWU5MzEyNTc2MzVkYzBhNzYwYjljN2EwOTAyMjYz%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDB0YWthX3lheW9pJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9N2IxNWFlODUxN2VkZTFjYmVjZDkwMWE4NzA0NDk3MjA%26blend-x%3D242%26blend-y%3D454%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26txt64%3D44OH44O844K_44OW44Oq44OD44Kv44K544O744K444Oj44OR44Oz5qCq5byP5Lya56S-%26txt-x%3D242%26txt-y%3D539%26txt-width%3D838%26txt-clip%3Dend%252Cellipsis%26txt-color%3D%25231E2121%26txt-font%3DHiragino%2520Sans%2520W6%26txt-size%3D28%26s%3D2c6db19feac2a4f55ea441e026b9c91c" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/taka_yayoi/items/50c5a75caff8a6fb721d" rel="noopener noreferrer" class="c-link"&gt;
          Databricksにおけるテーブルのクローン #deltalake - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          Clone a table on Databricks | Databricks on AWS [2022/10/28時点]の翻訳です。本書は抄訳であり内容の正確性を保証するものではありません。正…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Based on the above,&lt;br&gt;&lt;br&gt;
it was confirmed that &lt;strong&gt;"tables created in Databricks can be used in Fabric."&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Once the hub storage is set up, achieving interoperability between Fabric and Databricks is relatively simple.&lt;/p&gt;

&lt;p&gt;▽ Previous article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMiVFRiVCQyU5QUZhYnJpYyUyMCVFMyU4MSVBNyVFNCVCRCU5QyVFNiU4OCU5MCVFMyU4MSU5NyVFMyU4MSU5RiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MiU5MkRhdGFicmlja3MlRTMlODElQTclRTUlODglQTklRTclOTQlQTglRTMlODElOTklRTMlODIlOEIlRUYlQkMlODhEYXRhYnJpYyVFMiU4MCVBNiZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPWQwMjM0NmMzNTEzZTY4MmRiMGI0ODkzMTRkZDVkZGE3%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D89cc3454413e6d48e05254cf27917076" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性③：Fabric で作成したテーブルをDatabricksで利用する（Databrickで閲覧・分析・編集可能） #BI - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめにFabricで作成したテーブルをDatabricksでもシームレスに参照・編集が可能できるのか？このような疑問を持つ方も少なくないと思います。そこで今回はFabric で作成したテー…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>databricks</category>
      <category>azure</category>
      <category>deltalake</category>
      <category>sql</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (3): Using Fabric Tables in Databricks for Viewing, Analyzing, and Editing</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 10:34:06 +0000</pubDate>
      <link>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-3-using-fabric-tables-in-databricks-for-viewing-analyzing-39o6</link>
      <guid>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-3-using-fabric-tables-in-databricks-for-viewing-analyzing-39o6</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;Can tables created in Fabric be seamlessly referenced and edited in Databricks?&lt;br&gt;&lt;br&gt;
Many people may have this question.&lt;/p&gt;

&lt;p&gt;In this article, we will specifically introduce the use case of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Using tables created in Fabric within Databricks.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For details on the prerequisite settings, please refer to the previous article.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;This article is part of a four-part series:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer"&gt;Overview and Purpose of Interoperability&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer"&gt;Detailed Configuration of Hub Storage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Using tables created in Fabric within Databricks (this article)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer"&gt;Using tables created in Databricks within Fabric&lt;/a&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;/blockquote&gt;
&lt;h1&gt;
  
  
  Linking Tables Created in Fabric to Databricks
&lt;/h1&gt;
&lt;h3&gt;
  
  
  Creating a New Table in Fabric
&lt;/h3&gt;

&lt;p&gt;Upload a CSV file to the Fabric Lakehouse.&lt;br&gt;&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcxzzllf6mte1ccquj4n6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcxzzllf6mte1ccquj4n6.png" alt="image.png" width="800" height="389"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
The CSV file used in this article is &lt;code&gt;sales.csv&lt;/code&gt; from the following Microsoft documentation:&lt;br&gt;&lt;br&gt;
&lt;a href="https://microsoftlearning.github.io/mslearn-fabric.ja-jp/Instructions/Labs/01-lakehouse.html" rel="noopener noreferrer"&gt;Create a Microsoft Fabric Lakehouse&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;

&lt;p&gt;From the CSV file, select &lt;strong&gt;[Load to Table] &amp;gt; [New Table]&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0x58re7crc7th2sg6f04.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0x58re7crc7th2sg6f04.png" alt="image.png" width="800" height="195"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Specify &lt;code&gt;ext&lt;/code&gt;, which is a shortcut created in the hub storage, as the schema.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2whh73kt23pxpx003rcd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2whh73kt23pxpx003rcd.png" alt="image.png" width="511" height="373"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Verifying the Created Table
&lt;/h3&gt;

&lt;p&gt;You can confirm that a new table has been created in the Lakehouse.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1c8e7xm0dpavkuoow7wd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1c8e7xm0dpavkuoow7wd.png" alt="image.png" width="800" height="297"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A &lt;code&gt;create_from_fabric_sales&lt;/code&gt; folder is created in the &lt;code&gt;ext&lt;/code&gt; folder of the hub storage.&lt;br&gt;&lt;br&gt;
(This means that the newly created table physically exists in the hub storage.)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7154fjipv5h6btc4amsh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7154fjipv5h6btc4amsh.png" alt="image.png" width="800" height="354"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You can also confirm that the table is in Delta format.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwd27bit08i4eewi7a6qn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwd27bit08i4eewi7a6qn.png" alt="image.png" width="800" height="187"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;At this point, as expected, the table created in Fabric is not yet visible in Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7uv6nmxm67bowc9fuhk9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7uv6nmxm67bowc9fuhk9.png" alt="image.png" width="800" height="266"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  Enabling Databricks to Access Fabric Tables
&lt;/h3&gt;

&lt;p&gt;Use the Databricks SQL Editor to create an external table.&lt;br&gt;&lt;br&gt;
Specify the &lt;strong&gt;hub storage folder path&lt;/strong&gt; (the folder of the table created in Fabric) in the &lt;strong&gt;Location&lt;/strong&gt; field.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F95p8lkdin7q8bwmfxr1b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F95p8lkdin7q8bwmfxr1b.png" alt="image.png" width="800" height="204"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;CREATE&lt;/span&gt; &lt;span class="k"&gt;TABLE&lt;/span&gt; &lt;span class="o"&gt;&amp;lt;&lt;/span&gt;&lt;span class="k"&gt;table_name&lt;/span&gt;&lt;span class="o"&gt;&amp;gt;&lt;/span&gt;
&lt;span class="k"&gt;USING&lt;/span&gt; &lt;span class="n"&gt;DELTA&lt;/span&gt;
&lt;span class="k"&gt;LOCATION&lt;/span&gt; &lt;span class="s1"&gt;'abfss://&amp;lt;container_name&amp;gt;@&amp;lt;ADLS2_name&amp;gt;.dfs.core.windows.net/folder_name/&amp;lt;table_folder_name&amp;gt;'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Then, you can view tables created in Fabric from the [Catalog].&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm82dbs21vcwarkrama3z.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm82dbs21vcwarkrama3z.png" alt="image.png" width="800" height="376"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Viewing and Analyzing Tables Created in Fabric with Databricks (BI Creation)
&lt;/h1&gt;

&lt;p&gt;From the [Dashboard] in Databricks, you can create a new dashboard and select an external table (i.e., a table created in Fabric) from [Data] &amp;gt; [Select Table].&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftisgx3r0sku69kvqdvg0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftisgx3r0sku69kvqdvg0.png" alt="image.png" width="800" height="508"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thus, it is possible to analyze tables created in Fabric using Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2udlz9uzlmgc1cr1bgq6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2udlz9uzlmgc1cr1bgq6.png" alt="image.png" width="800" height="366"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Editing Tables Created in Fabric with Databricks (DML)
&lt;/h1&gt;

&lt;p&gt;Try executing an &lt;code&gt;UPDATE&lt;/code&gt; statement (DML statement) from the SQL Editor in Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs5a34tc4u42anmnh4q7y.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs5a34tc4u42anmnh4q7y.png" alt="image.png" width="800" height="179"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;UPDATE&lt;/span&gt; &lt;span class="n"&gt;create_from_fabric_sales&lt;/span&gt;
&lt;span class="k"&gt;SET&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'No.1 Item'&lt;/span&gt;
&lt;span class="k"&gt;WHERE&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="s1"&gt;'Road-150 Red, 48'&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Of course, you can confirm that the changes have been reflected on the Databricks side.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flasox5tqd3m4fntjbtdb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flasox5tqd3m4fntjbtdb.png" alt="image.png" width="800" height="511"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Although the edit was made from Databricks, the changes were successfully reflected on the Fabric side as well.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F65r3o65g0pli04xkiqqz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F65r3o65g0pli04xkiqqz.png" alt="image.png" width="800" height="372"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="k"&gt;SUM&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;Quantity&lt;/span&gt; &lt;span class="o"&gt;*&lt;/span&gt; &lt;span class="n"&gt;UnitPrice&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="k"&gt;AS&lt;/span&gt; &lt;span class="n"&gt;Revenue&lt;/span&gt;
&lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;Fabric_Lakehouse&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ext&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;create_from_fabric_sales&lt;/span&gt;
&lt;span class="k"&gt;GROUP&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;Item&lt;/span&gt;
&lt;span class="k"&gt;ORDER&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;Revenue&lt;/span&gt; &lt;span class="k"&gt;DESC&lt;/span&gt;&lt;span class="p"&gt;;&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;Therefore, it is possible to edit tables created in Fabric using Databricks (DML statements).&lt;/p&gt;
&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;From the above, we have confirmed that&lt;br&gt;&lt;br&gt;
&lt;strong&gt;"Tables created in Fabric can be used in Databricks."&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Once the hub storage is set up, it is relatively easy to achieve interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;In the next article, we will introduce the reverse case:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;"Using tables created in Databricks in Fabric."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;▽ Next article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMyVFRiVCQyU5QURhdGFicmljayUyMCVFMyU4MSVBNyVFNCVCRCU5QyVFNiU4OCU5MCVFMyU4MSU5NyVFMyU4MSU5RiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MiU5MkZhYnJpYyVFMyU4MSVBNyVFNSU4OCVBOSVFNyU5NCVBOCVFMyU4MSU5OSVFMyU4MiU4QiVFRiVCQyU4OEZhYnJpYyVFMyU4MSVBNyVFOSU5NiVCMiVFMiU4MCVBNiZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPTEwMTgzZDJlN2VmOTJlYzZjMmQ3MGNlNTJjOGRkZjNm%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D077c2fba487d7ed95768554cc83c0ac1" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性④：Databrick で作成したテーブルをFabricで利用する（Fabricで閲覧・分析・編集可能） #SQL - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめにFabricで作成したテーブルをDatabricksでもシームレスに参照・編集が可能できるのか？このような疑問を持つ方も少なくないと思います。そこで今回はDatabricksで作成し…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;



&lt;p&gt;▽ Previous article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMSVFRiVCQyU5QWh1YiVFMyU4MiVCOSVFMyU4MyU4OCVFMyU4MyVBQyVFMyU4MyVCQyVFMyU4MiVCOCVFOCVBOCVBRCVFNSVBRSU5QSVFNiU5NiVCOSVFNiVCMyU5NSUyMC1EYXRhYnJpY2tzJTIwJUUzJTgxJUE3JUU0JUJEJTlDJUU2JTg4JTkwJUUzJTgxJTk3JUUzJTgxJTlGJUUzJTgzJTg2JUUzJTgzJUJDJUUzJTgzJTk2JUUzJTgzJUFCJUUzJTgyJTkyRmFiJUUyJTgwJUE2JnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9MjM4ZTdmYWNlYmVkYzQ4ZDYzZmQzZTNhYjkzNzM5NjQ%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3Da729c7480d7acdfdd31cc6768c35f55a" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性②：hubストレージ設定方法 -Databricks で作成したテーブルをFabric で利用する、Fabric で作成したテーブルをDatabricksで利用する- #Azure - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめに今回はDatabricks で作成したテーブルをFabric で利用するFabric で作成したテーブルをDatabricksで利用するというユースケースを実施するための設定方法につ…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>databricks</category>
      <category>azure</category>
      <category>bi</category>
      <category>venderfree</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (2): Configuring Hub Storage</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 09:48:01 +0000</pubDate>
      <link>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-2-configuring-hub-storage-4l85</link>
      <guid>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-2-configuring-hub-storage-4l85</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;This article provides a detailed guide on &lt;strong&gt;how to configure settings&lt;/strong&gt; for the following use cases:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Databricks within Fabric&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Fabric within Databricks&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;:::note info&lt;br&gt;
This article is part of a four-part series:&lt;br&gt;
1. &lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer"&gt;Overview &amp;amp; Purpose of Interoperability&lt;/a&gt;&lt;br&gt;&lt;br&gt;
2. &lt;strong&gt;Detailed Configuration of Hub Storage (This Article)&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
3. &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer"&gt;Using Tables Created in Fabric within Databricks&lt;/a&gt;&lt;br&gt;&lt;br&gt;
4. &lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer"&gt;Using Tables Created in Databricks within Fabric&lt;/a&gt;&lt;br&gt;&lt;br&gt;
:::&lt;/p&gt;
&lt;h1&gt;
  
  
  Preparing Azure Data Lake Gen2 (ADLS2) as the Hub
&lt;/h1&gt;
&lt;h3&gt;
  
  
  ① Deploy a storage account in Azure Portal as the hub
&lt;/h3&gt;

&lt;p&gt;:::note warn&lt;br&gt;
Enable hierarchical namespace.&lt;br&gt;
:::&lt;/p&gt;
&lt;h3&gt;
  
  
  ② Create a container named 'hub' and a directory named 'ext'
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F59atbvryl38i2jroh241.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F59atbvryl38i2jroh241.png" alt="image.png" width="800" height="151"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Connecting Fabric to Hub Storage
&lt;/h1&gt;
&lt;h3&gt;
  
  
  ① Create a Lakehouse
&lt;/h3&gt;

&lt;p&gt;:::note warn&lt;br&gt;
Enable Lakehouse schema (public preview).&lt;br&gt;
:::&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F17nlumakj3nx01q7mgp9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F17nlumakj3nx01q7mgp9.png" alt="image.png" width="800" height="256"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h3&gt;
  
  
  ② Specify the hub storage in the new schema shortcut of the Lakehouse
&lt;/h3&gt;

&lt;p&gt;From [Tables] in the Lakehouse, click the three-dot menu and select [New Schema Shortcut].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiw1kf6ja3nijf42bnx65.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiw1kf6ja3nijf42bnx65.png" alt="image.png" width="800" height="625"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Select [Azure Data Lake Gen2].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp2cqeyvynuj0m9v23x9b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp2cqeyvynuj0m9v23x9b.png" alt="image.png" width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Enter the details for creating a new connection.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fob8ff4v75goe4962p13a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fob8ff4v75goe4962p13a.png" alt="image.png" width="800" height="444"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;/p&gt;
&lt;h3&gt;
  
  
  How to check ADLS2 access URL:
&lt;/h3&gt;

&lt;p&gt;You can confirm it from the storage account's [Endpoints] section under 'Data Lake Storage'.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fncngv3z9ggqxgml10ewq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fncngv3z9ggqxgml10ewq.png" alt="image.png" width="600" height="892"&gt;&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;

&lt;p&gt;Enable the 'ext' directory and click [Next].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdn2r7tb9ypeaj28dibld.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdn2r7tb9ypeaj28dibld.png" alt="image.png" width="800" height="444"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Click [Create].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmoyldilqcyv59e17xeuv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmoyldilqcyv59e17xeuv.png" alt="image.png" width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The 'ext' directory is created as an external schema shortcut.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F34nrtxwf95t0wsw4t1pn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F34nrtxwf95t0wsw4t1pn.png" alt="image.png" width="800" height="719"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h1&gt;
  
  
  Connecting Databricks to Hub Storage
&lt;/h1&gt;
&lt;h3&gt;
  
  
  ① Create an access connector for Azure Databricks in the Azure Portal
&lt;/h3&gt;

&lt;p&gt;Follow the steps under &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/data-governance/unity-catalog/azure-managed-identities#config-managed-id" rel="noopener noreferrer"&gt;"Step 1: Create an Access Connector for Azure Databricks" in the "Use Azure Managed Identity to Access Storage in Unity Catalog"&lt;/a&gt; guide, using a system-assigned managed identity.&lt;/p&gt;
&lt;h3&gt;
  
  
  ② Grant the connector access to the hub storage from the Azure Portal
&lt;/h3&gt;

&lt;p&gt;Follow &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/data-governance/unity-catalog/azure-managed-identities#config-managed-id" rel="noopener noreferrer"&gt;"Step 2: Grant Managed Identity Access to the Storage Account" in the same guide&lt;/a&gt;.&lt;/p&gt;
&lt;h3&gt;
  
  
  ③ Create storage credentials in Databricks
&lt;/h3&gt;

&lt;p&gt;Log in to Databricks and navigate to [Catalog] &amp;gt; [+] &amp;gt; [Add Storage Credentials].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo848dqgr4hdtkxajfmzs.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo848dqgr4hdtkxajfmzs.png" alt="image.png" width="800" height="307"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Add new storage credentials.&lt;br&gt;
|  | Input Value |&lt;br&gt;
|:-:|:-:|&lt;br&gt;
|Storage Credentials or Service Credentials| Storage Credentials |&lt;br&gt;
|Credential Name| Any name |&lt;br&gt;
|Access Connector ID| Resource ID of the connector created in step ① (can be confirmed in Azure Portal) |&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvyzntl21l0ds0eiafyy6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvyzntl21l0ds0eiafyy6.png" alt="image.png" width="800" height="739"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After creation, click the newly created credential name.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flc56yb266v7ymkz9dmn1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flc56yb266v7ymkz9dmn1.png" alt="image.png" width="800" height="177"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Click [Permissions] &amp;gt; [Grant].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft5921d1irf5cotp74d5i.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft5921d1irf5cotp74d5i.png" alt="image.png" width="800" height="193"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Grant [ALL PRIVILEGES] to necessary users.&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvge8jyfa5jcol2q8xvfp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvge8jyfa5jcol2q8xvfp.png" alt="image.png" width="800" height="456"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
Reference for steps and required permissions:&lt;br&gt;&lt;br&gt;
&lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/connect/unity-catalog/cloud-storage/storage-credentials#next-steps" rel="noopener noreferrer"&gt;Create Storage Credentials for Connecting to Azure Data Lake Storage Gen2&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;
&lt;h3&gt;
  
  
  ④ Add an external location in Databricks
&lt;/h3&gt;

&lt;p&gt;Log in to Databricks and navigate to [Catalog] &amp;gt; [+] &amp;gt; [Add External Location].&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz2ua8koc5ywms36u7oc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyz2ua8koc5ywms36u7oc.png" alt="image.png" width="800" height="221"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Create a new external location.&lt;br&gt;
|  | Input Value |&lt;br&gt;
|:-:|:-:|&lt;br&gt;
|External Location Name| Any name |&lt;br&gt;
|URL| abfss://directory-name (hub) @ storage-account-name.dfs.windows.net |&lt;br&gt;
|Storage Credentials| Select the credentials created in step ③ |&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9uggh8hwfbamejjke773.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9uggh8hwfbamejjke773.png" alt="image.png" width="800" height="570"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;/p&gt;
&lt;h3&gt;
  
  
  How to determine the URL:
&lt;/h3&gt;

&lt;p&gt;Refer to the storage account's [Endpoints] section used in step ② of "Connecting Fabric to Hub Storage".&lt;br&gt;
:::&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
Reference for steps and required permissions:&lt;br&gt;&lt;br&gt;
&lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/connect/unity-catalog/cloud-storage/external-locations" rel="noopener noreferrer"&gt;Create an External Location to Connect Cloud Storage to Azure Databricks&lt;/a&gt;&lt;br&gt;
:::&lt;/p&gt;
&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Now everything is set up!&lt;br&gt;&lt;br&gt;
Next, let's proceed with the actual interoperability of tables.&lt;/p&gt;

&lt;p&gt;▽ Next Article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMiVFRiVCQyU5QUZhYnJpYyUyMCVFMyU4MSVBNyVFNCVCRCU5QyVFNiU4OCU5MCVFMyU4MSU5NyVFMyU4MSU5RiVFMyU4MyU4NiVFMyU4MyVCQyVFMyU4MyU5NiVFMyU4MyVBQiVFMyU4MiU5MkRhdGFicmlja3MlRTMlODElQTclRTUlODglQTklRTclOTQlQTglRTMlODElOTklRTMlODIlOEIlRUYlQkMlODhEYXRhYnJpYyVFMiU4MCVBNiZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzMUUyMTIxJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPWQwMjM0NmMzNTEzZTY4MmRiMGI0ODkzMTRkZDVkZGE3%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D89cc3454413e6d48e05254cf27917076" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性③：Fabric で作成したテーブルをDatabricksで利用する（Databrickで閲覧・分析・編集可能） #BI - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめにFabricで作成したテーブルをDatabricksでもシームレスに参照・編集が可能できるのか？このような疑問を持つ方も少なくないと思います。そこで今回はFabric で作成したテー…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;
  

&lt;p&gt;▽ Previous Article&lt;br&gt;&lt;br&gt;
&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMCVFRiVCQyU5QWh1YiVFMyU4MiVCOSVFMyU4MyU4OCVFMyU4MyVBQyVFMyU4MyVCQyVFMyU4MiVCOCVFMyU4MSVBRSVFNyU5QiVBRSVFNyU5QSU4NCUyMC1EYXRhYnJpY2tzJTIwJUUzJTgxJUE3JUU0JUJEJTlDJUU2JTg4JTkwJUUzJTgxJTk3JUUzJTgxJTlGJUUzJTgzJTg2JUUzJTgzJUJDJUUzJTgzJTk2JUUzJTgzJUFCJUUzJTgyJTkyRmFicmljJUUyJTgwJUE2JnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9NTUxOTY4M2NiNmQ1MDg1ZTBkM2Y0MTEzYmFiNDc5MWY%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3Dd9e18c1adaf82cc577e53a8d830a3a93" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/48c7b1e54796f4f569f3" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性①：hubストレージの目的 -Databricks で作成したテーブルをFabricで利用する、Fabricで作成したテーブルをDatabricksで利用する- #Azure - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめにそれぞれの特徴がありつつも、大枠ではできることが似ているMicrosoft FabricとDatabricks。Azure Databricks Unity カタログのミラーリングを通…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>databricks</category>
      <category>azure</category>
      <category>microsoftfabric</category>
      <category>interoperability</category>
    </item>
    <item>
      <title>Fabric &amp; Databricks Interoperability (1): Purpose of Hub Storage for Table Sharing</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 09:38:46 +0000</pubDate>
      <link>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-1-purpose-of-hub-storage-for-table-sharing-30pl</link>
      <guid>https://dev.to/_d2a1ea24c442526a9777/fabric-databricks-interoperability-1-purpose-of-hub-storage-for-table-sharing-30pl</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;Although they each have their own characteristics, Microsoft Fabric and Databricks are broadly similar in what they can do.&lt;/p&gt;

&lt;p&gt;Through &lt;a href="https://learn.microsoft.com/ja-jp/fabric/database/mirrored-database/azure-databricks" rel="noopener noreferrer"&gt;Azure Databricks Unity Catalog mirroring&lt;/a&gt;, we are now able to reference Databricks-managed data in Fabric, but editing the data is still not possible.&lt;/p&gt;

&lt;p&gt;This brings up the following concerns:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Is it possible for our department to use tables managed by other departments in Databricks via our Fabric?&lt;/li&gt;
&lt;li&gt;I want to make the tables I create open for modification and reference, regardless of the tool used!&lt;/li&gt;
&lt;li&gt;We are currently using Databricks, but we might migrate to Fabric in the future... we want to maintain a vendor-free stance.&lt;/li&gt;
&lt;li&gt;Business-side employees use Fabric, but engineers use Databricks; there are times when we need to reference the same table.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In this article, I will introduce use cases for seamlessly utilizing Fabric and Databricks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Databricks in Fabric&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Using tables created in Fabric in Databricks&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The goal of this article:&lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6lp4nkc7c21gx1jxdmux.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6lp4nkc7c21gx1jxdmux.png" alt="image.png" width="800" height="259"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;:::note info&lt;br&gt;
This article consists of four parts:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Overview and purpose of interoperability (this article)&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer"&gt;Detailed setup of hub storage&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://qiita.com/ReijiOtake/items/86b44b2c30986c65db08" rel="noopener noreferrer"&gt;Using tables created in Fabric in Databricks&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://qiita.com/ReijiOtake/items/088abbd5f5ce06035501" rel="noopener noreferrer"&gt;Using tables created in Databricks in Fabric&lt;/a&gt;
:::&lt;/li&gt;
&lt;/ol&gt;
&lt;h1&gt;
  
  
  Prerequisite: Fabric and Databricks have similar functions... which one should we actually use?
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyimt27gcmj2amjs7y2pk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyimt27gcmj2amjs7y2pk.png" alt="image.png" width="800" height="427"&gt;&lt;/a&gt;&lt;br&gt;
Fabric and Databricks are both attracting attention as Lakehouse platforms that handle data end-to-end.&lt;/p&gt;

&lt;p&gt;As someone new to the industry, my first impression of using both was that &lt;strong&gt;"They can probably do about the same things?"&lt;/strong&gt;.&lt;br&gt;
They both support ETL processes and AI model creation.&lt;/p&gt;

&lt;p&gt;Fabric is appealing because of its beginner-friendly GUI, designed for intuitive operations.&lt;br&gt;
On the other hand, Databricks is more code-based, so it seems to require a slightly higher skill level.&lt;br&gt;
Additionally, Databricks offers more customization options for computer resources, and if you stop the cluster frequently, it can be more cost-effective than Fabric.&lt;/p&gt;

&lt;p&gt;I believe the choice between these platforms depends on whether you prioritize ease of use or flexibility.&lt;/p&gt;

&lt;p&gt;▽Reference&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/akihiro_suto/items/afaadd078c5a87772417" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Fadvent-calendar-ogp-background-7940cd1c8db80a7ec40711d90f43539e.jpg%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnMzLWFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkZxaWl0YS1pbWFnZS1zdG9yZSUyRjAlMkY4MDQ2NyUyRjUyNDI4NjA3YjlmOTgyMzMxNTQ2YjgyYTliN2E0OWQ2MjY2ZmZlY2QlMkZ4X2xhcmdlLnBuZyUzRjE3MzMzMDY4OTU_aXhsaWI9cmItNC4wLjAmYXI9MSUzQTEmZml0PWNyb3AmbWFzaz1lbGxpcHNlJmZtPXBuZzMyJnM9MTI1YTVlN2EzMmQ1YjAzZjY0YWEzYmFmMWRhOTE2ZGM%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253Dc73e11c66058241a1f1901961e3e4b21%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9QXp1cmUlMjBEYXRhYnJpY2tzJTIwJUUzJTgxJUE4JTIwTWljcm9zb2Z0JTIwRmFicmljJTIwJUUzJTgxJUFFJUU5JTk2JUEyJUU0JUJGJTgyJUUzJTgyJTkyJUU4JTgwJTgzJUUzJTgxJTg4JUUzJTgyJThCJUYwJTlGJUE3JTkwJnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMzQTNDM0MmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9MTliMTFiZDdiOTRjMWM5NDlhZjA2MzQyNDYzZmZkMzE%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBha2loaXJvX3N1dG8mdHh0LWNvbG9yPSUyMzNBM0MzQyZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT0zNiZ0eHQtcGFkPTAmcz1lODU5ZDAzMTNjM2Y0ZmIwMjBjYjJlZmIwMWNhNmUyNg%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3De21205b66d6c9e6d192abef7450edaa6" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/akihiro_suto/items/afaadd078c5a87772417" rel="noopener noreferrer" class="c-link"&gt;
          Azure Databricks と Microsoft Fabric の関係を考える🧐 #PowerBI - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめに本記事はDatabricks アドベントカレンダー2024 7日目の記事です。https://qiita.com/advent-calendar/2024/databricks/本記事…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Considering interoperability methods
&lt;/h1&gt;

&lt;p&gt;Here, I will explore methods for achieving interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71a20yxueztqddbfhaep.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F71a20yxueztqddbfhaep.png" alt="image.png" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  ① Unity Catalog Mirroring (Currently DML from Fabric to Databricks is not supported)
&lt;/h3&gt;

&lt;p&gt;Using &lt;a href="https://learn.microsoft.com/ja-jp/fabric/database/mirrored-database/azure-databricks" rel="noopener noreferrer"&gt;Azure Databricks Unity Catalog mirroring (preview)&lt;/a&gt;, it is possible to &lt;strong&gt;reference (SELECT statements) Databricks tables from Fabric&lt;/strong&gt;, but &lt;strong&gt;editing (DML statements)&lt;/strong&gt; is currently not supported.&lt;/p&gt;

&lt;p&gt;Thus, this method is not suitable for interoperability.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Please let me know if my understanding is incorrect 🙇&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  ② Specifying OneLake as an external location from Databricks (Currently not supported)
&lt;/h3&gt;

&lt;p&gt;I tried configuring an external location for OneLake via &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/connect/unity-catalog/cloud-storage/external-locations" rel="noopener noreferrer"&gt;cloud storage connection to Azure Databricks&lt;/a&gt;, but it is not currently supported.&lt;/p&gt;

&lt;p&gt;I was hoping this method would work, but unfortunately, it doesn't...&lt;br&gt;
So this method is also not suitable for interoperability.&lt;/p&gt;

&lt;p&gt;▽Reference&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ryoma-nagata/items/39fd52ab81015e3c9527#3-%E5%A4%96%E9%83%A8%E3%83%AD%E3%82%B1%E3%83%BC%E3%82%B7%E3%83%A7%E3%83%B3" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Fadvent-calendar-ogp-background-7940cd1c8db80a7ec40711d90f43539e.jpg%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnFpaXRhLWltYWdlLXN0b3JlLnMzLmFwLW5vcnRoZWFzdC0xLmFtYXpvbmF3cy5jb20lMkYwJTJGMjgxODE5JTJGcHJvZmlsZS1pbWFnZXMlMkYxNjQ0MzAyODYyP2l4bGliPXJiLTQuMC4wJmFyPTElM0ExJmZpdD1jcm9wJm1hc2s9ZWxsaXBzZSZmbT1wbmczMiZzPWU0NjRjMjU2MzJjZGJhZjliZWM1NWVkMTFmN2IzZjY2%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253Db1a7084443eb6442d076f44fedc0b53c%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9QXp1cmUlMjBEYXRhYnJpY2tzJTIwJUUzJTgxJThCJUUzJTgyJTg5JTIwT25lTGFrZSUyMCVFNCVCOCU4QSVFMyU4MSVBRSVFMyU4MyU4NyVFMyU4MyVCQyVFMyU4MiVCRiVFMyU4MSVBQiVFMyU4MiVBMiVFMyU4MiVBRiVFMyU4MiVCQiVFMyU4MiVCOSVFMyU4MSU5OSVFMyU4MiU4QiVFNiU5NiVCOSVFNiVCMyU5NSUyMDIwMjQlMkYxMiUyMCVFNyU4OSU4OCZ0eHQtYWxpZ249bGVmdCUyQ3RvcCZ0eHQtY29sb3I9JTIzM0EzQzNDJnR4dC1mb250PUhpcmFnaW5vJTIwU2FucyUyMFc2JnR4dC1zaXplPTU2JnR4dC1wYWQ9MCZzPTM2NTdmMjI4NjdiZmQ1ZGUwMTU0ODcwY2E4YWE1NTU2%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDByeW9tYS1uYWdhdGEmdHh0LWNvbG9yPSUyMzNBM0MzQyZ0eHQtZm9udD1IaXJhZ2lubyUyMFNhbnMlMjBXNiZ0eHQtc2l6ZT0zNiZ0eHQtcGFkPTAmcz0xNTRkMWM1OWE4OGFmOTAwYzU0MTRiZDgzODY0Y2FlNg%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3D4b37c422586c3667e7dfc23cb7a9875e" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ryoma-nagata/items/39fd52ab81015e3c9527#3-%E5%A4%96%E9%83%A8%E3%83%AD%E3%82%B1%E3%83%BC%E3%82%B7%E3%83%A7%E3%83%B3" rel="noopener noreferrer" class="c-link"&gt;
          Azure Databricks から OneLake 上のデータにアクセスする方法 2024/12 版 #Microsoft - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめにAzure Databricks の Unity Catalog ミラーリング を通して、Databricks の管理するデータについて Fabric で利用できるようになりましたが、Fa…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h3&gt;
  
  
  ③ Hub storage as a solution
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0go23z4rkvvzwxfxuemj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0go23z4rkvvzwxfxuemj.png" alt="image.png" width="800" height="446"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Since mirroring and external locations didn't work, I decided to store the actual tables in Azure Data Lake Gen 2 (ADLS2). &lt;br&gt;
By using a schema shortcut to ADLS2 from Fabric, and specifying ADLS2 as the storage location in Databricks' catalog, both Fabric and Databricks can perform SELECT and DML operations.&lt;/p&gt;

&lt;p&gt;This means that interoperability between Fabric and Databricks is now possible!&lt;/p&gt;

&lt;p&gt;From this point forward, I will refer to this ADLS2 as &lt;strong&gt;hub storage&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;▽For the specific setup method, refer to the following article:&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMSVFRiVCQyU5QWh1YiVFMyU4MiVCOSVFMyU4MyU4OCVFMyU4MyVBQyVFMyU4MyVCQyVFMyU4MiVCOCVFOCVBOCVBRCVFNSVBRSU5QSVFNiU5NiVCOSVFNiVCMyU5NSUyMC1EYXRhYnJpY2tzJTIwJUUzJTgxJUE3JUU0JUJEJTlDJUU2JTg4JTkwJUUzJTgxJTk3JUUzJTgxJTlGJUUzJTgzJTg2JUUzJTgzJUJDJUUzJTgzJTk2JUUzJTgzJUFCJUUzJTgyJTkyRmFiJUUyJTgwJUE2JnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9MjM4ZTdmYWNlYmVkYzQ4ZDYzZmQzZTNhYjkzNzM5NjQ%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3Da729c7480d7acdfdd31cc6768c35f55a" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性②：hubストレージ設定方法 -Databricks で作成したテーブルをFabric で利用する、Fabric で作成したテーブルをDatabricksで利用する- #Azure - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめに今回はDatabricks で作成したテーブルをFabric で利用するFabric で作成したテーブルをDatabricksで利用するというユースケースを実施するための設定方法につ…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Hub storage works thanks to Delta Lake
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp1x17twwkdfvc8fgxlcg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp1x17twwkdfvc8fgxlcg.png" alt="image.png" width="370" height="302"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As explained above, hub storage allows interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;But why does this interoperability work?&lt;/p&gt;

&lt;p&gt;The key lies in &lt;strong&gt;Delta Lake&lt;/strong&gt;, the mechanism behind it.&lt;/p&gt;

&lt;h3&gt;
  
  
  The mechanism of Delta Lake
&lt;/h3&gt;

&lt;p&gt;Delta Lake is an open-source storage layer that provides transaction and schema management on top of a data lake. It uses a combination of Parquet and JSON as its underlying data formats. Parquet is a columnar compression format that enables fast queries and data compression, while JSON is used as a transaction log to record data change history and versioning.&lt;/p&gt;

&lt;p&gt;By leveraging the Delta Lake mechanism, hub storage enables advanced data sharing and operations. When using Fabric or Databricks, it’s crucial to understand the underlying infrastructure to fully take advantage of the features provided by Delta Lake.&lt;/p&gt;

&lt;p&gt;▽For more details on Delta Lake, refer to the official documentation:&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/delta/" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Flearn.microsoft.com%2Fen-us%2Fmedia%2Fopen-graph-image.png" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/delta/" rel="noopener noreferrer" class="c-link"&gt;
          Delta Lake とは - Azure Databricks | Microsoft Learn
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          Databricks レイクハウスに電源を供給するために使用される Delta Lake ストレージ プロトコルについて説明します。
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
        learn.microsoft.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;In this article, I summarized the purpose and methods of interoperability between Fabric and Databricks.&lt;/p&gt;

&lt;p&gt;The next article will provide a detailed guide for setting up hub storage.&lt;/p&gt;

&lt;p&gt;▽Next article:&lt;/p&gt;


&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
      &lt;div class="c-embed__cover"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" class="c-link s:max-w-50 align-middle" rel="noopener noreferrer"&gt;
          &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fqiita-user-contents.imgix.net%2Fhttps%253A%252F%252Fqiita-user-contents.imgix.net%252Fhttps%25253A%25252F%25252Fcdn.qiita.com%25252Fassets%25252Fpublic%25252Farticle-ogp-background-afbab5eb44e0b055cce1258705637a91.png%253Fixlib%253Drb-4.0.0%2526w%253D1200%2526blend64%253DaHR0cHM6Ly9xaWl0YS11c2VyLXByb2ZpbGUtaW1hZ2VzLmltZ2l4Lm5ldC9odHRwcyUzQSUyRiUyRnNlY3VyZS5ncmF2YXRhci5jb20lMkZhdmF0YXIlMkY2YWQ3NTBlZjVkZWViOGYzMzdjMjI1YmYyYzE3NmMyZj9peGxpYj1yYi00LjAuMCZhcj0xJTNBMSZmaXQ9Y3JvcCZtYXNrPWVsbGlwc2UmZm09cG5nMzImcz1iZGY4MTY2YmY2MDEyNTVlNDBmYWRlYzdhNzQ2YTExOQ%2526blend-x%253D120%2526blend-y%253D467%2526blend-w%253D82%2526blend-h%253D82%2526blend-mode%253Dnormal%2526s%253D1d50db4d1ee7e3e652ef7efcf36b7dc8%3Fixlib%3Drb-4.0.0%26w%3D1200%26fm%3Djpg%26mark64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTk2MCZoPTMyNCZ0eHQ9RmFicmljJUUzJTgxJUE4RGF0YWJyaWNrcyVFMyU4MSVBRSVFNyU5QiVCOCVFNCVCQSU5MiVFOSU4MSU4QiVFNyU5NCVBOCVFNiU4MCVBNyVFMiU5MSVBMSVFRiVCQyU5QWh1YiVFMyU4MiVCOSVFMyU4MyU4OCVFMyU4MyVBQyVFMyU4MyVCQyVFMyU4MiVCOCVFOCVBOCVBRCVFNSVBRSU5QSVFNiU5NiVCOSVFNiVCMyU5NSUyMC1EYXRhYnJpY2tzJTIwJUUzJTgxJUE3JUU0JUJEJTlDJUU2JTg4JTkwJUUzJTgxJTk3JUUzJTgxJTlGJUUzJTgzJTg2JUUzJTgzJUJDJUUzJTgzJTk2JUUzJTgzJUFCJUUzJTgyJTkyRmFiJUUyJTgwJUE2JnR4dC1hbGlnbj1sZWZ0JTJDdG9wJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9NTYmdHh0LXBhZD0wJnM9MjM4ZTdmYWNlYmVkYzQ4ZDYzZmQzZTNhYjkzNzM5NjQ%26mark-x%3D120%26mark-y%3D112%26blend64%3DaHR0cHM6Ly9xaWl0YS11c2VyLWNvbnRlbnRzLmltZ2l4Lm5ldC9-dGV4dD9peGxpYj1yYi00LjAuMCZ3PTgzOCZoPTU4JnR4dD0lNDBSZWlqaU90YWtlJnR4dC1jb2xvcj0lMjMxRTIxMjEmdHh0LWZvbnQ9SGlyYWdpbm8lMjBTYW5zJTIwVzYmdHh0LXNpemU9MzYmdHh0LXBhZD0wJnM9YzZmYWVkZTgyMDU5YzQwNDk3NWU4MjAwNDZmMjA4NGM%26blend-x%3D242%26blend-y%3D480%26blend-w%3D838%26blend-h%3D46%26blend-fit%3Dcrop%26blend-crop%3Dleft%252Cbottom%26blend-mode%3Dnormal%26s%3Da729c7480d7acdfdd31cc6768c35f55a" height="420" class="m-0" width="800"&gt;
        &lt;/a&gt;
      &lt;/div&gt;
    &lt;div class="c-embed__body"&gt;
      &lt;h2 class="fs-xl lh-tight"&gt;
        &lt;a href="https://qiita.com/ReijiOtake/items/7ce8e1a743d05efe925c" rel="noopener noreferrer" class="c-link"&gt;
          FabricとDatabricksの相互運用性②：hubストレージ設定方法 -Databricks で作成したテーブルをFabric で利用する、Fabric で作成したテーブルをDatabricksで利用する- #Azure - Qiita
        &lt;/a&gt;
      &lt;/h2&gt;
        &lt;p class="truncate-at-3"&gt;
          はじめに今回はDatabricks で作成したテーブルをFabric で利用するFabric で作成したテーブルをDatabricksで利用するというユースケースを実施するための設定方法につ…
        &lt;/p&gt;
      &lt;div class="color-secondary fs-s flex items-center"&gt;
          &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fcdn.qiita.com%2Fassets%2Ffavicons%2Fpublic%2Fproduction-c620d3e403342b1022967ba5e3db1aaa.ico" width="120" height="120"&gt;
        qiita.com
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;


</description>
      <category>interoperability</category>
      <category>azure</category>
      <category>databricks</category>
      <category>microsoftfabric</category>
    </item>
    <item>
      <title>How a New Graduate with No Industry Experience Passed the Databricks Certified Data Engineer Associate Exam in 3 Weeks</title>
      <dc:creator>Reiji Otake</dc:creator>
      <pubDate>Sun, 16 Feb 2025 07:56:12 +0000</pubDate>
      <link>https://dev.to/_d2a1ea24c442526a9777/how-a-new-graduate-with-no-industry-experience-passed-the-databricks-certified-data-engineer-j8</link>
      <guid>https://dev.to/_d2a1ea24c442526a9777/how-a-new-graduate-with-no-industry-experience-passed-the-databricks-certified-data-engineer-j8</guid>
      <description>&lt;h1&gt;
  
  
  Introduction
&lt;/h1&gt;

&lt;p&gt;As the title suggests, I am posting this as a memorandum to keep a record of my learning.&lt;br&gt;&lt;br&gt;
It's just a simple summary, but I hope it can be helpful for your studies as well.  &lt;/p&gt;

&lt;p&gt;If you have any additional questions, please feel free to leave a comment.  &lt;/p&gt;

&lt;h1&gt;
  
  
  Personal Impressions &amp;amp; Key Points
&lt;/h1&gt;

&lt;h3&gt;
  
  
  Grasping the Overall Picture
&lt;/h3&gt;

&lt;p&gt;By reading the &lt;a href="https://www.amazon.co.jp/%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9-%E3%82%AF%E3%82%A4%E3%83%83%E3%82%AF%E3%82%B9%E3%82%BF%E3%83%BC%E3%83%88%E3%82%AC%E3%82%A4%E3%83%89-%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9%E3%83%BB%E3%82%B8%E3%83%A3%E3%83%91%E3%83%B3-ebook/dp/B09V1YXFVQ/ref=sr_1_3?__mk_ja_JP=%E3%82%AB%E3%82%BF%E3%82%AB%E3%83%8A&amp;amp;crid=1B8GH67RI7HP1&amp;amp;dib=eyJ2IjoiMSJ9.QjgTf6G7XxomSI9f9hCE1K2qYT1U7IZIh47ExDCXNszKVaBTr_Z4GGJHOz4CG5IwyUn5ieAazLo8vLyGko-HKFbvsy69Wv-5RtjgXMhJ60h_C4-kOkMUPFbeuY7YBT6y0BJEw4UoKmML9hCZntFsVOsfsey_Pvw2CXddGPhE_rqzdqQwHkdR_I4c9vNxsOdEj1INDE93secmQ3SOoA9KEHxTGPWeWe1phgKmfwjolec6OBZq1QpqcyYztj6M0oK9eIlt3QVlNcp4QBaIZWtvMj_sy_DhYwd5FGPITqR9cyP-beIYhV1_NZ8j6RiNxzK9Y-xUVdY8M_-CePTy5bRqKQxtj-IN1fEItWjvniROMfwARoWjOqiGINi_pYxtMhsDXVxxU4Mu2LOTXJ8BRTnu2nMmzxVAFivSK8z-kRlJxj5dxpE_xV5aWc5uclkGeL8a.PXtPo4zK3rMLq5npaos2WoNM2L192fU3bhVfq5eskkE&amp;amp;dib_tag=se&amp;amp;keywords=Databricks&amp;amp;qid=1735199303&amp;amp;sprefix=databricks%2Caps%2C170&amp;amp;sr=8-3" rel="noopener noreferrer"&gt;Databricks Quick Start Guide&lt;/a&gt;, you can get an overview of the exam's flow and key points.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Dealing with SQL Questions
&lt;/h3&gt;

&lt;p&gt;For those who have obtained Oracle Silver, the SQL section is relatively easy.&lt;br&gt;&lt;br&gt;
There were more than five SQL questions on the exam.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Utilizing the Databricks Environment
&lt;/h3&gt;

&lt;p&gt;If you have access to a Databricks environment, actually using it and working through tutorials will deepen your understanding.&lt;br&gt;&lt;br&gt;
Fortunately, I had access to a study environment where I could freely experiment.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Leveraging Official Documentation
&lt;/h3&gt;

&lt;p&gt;Going back to the official documentation is ultimately the fastest way to find answers.&lt;br&gt;&lt;br&gt;
In my case, I had a shallow understanding of Unity Catalog, DLT, structured streaming, and Git, so I studied by reading the official documentation.  &lt;/p&gt;

&lt;h1&gt;
  
  
  Study Timeline
&lt;/h1&gt;

&lt;h3&gt;
  
  
  24/12/23 Asked Senior Employees Who Passed for Recommended Study Materials
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://qiita.com/nttd-saitouyun/items/e7d1ca77e23b8e635518#%E3%83%87%E3%83%BC%E3%82%BF%E3%82%A8%E3%83%B3%E3%82%B8%E3%83%8B%E3%82%A2-%E3%82%A2%E3%82%BD%E3%82%B7%E3%82%A8%E3%82%A4%E3%83%88--databricks-certified-data-engineer-associate" rel="noopener noreferrer"&gt;Databricks Certifications: I Took Them All and Summarized Systematically&lt;/a&gt;  &lt;/p&gt;

&lt;h3&gt;
  
  
  2024/12/27 Databricks Quick Start Guide
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.amazon.co.jp/%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9-%E3%82%AF%E3%82%A4%E3%83%83%E3%82%AF%E3%82%B9%E3%82%BF%E3%83%BC%E3%83%88%E3%82%AC%E3%82%A4%E3%83%89-%E3%83%87%E3%83%BC%E3%82%BF%E3%83%96%E3%83%AA%E3%83%83%E3%82%AF%E3%82%B9%E3%83%BB%E3%82%B8%E3%83%A3%E3%83%91%E3%83%B3-ebook/dp/B09V1YXFVQ/ref=sr_1_3?__mk_ja_JP=%E3%82%AB%E3%82%BF%E3%82%AB%E3%83%8A&amp;amp;crid=1B8GH67RI7HP1&amp;amp;dib=eyJ2IjoiMSJ9.QjgTf6G7XxomSI9f9hCE1K2qYT1U7IZIh47ExDCXNszKVaBTr_Z4GGJHOz4CG5IwyUn5ieAazLo8vLyGko-HKFbvsy69Wv-5RtjgXMhJ60h_C4-kOkMUPFbeuY7YBT6y0BJEw4UoKmML9hCZntFsVOsfsey_Pvw2CXddGPhE_rqzdqQwHkdR_I4c9vNxsOdEj1INDE93secmQ3SOoA9KEHxTGPWeWe1phgKmfwjolec6OBZq1QpqcyYztj6M0oK9eIlt3QVlNcp4QBaIZWtvMj_sy_DhYwd5FGPITqR9cyP-beIYhV1_NZ8j6RiNxzK9Y-xUVdY8M_-CePTy5bRqKQxtj-IN1fEItWjvniROMfwARoWjOqiGINi_pYxtMhsDXVxxU4Mu2LOTXJ8BRTnu2nMmzxVAFivSK8z-kRlJxj5dxpE_xV5aWc5uclkGeL8a.PXtPo4zK3rMLq5npaos2WoNM2L192fU3bhVfq5eskkE&amp;amp;dib_tag=se&amp;amp;keywords=Databricks&amp;amp;qid=1735199303&amp;amp;sprefix=databricks%2Caps%2C170&amp;amp;sr=8-3" rel="noopener noreferrer"&gt;Databricks Quick Start Guide&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;Recommended for getting a broad overview.  &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Recommended Points&lt;/strong&gt;  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Affordable Price&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;The Kindle version is available for 99 yen (this is the regular price, not a sale).
&lt;/li&gt;
&lt;li&gt;The same content is available for free on Qiita, but the Kindle version is compiled into a single volume, making it significantly easier to read. (Well worth the 99 yen.)
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Reliable Information Source&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;Published by Databricks Japan, ensuring high reliability.
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Easy-to-Understand Explanations&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;Clearly explains the background and history of Lakehouse.
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Practical Content&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;Includes ETL hands-on exercises, making it easy to apply in real-world scenarios.
&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;/blockquote&gt;

&lt;p&gt;　&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Caution&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Since it was published in 2022, some information may be slightly outdated.  &lt;/p&gt;
&lt;/blockquote&gt;

&lt;h3&gt;
  
  
  2025/1/7~8 Official Practice Questions
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://qiita.com/kohei-arai/items/5b54a89cbaec801f1972" rel="noopener noreferrer"&gt;Translated &amp;amp; Explained Databricks Certified Data Engineer Associate Practice Questions&lt;/a&gt;&lt;br&gt;&lt;br&gt;
Questions 1-30 have been translated into Japanese and explained.  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://qiita.com/nakazax/items/8f35cecb8f658b35e314" rel="noopener noreferrer"&gt;Unofficial Explanation of Databricks Certified Data Engineer Associate Practice Exam Answers (As of January 2024)&lt;/a&gt;&lt;br&gt;&lt;br&gt;
Questions 31-45 are in English, but explanations are provided in Japanese.  &lt;/p&gt;

&lt;h3&gt;
  
  
  2025/1/8~6 Udemy Course Recommended by Senior Employees
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.udemy.com/course/practice-exams-databricks-certified-data-engineer-associate/?couponCode=ST16MT28125CROW" rel="noopener noreferrer"&gt;Practice Exams: Databricks Certified Data Engineer Associate&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;Accuracy Rates:&lt;br&gt;&lt;br&gt;
Practice 1: 51% (23/45)&lt;br&gt;&lt;br&gt;
Practice 2: 62% (28/45)  &lt;/p&gt;

&lt;h3&gt;
  
  
  2025/1/16~19 Struggled with Structured Streaming, Delta Live Tables, and Git as a New Graduate
&lt;/h3&gt;

&lt;p&gt;I found them too difficult to understand, so I searched through documentation and tutorials.  &lt;/p&gt;

&lt;h4&gt;
  
  
  Relevant Resources:
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/delta-live-tables/tutorial-pipelines.html#language-sql" rel="noopener noreferrer"&gt;Delta Live Tables SQL Tutorial&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/structured-streaming/incremental.html" rel="noopener noreferrer"&gt;Incremental Processing with Structured Streaming&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/structured-streaming/tutorial.html" rel="noopener noreferrer"&gt;Structured Streaming Tutorial&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/jobs/jobs-quickstart.html" rel="noopener noreferrer"&gt;Quickstart Guide for Databricks Jobs&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/jobs/index.html" rel="noopener noreferrer"&gt;Overview of Databricks Jobs&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://learn.microsoft.com/ja-jp/azure/databricks/delta-live-tables/updates#development-and-production-modes" rel="noopener noreferrer"&gt;Azure Databricks Delta Live Tables Update Modes&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.databricks.com/jp/spark/about" rel="noopener noreferrer"&gt;Introduction to Apache Spark&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://docs.databricks.com/ja/delta-live-tables/index.html" rel="noopener noreferrer"&gt;Databricks Delta Live Tables Documentation&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.databricks.com/jp/product/delta-live-tables" rel="noopener noreferrer"&gt;Databricks Delta Live Tables Overview&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://qiita.com/taka_yayoi/items/e881769270f0ec0b7d06#databricks%E3%82%B8%E3%83%A7%E3%83%96%E3%81%A8%E3%81%AF" rel="noopener noreferrer"&gt;What is Databricks Jobs?&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://qiita.com/ryoma-nagata/items/74e1bd9ebaf0413c9fd6" rel="noopener noreferrer"&gt;Understanding Databricks Workflows&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Git/GitHub Resources:
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://www.youtube.com/watch?v=Dqgyc_S3L0s" rel="noopener noreferrer"&gt;Git/GitHub Introduction: Learn the Basics in 30 Minutes&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.youtube.com/watch?v=1l8oBEown8c" rel="noopener noreferrer"&gt;What is Git/GitHub? 10-Minute Beginner Guide&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2025/1/20 Retook the Official Practice Exam
&lt;/h3&gt;

&lt;p&gt;Accuracy Rate: &lt;strong&gt;86%&lt;/strong&gt;  &lt;/p&gt;

&lt;h3&gt;
  
  
  2025/1/20 Registered for the Exam
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.databricks.com/jp/learn/certification/data-engineer-associate" rel="noopener noreferrer"&gt;Databricks Certified Data Engineer Associate Exam Registration&lt;/a&gt;  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Warning:&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
I do not recommend using a JCB card for payment.&lt;br&gt;&lt;br&gt;
(My confirmation email never arrived, but the payment was deducted. The follow-up with support was very time-consuming.)  &lt;/p&gt;

&lt;h3&gt;
  
  
  1/22 Passed the Exam!! 🎉
&lt;/h3&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Even as an industry newcomer with only 9 months of experience, I was able to pass on the first attempt!&lt;br&gt;&lt;br&gt;
I encourage you to give it a try!  &lt;/p&gt;

&lt;p&gt;Since then, I have continued using Databricks in my work and have found that the knowledge gained from certification studies has been directly applicable.  &lt;/p&gt;

&lt;p&gt;This was just a brief summary, but I hope it helps with your studies!  &lt;/p&gt;

</description>
      <category>databricks</category>
      <category>beginners</category>
      <category>certification</category>
    </item>
  </channel>
</rss>
