<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Ansam Yousry</title>
    <description>The latest articles on DEV Community by Ansam Yousry (@ansamal93038737).</description>
    <link>https://dev.to/ansamal93038737</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1106622%2F9667ec72-85f7-4577-8c71-f3a36672950d.png</url>
      <title>DEV Community: Ansam Yousry</title>
      <link>https://dev.to/ansamal93038737</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ansamal93038737"/>
    <language>en</language>
    <item>
      <title>Differences between dbt Core and dbt Cloud</title>
      <dc:creator>Ansam Yousry</dc:creator>
      <pubDate>Thu, 22 Jun 2023 19:19:15 +0000</pubDate>
      <link>https://dev.to/ansamal93038737/differences-between-dbt-core-and-dbt-cloud-3k7b</link>
      <guid>https://dev.to/ansamal93038737/differences-between-dbt-core-and-dbt-cloud-3k7b</guid>
      <description>&lt;p&gt;dbt (Data Build Tool) is an open-source tool that enables data analysts and engineers to build data transformation pipelines. There are two main versions of dbt: dbt Core and dbt Cloud.&lt;/p&gt;

&lt;p&gt;In this article, we'll discuss the differences between dbt Core and dbt Cloud. For more detailed information, check out this &lt;a href="https://medium.com/illumination/differences-between-dbt-core-and-dbt-cloud-4e4064a97aaf"&gt;medium post&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  dbt Core
&lt;/h2&gt;

&lt;p&gt;dbt Core is the open-source version of dbt that can be installed and run locally. It is a command-line tool that allows you to write SQL code to transform your data. dbt Core has the following features:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Can be run on your local machine or on a server.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Allows you to write SQL code to transform your data.&lt;br&gt;
Integrates with popular data warehouses such as Snowflake, BigQuery, and Redshift.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Supports version control using Git.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Provides testing functionality to ensure the quality and accuracy of your data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Allows you to document your data pipelines using Markdown.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;dbt Core is a powerful tool that provides a lot of functionality for transforming your data but requires some setup and configuration.&lt;/p&gt;

&lt;h2&gt;
  
  
  dbt Cloud
&lt;/h2&gt;

&lt;p&gt;dbt Cloud is a cloud-based platform that provides a user-friendly interface for building and managing data transformation pipelines. It is a paid service that provides additional features on top of dbt Core. dbt Cloud has the following features:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Provides a web-based interface for building and managing data pipelines.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Automates the deployment and running of your dbt projects.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Provides a collaborative environment for teams to work on data pipelines together.&lt;br&gt;
Includes a built-in scheduler for running data pipelines on a schedule.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Provides additional functionality for monitoring and alerting on data pipelines.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;dbt Cloud is a powerful tool that provides a lot of functionality for building and managing data pipelines without requiring a lot of setup and configuration. It is especially useful for teams who want to collaborate on data pipelines and don't want to manage their own infrastructure.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;dbt is a powerful tool for building and managing data transformation pipelines. dbt Core is the open-source version that can be installed locally, while dbt Cloud is a cloud-based platform that provides additional features on top of dbt Core. Which version you choose depends on your needs and preferences. If you're just getting started with dbt, dbt Core is a great place to start. If you're working on a team or need additional functionality, dbt Cloud may be a better fit.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>data</category>
      <category>ai</category>
    </item>
    <item>
      <title>Apache Airflow: A Beginner's Guide</title>
      <dc:creator>Ansam Yousry</dc:creator>
      <pubDate>Thu, 22 Jun 2023 18:50:23 +0000</pubDate>
      <link>https://dev.to/ansamal93038737/apache-airflow-a-beginners-guide-6nh</link>
      <guid>https://dev.to/ansamal93038737/apache-airflow-a-beginners-guide-6nh</guid>
      <description>&lt;p&gt;Apache Airflow is an open-source platform to programmatically author, schedule, and monitor workflows. It was developed by Airbnb to manage their data pipelines. The platform is used to create complex workflows composed of multiple tasks, and it allows for easy maintenance and monitoring of these workflows.&lt;/p&gt;

&lt;p&gt;In this article, we'll cover the basics of Apache Airflow and how to install it using Docker. For more detailed instructions, check out this &lt;a href="https://medium.com/illumination/airflow-from-zero-and-how-to-install-airflow-in-docker-fb5c4a0f992b"&gt;medium post&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Apache Airflow?
&lt;/h2&gt;

&lt;p&gt;Apache Airflow is a platform that allows you to create, schedule, and monitor workflows. Workflows are created using Python code and are made up of multiple tasks that are executed independently. Each task can be run on a different machine, making it possible to scale workflows as needed.&lt;/p&gt;

&lt;p&gt;Airflow has a web interface that allows you to monitor the status of your workflows, view logs, and see the dependencies between tasks. It also has a powerful scheduler that can handle complex dependencies and retry failed tasks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Installing Apache Airflow with Docker
&lt;/h2&gt;

&lt;p&gt;Installing Apache Airflow can be a complex process, but using Docker can make it much simpler. Docker is a platform that allows you to create, deploy, and run applications in containers. Containers are lightweight, portable, and self-contained, making them ideal for running complex applications like Apache Airflow.&lt;/p&gt;

&lt;p&gt;To install Apache Airflow with Docker, follow these steps:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Install Docker on your machine. You can download Docker from the official website.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Create a directory for your Airflow installation. This directory will contain your Airflow configuration files and your Docker Compose file.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Create a Docker Compose file in your Airflow directory. This file will define the services that make up your Airflow installation. Here's an example:&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;`version: '2'&lt;/p&gt;

&lt;p&gt;services:&lt;br&gt;
  postgres:&lt;br&gt;
    image: postgres:9.6&lt;br&gt;
    environment:&lt;br&gt;
      POSTGRES_USER: airflow&lt;br&gt;
      POSTGRES_PASSWORD: airflow&lt;br&gt;
      POSTGRES_DB: airflow&lt;br&gt;
    volumes:&lt;br&gt;
      - postgres_data:/var/lib/postgresql/data&lt;/p&gt;

&lt;p&gt;webserver:&lt;br&gt;
    image: apache/airflow:2.1.2&lt;br&gt;
    depends_on:&lt;br&gt;
      - postgres&lt;br&gt;
    environment:&lt;br&gt;
      - LOAD_EX=n&lt;br&gt;
      - FERNET_KEY=your_fernet_key_here&lt;br&gt;
      - EXECUTOR=Local&lt;br&gt;
      - POSTGRES_USER=airflow&lt;br&gt;
      - POSTGRES_PASSWORD=airflow&lt;br&gt;
      - POSTGRES_DB=airflow&lt;br&gt;
    ports:&lt;br&gt;
      - "8080:8080"&lt;br&gt;
    volumes:&lt;br&gt;
      - ./dags:/opt/airflow/dags&lt;br&gt;
      - ./logs:/opt/airflow/logs&lt;br&gt;
      - ./plugins:/opt/airflow/plugins&lt;br&gt;
    command: webserver&lt;/p&gt;

&lt;p&gt;volumes:&lt;br&gt;
  postgres_data:`&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Start your Airflow installation by running docker-compose up in your Airflow directory. This will start the Postgres database and the Airflow webserver.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Access the Airflow web interface by navigating to &lt;a href="http://localhost:8080"&gt;http://localhost:8080&lt;/a&gt; in your web browser. You should see the Airflow dashboard, which shows the status of your workflows.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Apache Airflow is a powerful platform for managing complex workflows. Using Docker to install Airflow can make the process much simpler and more manageable. With Airflow, you can create, schedule, and monitor your workflows with ease, allowing you to focus on your data instead of worrying about the infrastructure.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
