<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Paige Tran</title>
    <description>The latest articles on DEV Community by Paige Tran (@bttminhphuc).</description>
    <link>https://dev.to/bttminhphuc</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F898300%2F6c562fa7-d678-427b-b905-dd9f06647d73.jpeg</url>
      <title>DEV Community: Paige Tran</title>
      <link>https://dev.to/bttminhphuc</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/bttminhphuc"/>
    <language>en</language>
    <item>
      <title>Solving the Top 5 DBT Problems with Edge Cases: Streamlining Data Build Processes</title>
      <dc:creator>Paige Tran</dc:creator>
      <pubDate>Sat, 27 May 2023 19:49:54 +0000</pubDate>
      <link>https://dev.to/bttminhphuc/solving-the-top-5-dbt-problems-with-edge-cases-streamlining-data-build-processes-2g4k</link>
      <guid>https://dev.to/bttminhphuc/solving-the-top-5-dbt-problems-with-edge-cases-streamlining-data-build-processes-2g4k</guid>
      <description>&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;:&lt;br&gt;
DBT (Data Build Tool) is a popular SQL modeling tool that enables data engineers and analysts to build, test, and deploy reliable data pipelines. While DBT is a powerful tool, it can encounter certain challenges during its implementation and usage. In this article, we will explore the top 5 problems faced in DBT and discuss how leveraging edge cases can provide effective solutions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Complex Data Transformations:&lt;/strong&gt;&lt;br&gt;
One of the common challenges in DBT is handling complex data transformations. Edge cases can be employed by creating specialized transformations for unique scenarios. By identifying and incorporating edge cases into the data build process, developers can handle complex data transformations more efficiently. These edge cases serve as specific examples that test the robustness and scalability of the transformations, ensuring accurate results.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;-- Example of a complex data transformation using edge cases
-- Edge case: Handling null values in the transformation
WITH edge_case AS (
  SELECT 
    column1,
    CASE
      WHEN column2 IS NULL THEN 'N/A'
      ELSE column2
    END AS transformed_column
  FROM source_table
)
SELECT *
FROM edge_case;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;2. Handling Large Volumes of Data:&lt;/strong&gt;&lt;br&gt;
Scaling DBT to handle large volumes of data can be a daunting task. Edge cases can be used to simulate and validate the performance of the data build process under extreme data conditions. By running DBT on subsets of the entire dataset or synthetic datasets representing edge cases, developers can optimize and fine-tune the SQL queries, models, and configurations to handle large-scale data processing effectively.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;-- Example of optimizing a DBT model for large data volumes using edge cases
-- Edge case: Simulating large dataset for testing and performance tuning
WITH edge_case AS (
  SELECT *
  FROM source_table
  WHERE created_date &amp;gt;= '2023-01-01' -- Edge case: Focus on a specific date range
)
SELECT *
FROM edge_case;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;3. Dependency Management:&lt;/strong&gt;&lt;br&gt;
Managing dependencies between DBT models is crucial for maintaining a well-structured and efficient data pipeline.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;-- Example of resolving complex dependencies between DBT models using edge cases
-- Edge case: Addressing circular dependencies
-- Model A
SELECT *
FROM model_b

-- Model B
SELECT *
FROM model_a
JOIN model_c ON model_a.id = model_c.id;

-- Model C
SELECT *
FROM source_table;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Edge cases can be leveraged to identify and resolve complex dependencies. By introducing edge cases that cover scenarios with intricate relationships between models, developers can identify potential issues, such as circular dependencies or performance bottlenecks, and implement necessary optimizations to ensure smooth dependency management.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Testing and Validation:&lt;/strong&gt;&lt;br&gt;
DBT offers built-in testing capabilities, but ensuring comprehensive testing and validation can still be a challenge.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;-- Example of testing and validating DBT models using edge cases
-- Edge case: Testing edge scenarios with extreme values
-- Model A
SELECT *
FROM source_table
WHERE column1 &amp;gt;= 0 -- Edge case: Testing positive values

-- Model B
SELECT *
FROM model_a
WHERE column2 IS NULL; -- Edge case: Testing null values
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Edge cases can be utilized to design test cases that cover a wide range of scenarios, including edge scenarios that test the limits of the data build process. By incorporating edge cases into the testing strategy, developers can identify and fix issues early on, ensuring the accuracy and reliability of the data pipelines.&lt;/p&gt;

&lt;p&gt;*&lt;em&gt;5. Collaboration and Version *&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Control:&lt;br&gt;
Collaboration and version control are crucial aspects of maintaining a well-maintained and scalable DBT project. Edge cases can help address collaboration challenges by creating branches for different edge case scenarios. These branches can be used to test and validate alternative approaches, allowing for experimentation without affecting the main production branch. By utilizing version control systems effectively and incorporating edge cases, developers can streamline collaboration and maintain a reliable history of changes.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;-- Example of using version control and collaboration with DBT using edge cases
-- Edge case: Creating a separate branch for experimenting with alternative approaches
-- Main branch
SELECT *
FROM source_table
WHERE column1 = 'A';

-- Experiment branch (edge case)
SELECT *
FROM source_table
WHERE column1 = 'B';

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;These coding examples illustrate how edge cases can be incorporated into DBT projects to address the top 5 problems. By considering specific scenarios, such as handling null values, simulating large datasets, resolving complex dependencies, testing extreme values, and utilizing separate branches for experimentation, developers can enhance the effectiveness and reliability of their DBT implementations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion:&lt;/strong&gt;&lt;br&gt;
DBT is a powerful data modeling tool that streamlines data build processes, but it can encounter various challenges. Leveraging edge cases can provide effective solutions to these problems. By incorporating specialized transformations, simulating large data volumes, managing complex dependencies, implementing comprehensive testing, and optimizing collaboration and version control, developers can enhance the reliability, scalability, and efficiency of DBT projects. With the use of edge cases, data engineers and analysts can build robust data pipelines that meet the evolving needs of data-driven organizations.&lt;/p&gt;

</description>
      <category>dbt</category>
      <category>database</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Optimizing DBT Modeling Code: A Guide to DBT Setup and Hand-on commands</title>
      <dc:creator>Paige Tran</dc:creator>
      <pubDate>Fri, 17 Feb 2023 13:17:33 +0000</pubDate>
      <link>https://dev.to/bttminhphuc/optimizing-dbt-modeling-code-a-guide-to-dbt-setup-and-hand-on-commands-2h2</link>
      <guid>https://dev.to/bttminhphuc/optimizing-dbt-modeling-code-a-guide-to-dbt-setup-and-hand-on-commands-2h2</guid>
      <description>&lt;h2&gt;
  
  
  1.   What is &lt;a href="https://www.getdbt.com/" rel="noopener noreferrer"&gt;DBT&lt;/a&gt;?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.getdbt.com/" rel="noopener noreferrer"&gt;DBT&lt;/a&gt; (Data Build Tool) is an open-source tool used for data modeling, data transformation, and data management. It is designed to help data teams build, maintain, and document their data models and transformations in an organized and efficient manner.&lt;/p&gt;

&lt;p&gt;DBT helps automate data processing pipelines and provides a framework for modeling data in a declarative way. It supports a variety of databases, including Amazon &lt;a href="https://aws.amazon.com/ru/redshift/" rel="noopener noreferrer"&gt;Redshift&lt;/a&gt;, &lt;a href="https://cloud.google.com/bigquery/" rel="noopener noreferrer"&gt;Google BigQuery&lt;/a&gt;, &lt;a href="https://signup.snowflake.com/" rel="noopener noreferrer"&gt;Snowflake&lt;/a&gt;, and more, making it a versatile tool for managing data in various environments. &lt;/p&gt;

&lt;h2&gt;
  
  
  2.   What code editor is better for DBT?
&lt;/h2&gt;

&lt;p&gt;There are several code editors that can be used for DBT, however, the most popular ones are Sublime Text and Visual Studio Code (VS Code). Here is a comparison between the two:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcfxi1nz6qyt89yf0i53a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcfxi1nz6qyt89yf0i53a.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I personally use VS code and I find it very useful. All in all, a combination of DBT and VS Code make a suitable environment for developing and maintaining high-quality data models with powerful debugging and testing capabilities,.&lt;/p&gt;

&lt;h2&gt;
  
  
  3.   Setting up DBT with a data warehouse
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Step 1: Install DBT&lt;/strong&gt;&lt;br&gt;
 DBT is a command-line tool that can be installed using pip, a Python package manager. You can install DBT by running the following command in your terminal: &lt;/p&gt;

&lt;p&gt;&lt;code&gt;pip install dbt&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 2: Set up your data warehouse&lt;/strong&gt;&lt;br&gt;
You will need to set up a data warehouse, such as Amazon Redshift, Google BigQuery, or Snowflake, to store your data. If you do not have a data warehouse set up yet, you will need to create one and upload your data onto it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 3: Configure your data warehouse connection&lt;/strong&gt;&lt;br&gt;
In order to connect DBT to your data warehouse, you will need to create a profile in your dbt_project.yml file. This profile should include the connection details for your data warehouse, such as the host, database name, username, and password.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 4: Create a DBT project&lt;/strong&gt;&lt;br&gt;
You can create a new DBT project by running the following command in your terminal: &lt;br&gt;
&lt;code&gt;dbt init&lt;/code&gt;&lt;br&gt;
This will create a new directory with a basic structure for a DBT project, including a &lt;code&gt;dbt_project.yml&lt;/code&gt; file, which you can use to configure your project.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 5: Define your models&lt;/strong&gt;&lt;br&gt;
DBT models are defined using SQL and you can use them to perform data transformations and modeling tasks. You can define your models in .sql files in the model’s directory of your DBT project.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 6: Run and test DBT&lt;/strong&gt;&lt;br&gt;
Once you have defined your models, you can run DBT by running the following command in your terminal:&lt;br&gt;
&lt;code&gt;dbt run&lt;/code&gt;&lt;br&gt;
This will compile your models and run any tests you have defined. If there are no errors, your models will be deployed to your data warehouse.&lt;/p&gt;

&lt;p&gt;Over time, you may need to make changes to your models or add new models. You can use DBT to manage these changes, track the state of your data models, and ensure that your data remains accurate and up-to-date.&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Some hand-on commands you should know for building your model
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;- Create a new branch&lt;/strong&gt; &lt;br&gt;
&lt;code&gt;git checkout -b [new_branch_name]&lt;/code&gt;&lt;br&gt;
In some cases, if your company is using Jira or any product management tool, you can set up a link to the tickets with your PR by including ticket numbers in the names of your new branches. This setup would enable tickets to immediately link with the PR and update other stakeholders on the progress of your changes.&lt;br&gt;
For example: I have ticket &lt;code&gt;spend-123&lt;/code&gt; to solve the bug of transaction data. I would create a new branch by command: &lt;br&gt;
&lt;code&gt;git checkout -b spend-123_transaction_debug&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Switch and update the branch&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
You can switch your branch to master in order to update it along with the latest changes in your pipeline &lt;br&gt;
&lt;code&gt;git checkout master&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcb05kyjj5244hp2mbsgm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcb05kyjj5244hp2mbsgm.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;git pull&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhgbzc5ouh9m3l1debo0o.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhgbzc5ouh9m3l1debo0o.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;To switch back to your branch run&lt;br&gt;
&lt;code&gt;git checkout [branch_name]&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Running your model&lt;/strong&gt;&lt;br&gt;
Run your built model &lt;br&gt;
&lt;code&gt;dbt run -m [model_name]&lt;/code&gt;&lt;br&gt;
OR target any dataset. &lt;br&gt;
For example, to run a model in preprod only, execute the following command:&lt;br&gt;
&lt;code&gt;dbt run -m [model_name] -t preprod&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- To test your built model, execute&lt;/strong&gt;&lt;br&gt;
&lt;code&gt;dbt test --models snapshot_core_transactions&lt;/code&gt; &lt;/p&gt;

&lt;p&gt;*&lt;em&gt;- Reset your branch with this command *&lt;/em&gt;&lt;br&gt;
&lt;code&gt;git reset --hard&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;*&lt;em&gt;- SQL fluff fix *&lt;/em&gt;&lt;br&gt;
SQL fluff refers to redundant or unnecessary code in SQL statements. It can come in various forms, such as unnecessary parentheses, redundant clauses, or excessive spaces. The presence of SQL fluff can make the code more difficult to read, maintain, and optimize, leading to slower performance and increased complexity.&lt;br&gt;
When mondeling data with DBT, it is important to write clean, efficient, and well-structured SQL code to make sure that your data models perform well and are easy to maintain. By reducing or removing SQL fluff you can simplify your code and make it easier to understand and maintain.&lt;br&gt;
In some cases, we can use this command to do a quick fix by implementing a specific rule specified in your run: &lt;br&gt;
&lt;code&gt;sqlfluff fix models/[folder_name]/model_name.sql –rules [rule_name]&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;*&lt;em&gt;- Commit and push the request *&lt;/em&gt;&lt;br&gt;
&lt;code&gt;git commit -m ‘message of your commit’&lt;/code&gt;&lt;br&gt;
&lt;code&gt;git push&lt;/code&gt; &lt;br&gt;
Copy your link URL after DBT creates your pull request and paste it. This would let your link to be submitted to your PR and fill more detailed information and requirements of your PR. &lt;/p&gt;

&lt;p&gt;In short, these are some basic tips to set up and run your model with DBT tool. &lt;/p&gt;

</description>
      <category>dbt</category>
      <category>datascience</category>
      <category>database</category>
      <category>analytics</category>
    </item>
    <item>
      <title>Start With Product Management — Top Key Metrics You Should Know</title>
      <dc:creator>Paige Tran</dc:creator>
      <pubDate>Mon, 03 Oct 2022 12:42:56 +0000</pubDate>
      <link>https://dev.to/bttminhphuc/start-with-product-management-top-key-metrics-you-should-know-2eh</link>
      <guid>https://dev.to/bttminhphuc/start-with-product-management-top-key-metrics-you-should-know-2eh</guid>
      <description>&lt;p&gt;I know many people switch from different backgrounds to Product Management. It is no doubt that Product management has its charm and that you can learn a lot from engineering, design, and branding to analytics. However, it is indeed still very ambiguous how to start.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa20h7ak2w75esfu2l233.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa20h7ak2w75esfu2l233.png" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
When I started to switch to Product, I took the first step — I believe it would be a fundamental step for anyone who wants to switch&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Shifting from depth-to-breadth&lt;br&gt;
In my case, I had years of working on Fraud and Payment analytics particularly. Hence, I start to learn about products by looking widely into product analytics.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Key metrics are the most simple and straightforward way to check the product's health, to know if you are building an impactful product. This article would benefit anyone who works in the product management space!&lt;/p&gt;

&lt;p&gt;Let’s cut the chase, there are the top 3 Product metrics categories I love the most:&lt;br&gt;
**&lt;/p&gt;
&lt;h2&gt;
  
  
  1. Product Awareness
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
This is the most important stage in your product funnel.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you are not being aware that the product is existing, you would never use the product.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fttxu1am2chfuyar8ekc9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fttxu1am2chfuyar8ekc9.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So how to measure this? The fundamental formula is equal to the number of customers who know about your product per total population you advertised or total customers you have. We commonly see it in the campaign reports. However, I personally think it is very important for anyone who builds the product to know this. There is some common metric to use in this case&lt;/p&gt;

&lt;p&gt;Active User (DAU, WAU, MAU): Defined as the number of users who use the product in the specific given time&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;DAU: Daily Active User&lt;/li&gt;
&lt;li&gt;WAU: Weekly Active User&lt;/li&gt;
&lt;li&gt;MAU: Monthly Active User&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Also, in some cases, you can expand the concept of &lt;strong&gt;Active users&lt;/strong&gt; in &lt;strong&gt;calendar time&lt;/strong&gt; or &lt;strong&gt;rolling time&lt;/strong&gt;. &lt;/p&gt;

&lt;p&gt;For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;WAU Calendar&lt;/strong&gt;: Number of Users who uses the product in a specific week from Monday to Sunday&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;WAU Rolling&lt;/strong&gt;: Number of Users who use the product within the last 7 days (No matter which current day is this)&lt;br&gt;
**&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h2&gt;
  
  
  How active are user metrics used to measure awareness?
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
It would narrow down the scope of the population who possibly knows about the change in your product. This metric is also used for the marketing campaign to advertise to the top users about the change since they have better engagement with your product.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;For example: If you want to release a button in your app. To know how many people are aware of your button existing, they would need to go to the app first. Active user metrics are the direct metric to measure this.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;**&lt;/p&gt;
&lt;h2&gt;
  
  
  2. Product Adoption
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
Product adoption, or user adoption, is the moment when users start to &lt;strong&gt;use your product or features&lt;/strong&gt;. At a basic level, adoption can be defined by the percentage of users who take the action on your product or feature for the first time&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwnbnpm2cenyt30pvr3iq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwnbnpm2cenyt30pvr3iq.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;For example: Follow up with your new button in the app from the previous example, now you would like to know how many users tap into the button the first time.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Some of the key metrics in this category would help:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Product Funnel Conversion Rate&lt;/strong&gt;: The ratio in each step is the straightforward way to measure the feature/product adoption. Tips are simple: Understand your funnel and when a user would use CTA (Call to Action)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi3tf6ha2aqt4djotmiwv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi3tf6ha2aqt4djotmiwv.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Speed of Adoption:&lt;/strong&gt; Defined as how long the user takes from the moment, they are aware of the product to the moment they act. This metric is very helpful to identify:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;(1) The product problem: If your product has a technical challenge that prevents the customer from acting&lt;/p&gt;

&lt;p&gt;(2) Product market fit: If the product does not bring benefits or is not helpful to the customer in this market&lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;
&lt;h2&gt;
  
  
  3. Product Retention &amp;amp; Engagement
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
After the customers used the feature/product, the important and sustainable business lies in how well we can retain customers over time.&lt;/p&gt;

&lt;p&gt;There are &lt;strong&gt;2 key metrics&lt;/strong&gt; in this category:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Retention:&lt;/strong&gt; this metric is defined as the percentage of customers from the moment they used the product (adoption) and still use the product over time.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdeax1p9dkznzah5tlwys.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdeax1p9dkznzah5tlwys.png" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
Looking at the example, we see after the first &lt;strong&gt;CTA (Call to action)&lt;/strong&gt;, the retention rate of January is pretty bad, less than 35% actually continue to use the product feature.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Churn:&lt;/strong&gt; One of my favourite metrics. This is a commonly used metric for measuring product engagement. It is defined as a percentage of customers that stopped using the product during a certain period.
In most of the companies I worked at, this is a very important metric because we do not want to lose the customer. Churn metric usually raises multiple high concerns about the business.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;For example: If you see a company with 80% growth rate from month 1 to month 2, it sounds like an amazing result. But if the churn rate is 90% of the month after, then the question is:&lt;br&gt;
&lt;/p&gt;


&lt;/blockquote&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt; “Is the company really growing?” The churn is fast as it grows.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;There are tons of other metrics we should look into when building a product feature. It is definitely case by case. I hope these top product metric categories would help you approach the problem faster as a reference&lt;/p&gt;

&lt;p&gt;I am always up for an open discussion if you have different ways and I love to learn from others too!&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Building your data portfolio is like making a dinner</title>
      <dc:creator>Paige Tran</dc:creator>
      <pubDate>Wed, 07 Sep 2022 13:26:06 +0000</pubDate>
      <link>https://dev.to/bttminhphuc/building-your-data-portfolio-is-like-making-a-dinner-758</link>
      <guid>https://dev.to/bttminhphuc/building-your-data-portfolio-is-like-making-a-dinner-758</guid>
      <description>&lt;p&gt;I wrote an original article: &lt;a href="https://medium.com/@bttminhphuc/building-your-data-portfolio-is-like-making-a-dinner-119f2de11178" rel="noopener noreferrer"&gt;Building your data portfolio is like making a dinner&lt;/a&gt;&lt;br&gt;
Some people asked me to share more in detail about how to build data projects or data portfolios productively after my previous blog about 4 Self-learning steps to get started in Data Analytics&lt;/p&gt;

&lt;p&gt;Hence, I separate this topic and write it in more detail. This article is focusing on some basic workflow that I believe would be beneficial for you to build your own portfolio. These steps are the daily process that I am applying for most of my task jobs. I will take one of my best analogies to describe it:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“How do you make a great dish for dinner?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Step 1: Collecting Data&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fawqfj0ttfrw1es5gnp25.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fawqfj0ttfrw1es5gnp25.png" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
This is the stage when you go to the supermarket and shop. Let’s imagine!&lt;/p&gt;

&lt;p&gt;Data is your raw ingredients. Do you have fresh or spoiled meat? Do you have all or missing ingredients? All the quality items that you shopped in a supermarket would impact your dish. Sometimes, the dish is not perfect because perhaps you are missing some ingredients. Sometimes, you can also borrow or combine ingredients from the different supermarkets just like you can improvise if you possibly utilize data from other sources.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;How to get the data for your practice?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;As shared in the &lt;a href="https://medium.com/@bttminhphuc/4-self-learning-steps-to-get-started-in-data-analytics-c38f6a907516" rel="noopener noreferrer"&gt;previous blog&lt;/a&gt;, you can visit some open source communities such as &lt;a href="https://www.kaggle.com/#" rel="noopener noreferrer"&gt;Kaggle&lt;/a&gt;, or &lt;a href="https://github.com/" rel="noopener noreferrer"&gt;Github&lt;/a&gt;. Some of the good references to collect the data&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://towardsdatascience.com/how-to-collect-data-for-your-analysis-a8bc58043e64" rel="noopener noreferrer"&gt;How to collect data for your analysis&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://towardsdatascience.com/6-web-scraping-tools-that-make-collecting-data-a-breeze-457c44e4411d" rel="noopener noreferrer"&gt;- Tools That Make Collecting Data A Breeze (For higher advanced levels so you can practice Web Scraping and get more data for your project)&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Step 2: EDA (Exploratory Data Analysis) and Cleaning Data&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi0ehkjd2vb5tc71jegns.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi0ehkjd2vb5tc71jegns.png" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
Whenever we talk about EDA or cleaning the data, food preparation is always one of the best examples of how we should do it. Firstly, we need a lot of patience and love for this stage! This stage is when you start cleaning your ingredient and preparing well (includes cutting, chopping, etc), understand better its quality and which part we should take, and which part we can improvise to make it better. In comparison, some of the great values in this stage for data enthusiasts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Understand better the range of data and be able to detect outliers and anomalies&lt;/li&gt;
&lt;li&gt;Identify the missing values&lt;/li&gt;
&lt;li&gt;Decode and conquer imbalance classes&lt;/li&gt;
&lt;li&gt;Re-format variable names and types
And so on!&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;To understand better EDA and cleaning data, I would recommend reading more about:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://towardsdatascience.com/an-extensive-guide-to-exploratory-data-analysis-ddd99a03199e" rel="noopener noreferrer"&gt;Step by Step Guide to Exploratory Data Analysis&lt;/a&gt;&lt;br&gt;
&lt;a href="https://towardsdatascience.com/the-ultimate-guide-to-data-cleaning-3969843991d4" rel="noopener noreferrer"&gt;The Ultimate Guide to Data Cleaning&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Step 3: Project techniques&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;This is the “cooking” stage when you select the right method to cook. Is this a stew dish or grilled dish? The decision is based mainly on the purpose of the dish (when you shopped, you target to make dish X). However, sometimes you could be more flexible based on the ingredient you got. For example, when half of the veggie is spoiled, can borrow some from a neighbour or combine it with some other veggie you have in the fridge. In comparison, sometimes to back up your analysis with a lack of data situation, you can manipulate the existing data or other market research data.&lt;/p&gt;

&lt;p&gt;Some examples of your technique’s selection&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Business analysis&lt;/strong&gt;: Clustering. Even if you only use Excel, SQL, or PowerBI, the most important path is you are clear on the methodology you applying.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Recommendation&lt;/strong&gt;: Collaborative Filtering&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Predictions&lt;/strong&gt;: Classification/Regression model&lt;br&gt;
And so on! Some examples of good technique projects&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.kaggle.com/code/pierpaolo28/stock-market-analysis-and-time-series-prediction/notebook" rel="noopener noreferrer"&gt;Stock Market Analysis and Time Series Prediction&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.kaggle.com/code/laowingkin/netflix-movie-recommendation" rel="noopener noreferrer"&gt;Netflix Movie Recommendation&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Step 4: Insights and recommendations&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;I love this the most:** “How do you serve your food?”**&lt;/p&gt;

&lt;p&gt;Many people underestimate this stage because of the quality of the cooking above. However, I believe that if you serve Phở by putting noodles, soup, and meat separately in different bowls, the dish is not the same anymore. Sometimes, it even confuses people regarding how to eat.&lt;/p&gt;

&lt;p&gt;In comparison, when you present your results, the key question is: &lt;strong&gt;“SO WHAT?”&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn2jhdt4xz5yxyao884pv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn2jhdt4xz5yxyao884pv.png" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
One of the greatest pieces of advice from my former boss is when you present your analysis:&lt;br&gt;
**&lt;br&gt;
“Focus on the storytelling and data insights, rather than the methods”**&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;No method is right or wrong to start learning something. The tips are from my experience. We all have a very different approach. Find the one that suits you.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;I hope you find these tips are helpful&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>database</category>
      <category>datascience</category>
      <category>analytics</category>
    </item>
    <item>
      <title>4 Self-learning steps to get started in Data Analytics</title>
      <dc:creator>Paige Tran</dc:creator>
      <pubDate>Mon, 15 Aug 2022 22:39:00 +0000</pubDate>
      <link>https://dev.to/bttminhphuc/4-self-learning-steps-to-get-started-in-data-analytics-45ne</link>
      <guid>https://dev.to/bttminhphuc/4-self-learning-steps-to-get-started-in-data-analytics-45ne</guid>
      <description>&lt;p&gt;A quick intro on my background, I do not study data science or computer science. Hence, I understand how hard it is for some people who want to start in this field (to become a data analyst, business analyst, data engineer, data scientist, etc) but wondering how to start.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;I would share my simple formula (of course my journey had a lot more failures, and lost feelings and certainly took more time) for how should we start and how can we nurture our skillset in Data Analytics. And most importantly, finding a job that you love. By the way, I love my job 😊&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Before jumping on the long journey, these steps would take months, or years (depending on your commitment level of yourself). Thus, ensure you have the strong motivation of starting something and pursuing it.&lt;/p&gt;

&lt;p&gt;Let’s cut the chase and get into my 4 favorite self-learning steps&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Learning Data Analytics Fundamentals&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Many people ask me the first question when they want to switch their career to data analytics “what tool/certificates I should start learning?”. The answer would always be: You should start with fundamentals.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--hRyi0Xlp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/yda40va2g85mlaxb26pn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--hRyi0Xlp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/yda40va2g85mlaxb26pn.png" alt="Image description" width="880" height="444"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;What are Data Analytics Fundamentals? Basic Statistic. This would enable you the ability to ask the right question and find the right methodology to answer this. In most case, an analyst always need to break down to smaller questions that are easier to answer&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;For example: How should we increase the price of the item without losing our growth? (Stop new customers coming and let our existing customers leave us and never come back). To answer this complex problem, instead of bumping into the toolkit and how to code, a simple approach can be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What drives the price? Cost, revenue. Do we have data on it?&lt;/li&gt;
&lt;li&gt;What drives the cost? What drives the revenue? Do we have data on it?&lt;/li&gt;
&lt;li&gt;What is the optimal point of price that maximise the revenue and minimise the churn? What methodology should we use to predict the impact of price?&lt;/li&gt;
&lt;/ul&gt;
&lt;/blockquote&gt;

&lt;p&gt;After breaking down into the smaller questions. How do we explore the data we have?&lt;/p&gt;

&lt;p&gt;This is when most statistic courses would enable you to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Define and recognize key descriptive statistics&lt;/li&gt;
&lt;li&gt;Describe and distinguish between the central limit theorem and the law of large numbers&lt;/li&gt;
&lt;li&gt;Identify strategies for constructing an unbiased sample&lt;/li&gt;
&lt;li&gt;Where to learn statistics? if you do not learn any of this before in university or school, you can refer to some online free course&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://corporatefinanceinstitute.com/course/statistics-fundamentals/?gclid=Cj0KCQjwl92XBhC7ARIsAHLl9alXZk1nYl0hEHXiwIIzBlxW5aWf2TRG5Bty23j8OlTj2xXOiouQ0XYaAiSTEALw_wcB"&gt;Statistic Fundamentals&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.coursera.org/learn/stanford-statistics"&gt;Introduction to statistic&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Start with the right Toolbox&lt;/strong&gt;&lt;br&gt;
Maybe this is the step when you should slow down and make some research on the difference in roles in data analytics. I particularly love this Infographic; this explains to me a lot about data roles and what tools I should learn. And of course, the role in each company is not always the same scope or responsibility. So, keep in mind this is a reference only.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--RQsTTr4E--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/pnmry3iqhj8nc4xknfum.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--RQsTTr4E--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/pnmry3iqhj8nc4xknfum.png" alt="Image description" width="880" height="745"&gt;&lt;/a&gt;&lt;br&gt;
Source: &lt;a href="https://www.datacamp.com/blog/infographic-the-anatomy-of-a-data-team-different-data-roles"&gt;Data camp&lt;/a&gt; (This is also one of my favorite sites to learn tools)&lt;/p&gt;

&lt;p&gt;Plus video (5mins) for your reference: &lt;a href="https://www.youtube.com/watch?v=X3paOmcrTjQ"&gt;Data Science in 5 mins&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Each role would require a deeper understanding and experience of the different toolboxes. Indeed, finding the right tools after collecting data has always resonated intrinsically with my experience working in data analytics.&lt;/p&gt;

&lt;p&gt;If you do not know what role you would follow in the future, then I would recommend starting with the top 3 below:&lt;/p&gt;

&lt;p&gt;*BI Tools:&lt;br&gt;
PowerBI and Looker are my top selection. BI tools bring huge support for delivering data insight to more people. As a data enthusiast, it is important to understand how to use this tool, and the best case to choose the visualization. This would boost your skills to explain the problem and present it to a wider group of stakeholders. There are tons of visualization tools out there (some for your &lt;a href="https://mopinion.com/business-intelligence-bi-tools-overview/"&gt;reference&lt;/a&gt;) that you can start and practice with such as Tableau, SAS, etc. However, I particularly like PowerBI and Looker.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;For PowerBI, you can easily access via Microsoft and many good practices linking with excel. I refer to &lt;a href="https://www.datacamp.com/courses/introduction-to-power-bi"&gt;BI courses&lt;/a&gt; to learn&lt;/li&gt;
&lt;li&gt;For Looker, this is the tool that was acquired by Google in 2019 and become one of my favorites because of its friendly UXI and practical function. Free Looker &lt;a href="https://www.coursera.org/learn/analyzing-and-visualizing-data-in-looker"&gt;course&lt;/a&gt; to kick start.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;*SQL&lt;br&gt;
Structured Query Language. Fundamentally, it’s used to get your question answered with data from a Relational Database (You can easily understand this term after step 1 fundamentals). This language would contribute tremendous value to answering the basic question that the visualization tool may not present or missing. This also can give you a better comprehensive picture of the current data you have. Below are some top courses I did start with to learn SQL. In my opinion, it is good to give you a strong sense of the real practical case (My top selection from a ton of courses out there).&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.datacamp.com/courses/introduction-to-sql"&gt;Introduction of SQL&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.datacamp.com/courses/joining-data-in-sql"&gt;Joining Data in SQL&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.datacamp.com/courses/intermediate-sql"&gt;Intermediate SQL&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;*Python:&lt;br&gt;
You must hear this programming language somewhere as its popularity in real practice. Python is particularly well suited for deploying machine learning at a large scale and very friendly application for any data exploration. This language is basically given you a deeper understanding of your raw data and how to manipulate it, answering the tougher questions and furthermore, automation. Let’s start with some of my top references:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.datacamp.com/courses/intro-to-python-for-data-science"&gt;Introduction to Python&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.datacamp.com/courses/intermediate-python"&gt;Intermediate Python&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.datacamp.com/courses/exploratory-data-analysis-in-python"&gt;Exploratory Data Analysis in Python&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Python Data Science Toolbox (&lt;a href="https://www.datacamp.com/courses/python-data-science-toolbox-part-1"&gt;Part 1&lt;/a&gt;) and (&lt;a href="https://www.datacamp.com/courses/python-data-science-toolbox-part-2"&gt;Part 2&lt;/a&gt;)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;3. Practice and Practice! Sharpen your skills&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Where to practice?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Certainly, there are many websites for you to practice, but I personally love 2 communities where you not only can practice and build your portfolio but most importantly, learn from others.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.kaggle.com/"&gt;Kaggle&lt;/a&gt;&lt;br&gt;
Kaggle is the world’s largest community of data scientists and machine learning enthusiasts. This platform is the fastest way to get started on a new data science project. It also provides a very friendly interface to practice Jupyter notebook with a single click and is easy to build with the backup of the huge repository (free code and data).&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Most importantly, I learned a lot from other’s works. I believe learning from people in the same community is the fastest ways to sharpen your skills.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://github.com/"&gt;Github&lt;/a&gt;&lt;br&gt;
I mainly use Github to read and learn experiences from others. Different than Kaggle, Github is more focused on function. Github is one of the biggest host sites for GIT (versioning control system). Basically, it allows you and others to manage the changes in the code at the same time without conflict. I am sure you will use it at least once in a real job when you have multiple people on the same project. I know many great friends are using it and highly recommend it if you want to start your portfolio.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Build your own portfolio&lt;/strong&gt;&lt;br&gt;
I would highly recommend that you should start building your portfolio as soon as you plan to practice. I write another article about &lt;a href="https://medium.com/@bttminhphuc/building-your-data-portfolio-is-like-making-a-dinner-119f2de11178"&gt;HOW TO BUILD A DATA PORTFOLIO&lt;/a&gt; and it explains the steps below further in detail. It contains a simple workflow that I apply to most of my task job.&lt;/p&gt;

&lt;p&gt;Step 1: Collecting Data&lt;br&gt;
Step 2: EDA (Exploratory Data Analysis) and Cleaning Data&lt;br&gt;
Step 3: Project techniques&lt;br&gt;
Step 4: Insights and recommendations&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;“Focus on the storytelling and data insights, rather than the methods”&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Ix1twSRh--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/6s086i7b2yvqbl5ngl3n.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Ix1twSRh--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/6s086i7b2yvqbl5ngl3n.jpeg" alt="Image description" width="880" height="587"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What’s Next?&lt;/strong&gt;&lt;br&gt;
Learning every day!!! I build for myself the habit of reading the news daily and listening to podcasts before going to sleep. It is not easy to build but once you do, it brings you tremendous value.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--noPwcCJe--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/c21ump5vhzw4wpoxhdve.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--noPwcCJe--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/c21ump5vhzw4wpoxhdve.jpeg" alt="Image description" width="880" height="430"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Where to read?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I would write another one about this. Different topics have different great sources to read, watch or listen to. So wait for it! Meanwhile, you always can start researching by yourself and build your own habit.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;No method is right or wrong to start learning something. The tips are from my experience. We all have a very different approach. Find the one that suits you. I hope these tips would be helpful somehow as references.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;I am always up for an open discussion if you have different ways and I love to learn from others too!&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>datascience</category>
      <category>analytics</category>
      <category>analyst</category>
    </item>
  </channel>
</rss>
