<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Fatemeh Vahabi</title>
    <description>The latest articles on DEV Community by Fatemeh Vahabi (@vahabifatima).</description>
    <link>https://dev.to/vahabifatima</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1173980%2Fb9d20cf6-d831-4798-9059-34f2fe910cd5.jpg</url>
      <title>DEV Community: Fatemeh Vahabi</title>
      <link>https://dev.to/vahabifatima</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/vahabifatima"/>
    <language>en</language>
    <item>
      <title>Data Engineering for Beginners: A Step-by-Step Guide</title>
      <dc:creator>Fatemeh Vahabi</dc:creator>
      <pubDate>Thu, 02 Nov 2023 20:25:39 +0000</pubDate>
      <link>https://dev.to/vahabifatima/data-engineering-for-beginners-a-step-by-step-guide-4np1</link>
      <guid>https://dev.to/vahabifatima/data-engineering-for-beginners-a-step-by-step-guide-4np1</guid>
      <description>&lt;p&gt;Introduction:&lt;br&gt;
In today's world, data engineering is one of the most important fields. Data engineering is a set of processes, tools and techniques that we use to collect, store, process and analyze data. This is more important in organizations and companies because strategic decisions are made based on accurate and reliable data. In this article I will provide a step-by-step guide to get started and understand the basic concepts of data engineering.&lt;br&gt;
Basic concepts of data engineering&lt;br&gt;
The basic concepts of data engineering include data definition, data sources, data processing and how to use data in decision making. In data engineering, we use the data to get information that can be collected, stored, processed and analyzed. Data sources can range from internal sources of the organization such as databases, file systems and logs to external sources such as sensors, social data and web data. Data processing includes the processes of data cleaning, transformation and analysis which are done to extract useful information and patterns from the data. Finally, data is used as a basis for strategic decision-making in organizations and companies.&lt;br&gt;
Data collection&lt;br&gt;
The data collection part of data engineering is the process in which data is collected from various sources and transferred to a central location. This process includes steps such as extracting, cleaning, combining and storing data. data sources are first identified which may include databases, file systems, sensors, social data and web data. Then, data is extracted from these sources and collected in raw form.&lt;br&gt;
The next step is data cleaning which includes removing duplicate data, filling in blank values, resolving conflicts and converting data to a standard format. This step is important to maintain the accuracy and usability of the data in next steps.&lt;br&gt;
The data is then combined to create a complete and integrated dataset. This process involves merging different data from different sources which may include combinations of rows, columns, tables or irregular data.&lt;br&gt;
Finally the data is moved to a central location, for example a central database or central file system, so that it is organized and accessible. The main goal of this part is to provide an orderly and optimal environment for performing data processing and analysis processes in the next parts. Data collection which is the most basic step in data engineering is necessary for the effective and efficient use of data in strategic and operational decisions of organizations.&lt;br&gt;
data storage&lt;br&gt;
The data warehousing part of data engineering is the process by which data is organized and permanently stored in a central location. This part of the data engineering process makes sure that data is available in a secure and accessible way and that data can be searched, retrieved and updated.&lt;br&gt;
In this section, databases are usually used as the main tool for data storage. Based on the organization's structure databases can be relation-based such as relational databases or non-relation-based such as NoSQL databases.&lt;br&gt;
In database design and selection factors like data volume, access speed, stability, security, analytical and organizational needs are considered. Also, various technologies such as relational databases, columnar databases, document databases and graph databases are also used for data storage.&lt;br&gt;
Data processing&lt;br&gt;
The fourth part of data engineering is the data processing and analysis part where data is analyzed and processed to extract useful information and patterns. This part of the data engineering process enables optimal use of data and helps organizations make better decisions based on data.&lt;br&gt;
Here data is processed with various techniques and algorithms like feature extraction, statistical refinement and analysis, modeling and machine learning, data quality improvement, data mining and artificial intelligence, natural language processing and other related methods.&lt;br&gt;
The main purpose of this section is to extract useful information and hidden patterns in data, predict events and behaviors, identify relationships and meaningfulness between data, improve data quality and provide methods for better decision making.&lt;br&gt;
Using this segment, organizations can use their data to analyze trends, predict performance, improve processes, identify customers and their behavior, increase productivity, reduce risks, and make strategic decisions. This part of data engineering enables organizations to exploit data as a major strategic asset and make decisions based on evidence and more accurate information.&lt;br&gt;
Data maintenance and management&lt;br&gt;
The part of data engineering is the part related to the deployment and implementation of data solutions. In this section the solutions designed in the previous parts of data engineering are transferred to the operational operations of the organization. In this part data solutions are implemented to use the data and information available in the organization. This process includes installation, configuration, testing and commissioning of data solutions.&lt;br&gt;
The main objective of this department is to ensure the successful implementation of data solutions in the organization. At this part the implementation of data solutions is carried out to improve the performance of the organization, make better decisions, increase competitive ability and improve business processes.&lt;br&gt;
To succeed in this sector, change management, training and preparation of employees, communication and coordination between teams, quality control and continuous support of data solutions are very important. Also, maintaining and updating the solutions and adapting them to the needs of the organization over time is of great importance.&lt;br&gt;
By using this segment organizations can implement and operate their data solutions effectively and efficiently in their processes. Deploying data solutions helps organizations use their data as a powerful tool for strategic decision-making and improve performance against competitors and in the marketplace.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>dseafrica</category>
      <category>dataengineering</category>
      <category>database</category>
    </item>
    <item>
      <title>The Complete Guide to Time Series Models</title>
      <dc:creator>Fatemeh Vahabi</dc:creator>
      <pubDate>Thu, 02 Nov 2023 18:13:01 +0000</pubDate>
      <link>https://dev.to/vahabifatima/the-complete-guide-to-time-series-models-397d</link>
      <guid>https://dev.to/vahabifatima/the-complete-guide-to-time-series-models-397d</guid>
      <description>&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;&lt;br&gt;
Time series analysis is a statistical technique which is used for analyzing and predicting the data that evolves over time.it has applications in different fields like economic, finance, forecasting. time series models are mathematical models which capture the patterns and relationships within the time-independent data. In this article we provide a complete guide for understanding and implementing time series models.&lt;br&gt;
&lt;strong&gt;What is time series?&lt;/strong&gt;&lt;br&gt;
Time series is a set of points which are collected at regular intervals over time. we can represent them as a set of observant shown over the time. time series data often shows trends, seasonality, and other patterns that can be analyzed and utilized for forecasting future values. The main goal in using a time series is to predict future values. The first step in time series is to draw a graph of the data. with drawing a graph, we can identify general information, including an upward or downward trend, the presence of a seasonal pattern, a periodic trend, and the presence of outliers in the data. After plotting the data, the data must be stationary in order to have a proper prediction. Data can be made static by using differentiation or decomposition into its constituent components. After stabilizing the data, the order of the moving average and the order of autoregression of the model can be identified using the graph. It is necessary to examine the obtained parameters for significance using T-test. If it is significant and there is no dependence in the remainder, a suitable prediction can be made with the help of past data, also the predicted values can be evaluated using the absolute average percentage of error.&lt;br&gt;
&lt;strong&gt;The component of time series models&lt;/strong&gt;&lt;br&gt;
Time series models are consist of 3 part:&lt;br&gt;
• Trends: long term pattern or movement in data.&lt;br&gt;
• Seasonality: predictable or repetitive patterns that happen in a specific time period.&lt;br&gt;
• Residuals: The random fluctuations or noise in the data that we cannot explain them by the trend or seasonality.&lt;br&gt;
&lt;strong&gt;Different kinds of time series models&lt;/strong&gt;&lt;br&gt;
• Autoregressive Integrated Moving Average (ARIMA)&lt;br&gt;
• Exponential Smoothing (ES)&lt;br&gt;
• Seasonal Autoregressive Integrated Moving Average (SARIMA)&lt;br&gt;
• Vector Autoregression (VAR)&lt;br&gt;
&lt;strong&gt;Analyzing and choosing model&lt;/strong&gt;&lt;br&gt;
For choosing the best time series model for a specific dataset, we can use different evaluating techniques. These include analyzing residual plots, calculating evaluation metrics like mean squared error (MSE) or Akaike Information Criterion (AIC), and performing cross-validation.&lt;br&gt;
&lt;strong&gt;Model Training and Forecasting&lt;/strong&gt;&lt;br&gt;
When we choosed the appropriate time series model we have to train the model with historical data. This happens with estimating the model parameters based on the data. we can use the model to predict future values by extrapolating the patterns captured in the training phase.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>dseafrica</category>
      <category>timeseriesmodel</category>
      <category>completeguide</category>
    </item>
    <item>
      <title>Data Visualization</title>
      <dc:creator>Fatemeh Vahabi</dc:creator>
      <pubDate>Fri, 13 Oct 2023 20:13:01 +0000</pubDate>
      <link>https://dev.to/vahabifatima/data-visualization-3ac8</link>
      <guid>https://dev.to/vahabifatima/data-visualization-3ac8</guid>
      <description>&lt;p&gt;&lt;strong&gt;What is data visualization?&lt;/strong&gt;&lt;br&gt;
Data visualization is transferring and changing information into some chart and visual stuff that we use it to make recognizing patterns in big datasets easier.&lt;br&gt;
Visualizing data is one of the important part in data science which shows that data should be visualized after processing and modeling to help us make clear results. Also, data visualization is a part of the data architecture system, which maps the flow of data and provides a plan for data management, while documenting the assets of an organization.&lt;br&gt;
Data Visualization has a very important role at analyzing data. For example, when a data scientist is writing advanced predictive analytics algorithms, visualizing the outputs is important to monitor the results and ensure that the models are working correctly. Because It is more simple to get result with visualization of the complicated algorithms than numerical outputs.&lt;br&gt;
Totally, Data Visualization is a form of relation which show density and complexity of data in plots. The Pictures give us the ability to compare data and use them for analyzing the process in a more simple way.&lt;br&gt;
&lt;strong&gt;We know that data visualization is important but why?&lt;/strong&gt;&lt;br&gt;
Data visualization, using visual data, provides a quick and effective way to communicate information.  It helps businesses to recognize which factors affect on the customer’s behavior and which region needs to be proved. Therefor, not only you can make the data more useful for beneficiaries, but also predict the sales rate with data comprehension and visualization.&lt;br&gt;
&lt;strong&gt;What are the features of data visualization?&lt;/strong&gt;&lt;br&gt;
It is accurate, useful, efficient and scalable.&lt;br&gt;
&lt;strong&gt;What are the types of analysis for data visualization ?&lt;/strong&gt;&lt;br&gt;
There are 3 types of analysis for data visualization:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Univariate analysis: Here we use a special feature for analyze all the dimension and features of data. One of the best and most efficient single feature plots for input the information about data distribution is distribution piece. When we want to analyze the effect on the output variable according to the input, we should use the distribution chart.&lt;/li&gt;
&lt;li&gt;Bivariate analysis: comparing data between two features means we have a bivariate analysis.&lt;/li&gt;
&lt;li&gt;Multivariate analysis: here we compare more than two variables.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;What are the different data visualization models?&lt;/strong&gt;&lt;br&gt;
At first the only way to visualize data was to use a Microsoft Excel to turn data into a table, bar chart or any other chart. we can use this method for out visualization but there are better ways for this work:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Infographics&lt;/li&gt;
&lt;li&gt;Bullet charts&lt;/li&gt;
&lt;li&gt;Thermal maps&lt;/li&gt;
&lt;li&gt;Time series graphs&lt;/li&gt;
&lt;li&gt;Line charts&lt;/li&gt;
&lt;li&gt;Tree diagrams&lt;/li&gt;
&lt;li&gt;Area charts&lt;/li&gt;
&lt;li&gt;Bubbly clouds&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>datascience</category>
      <category>datavisualization</category>
      <category>dseafrica</category>
      <category>datamining</category>
    </item>
    <item>
      <title>Data Science RoadMap</title>
      <dc:creator>Fatemeh Vahabi</dc:creator>
      <pubDate>Wed, 04 Oct 2023 09:29:53 +0000</pubDate>
      <link>https://dev.to/vahabifatima/data-science-roadmap-1pl</link>
      <guid>https://dev.to/vahabifatima/data-science-roadmap-1pl</guid>
      <description>&lt;p&gt;Data science includes tools, algorithm and Machine Learning different rules in order to find the hidden pattern in input data. But what is the difference between the work that statistic scientist and data analyst do during recent years?&lt;br&gt;
&lt;strong&gt;What is Data science ?&lt;/strong&gt;&lt;br&gt;
Data science is a very important major in the IT technology that gives us the ability of using data in an organized way. Learning this useful field also gives us this ability to recognize the hidden pattern in the data and do more detailed analysis.&lt;br&gt;
Data science can help us understanding the world around us and making better decision. With learning and using concepts and techniques in data science, we can dominate a smarter and proved world. Finally with growth of data science, It is important to learn continuously and With the advancement of technology and new methods, we should always seek to improve and expand our knowledge and skills in data science.&lt;br&gt;
Data science use Predictive Analytics and Prescriptive Analytics and Machine Learning models for predicting and making decisions. But what each idioms means? &lt;br&gt;
Predictive Analytics helps us to become able to predict the possibility of a special event. For example if you have a company which lends money to its customers, It is important to know that they will repay this loan. For this purpose, you can build a model that can perform predictive analysis on customers' payment history and predict whether they will repay on time or not.&lt;br&gt;
Prescriptive Analytics is a relatively new field that focuses on providing data-driven recommendations. In other words, in addition to predicting probabilities, prescriptive analytics also suggest a range of related actions and outcomes. For example, data collected by vehicles and algorithms can be used to train self-driving cars and make them smarter.&lt;br&gt;
Supervised machine learning can be used to predict future events. For example, machine learning can use a company's transaction data to predict future financial trends or train a model to detect fraud based on fake purchase records.&lt;br&gt;
&lt;strong&gt;What is the roadmap of Data science?&lt;/strong&gt;&lt;br&gt;
In order to learn data science, first step is to know the based rules. We should learn the context like data collection, data cleaning, analyze and extract information from the data and also interpret the results.&lt;br&gt;
The most important step in this way, is that we should know the programming languages and tools related to the data science. Programming languages like Python with well-known libraries like NumPy and Pandas can help us in analyzing the data.&lt;br&gt;
Learning Statistical concepts is a Basic component of data science. We need to be familiar with mean, variance, distributions, correlation coefficients and statistical hypotheses in order to analyze data properly. Also, to evaluate data science models that use machine algorithms, we need to know metrics such as precision, accuracy, and performance characteristic curve.&lt;br&gt;
Next, we have to learn machine algorithms. Algorithms such as linear regression, decision tree, support vector machine (SVM) and neural networks are powerful tools that can help us predict and analyze data.&lt;br&gt;
Data visualization skills are also important skills in data science. We should be able to visualize data in a way that is understandable and interpretable for others through the use of charts, maps and dashboards.&lt;br&gt;
Also, when we want to use data science in a practical way, we need to get to know the related challenges. These challenges may include difficulties in data collection and preprocessing, big data processing, data management, data privacy and security, and correct interpretation of results.&lt;br&gt;
We should know that working with data in SQL is a part of this job. So It is obvious that learning SQL is important.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>roadmap</category>
      <category>machinelearning</category>
      <category>dseafrica</category>
    </item>
  </channel>
</rss>
