<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Zara Ziad</title>
    <description>The latest articles on DEV Community by Zara Ziad (@zaraziad).</description>
    <link>https://dev.to/zaraziad</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F820403%2F3c783183-8de8-4778-8ec1-7e15a60210a2.jpeg</url>
      <title>DEV Community: Zara Ziad</title>
      <link>https://dev.to/zaraziad</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/zaraziad"/>
    <language>en</language>
    <item>
      <title>Finding and removing duplicates through software tools</title>
      <dc:creator>Zara Ziad</dc:creator>
      <pubDate>Thu, 07 Jul 2022 11:53:46 +0000</pubDate>
      <link>https://dev.to/zaraziad/finding-and-removing-duplicates-through-software-tools-53de</link>
      <guid>https://dev.to/zaraziad/finding-and-removing-duplicates-through-software-tools-53de</guid>
      <description>&lt;p&gt;In Datafold’s recent survey on &lt;a href="https://www.datafold.com/blog/the-state-of-data-quality-in-2021"&gt;The State of Data Quality in 2021&lt;/a&gt;, most respondents said that they do not use any tool or software to clean data, and manually perform data quality verification and fixes. Nowadays, process optimization and operational efficiency are impossible without technology. A team that relies on manual effort to execute complicated tasks, such as bulk data cleansing and deduplication, is bound to make mistakes and produce inconsistent results.  &lt;/p&gt;

&lt;p&gt;In this blog, we will look at what a &lt;a href="https://dataladder.com/data-deduplication-software/"&gt;data deduplication software&lt;/a&gt; is, the most crucial features and functionalities found in such a tool, and how it can help your data quality teams to save time and money. Let's get started. &lt;/p&gt;

&lt;h2&gt;
  
  
  Data deduplication software
&lt;/h2&gt;

&lt;p&gt;Data deduplication software is a tool that efficiently executes the data deduplication process which includes: &lt;br&gt;
Comparing two or more records, &lt;br&gt;
Computing the probability of these records being similar or belonging to the same entity, &lt;br&gt;
Deciding which information to retain and which to override, &lt;br&gt;
Merging records to get a single comprehensive record for an entity, &lt;br&gt;
Deleting and discarding duplicate records. &lt;/p&gt;

&lt;p&gt;The data deduplication process requires advanced knowledge about data and how to deal with it to get optimal results. Otherwise, you may end up losing crucial information. Data deduplication tools come with advanced data profiling, cleaning, and matching algorithms – that are capable of processing millions of records in a matter of minutes. This is where automated tools can perform better and more quickly, accurately, and consistently – as compared to human effort.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites of accurate data deduplication
&lt;/h2&gt;

&lt;p&gt;Since datasets contain a variety of data quality errors, the accuracy of your data deduplication process depends on a number of pre-requirements, such as: &lt;br&gt;
Data residing at disparate sources must be combined or integrated together for the process. &lt;br&gt;
All records must follow a standardized format and pattern. &lt;br&gt;
Data values must be present in the same data type and unit of measurement. &lt;br&gt;
Data match definitions must be configured to define which attributes should be used for matching data. &lt;br&gt;
Rules must be defined to compare duplicate records and make decisions to merge and override data. &lt;br&gt;
There must be a way to export or store duplicate records in case you don’t want to delete them. &lt;/p&gt;

&lt;h2&gt;
  
  
  Features of data deduplication software
&lt;/h2&gt;

&lt;p&gt;If you take the prerequisites into consideration, it becomes clear that a data deduplication tool must be equipped with all these features. Let’s discuss the most crucial features to look for in data deduplication software: &lt;br&gt;
&lt;strong&gt;Data ingestion:&lt;/strong&gt; the ability to import, connect, or integrate data residing across disparate sources, including local files, databases, third-party applications, etc. &lt;br&gt;
&lt;strong&gt;Data profiling:&lt;/strong&gt; the ability to uncover hidden details about your data in terms of blank values, incorrect patterns, formats, and datatypes, to identify possible cleaning opportunities. &lt;br&gt;
&lt;strong&gt;Data cleansing and standardization:&lt;/strong&gt; the ability to eliminate incorrect information and transform data values to follow a standardized view across all sources. &lt;br&gt;
&lt;strong&gt;Data parsing and merging:&lt;/strong&gt; the ability to divide one column into multiple columns, or inversely, combining two or more columns into one. &lt;br&gt;
&lt;strong&gt;Data match configuration:&lt;/strong&gt; the ability to configure match definitions and tune algorithms according to the nature of data to get optimal results. &lt;br&gt;
&lt;strong&gt;Data deduplication:&lt;/strong&gt; the ability to execute different types of matching (exact or fuzzy) and compute the probability score of two or more records belonging to the same entity. &lt;br&gt;
&lt;strong&gt;Data merge and survivorship:&lt;/strong&gt; the ability to merge and override duplicate records and get a single comprehensive view of entities. &lt;br&gt;
Data export or load: the ability to transfer a single source of truth to the required destination, such as a local file, database, or any third-party application. &lt;/p&gt;

&lt;h2&gt;
  
  
  Factors to consider while employing data deduplication software
&lt;/h2&gt;

&lt;p&gt;There are quite a few vendors in the market that offer the features mentioned above in their data deduplication tools. But there are some factors to consider while choosing such a tool: &lt;br&gt;
What does your organization require? &lt;br&gt;
Data quality means something different for every organization. For this reason, instead of buying a tool that you heard works for somewhere else, you need to find out what will possibly work for you. Here, a list of data quality KPIs will help you understand what you are looking to achieve and whether the solution under consideration can help you implement that vision. &lt;br&gt;
How much time and budget are you willing to invest in this tool? &lt;br&gt;
Adapting to technological changes in an organization asks for time and money. You may need to assess what budget you are willing to invest in this tool. Also consider the fact that it might take some time for your team members to learn the new tool and use it efficiently. &lt;br&gt;
What does your data quality team prefer? &lt;br&gt;
This is a key player in your decision to choose a data deduplication tool. &lt;a href="https://dataladder.com/building-a-data-quality-team-roles-and-responsibilities-to-consider/"&gt;Data quality team&lt;/a&gt; members are often present in organizations as data analysts, data stewards, or data managers. These individuals spend most of their day dealing with multiple data applications, sources, and tools. Let them decide which tool helps them get their job done most efficiently.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Data deduplication is the first step in enabling a reliable data culture at a company and creating a single source of truth that is accessible to everyone. When your datasets are free from duplicates, you can get many benefits, such as accurate data analysis, customer personalization, data compliance, brand loyalty, and operational efficiency. Investing in such tools will definitely reduce rework and free up your team members to focus on more important tasks. &lt;/p&gt;

</description>
      <category>data</category>
      <category>deduplication</category>
      <category>dataquality</category>
      <category>mergepurge</category>
    </item>
    <item>
      <title>A codeless solution for cleaning and verifying your address data</title>
      <dc:creator>Zara Ziad</dc:creator>
      <pubDate>Wed, 23 Feb 2022 10:10:15 +0000</pubDate>
      <link>https://dev.to/zaraziad/a-codeless-solution-for-cleaning-and-verifying-your-address-data-jbe</link>
      <guid>https://dev.to/zaraziad/a-codeless-solution-for-cleaning-and-verifying-your-address-data-jbe</guid>
      <description>&lt;p&gt;Today, data has become one of the greatest assets of an organization. Whether you want to design customer journeys or forecast business future, data is the main ingredient that helps to attain successful outcomes. This is why business owners invest in developing custom solutions for keeping their data clean – especially a customer or contacts database.  &lt;/p&gt;

&lt;p&gt;But since multiple employees at a company work with, manipulate, and use the contacts dataset, it is soon filled with inconsistencies and inaccuracies. And then the company’s IT staff is expected to build an in-house solution that magically gets rid of all errors present in the database.&lt;/p&gt;

&lt;h2&gt;
  
  
  Coding every solution from scratch
&lt;/h2&gt;

&lt;p&gt;Although it is possible to write code for cleaning and standardizing datasets, but it is definitely an inefficient solution – considering the number of resources (time, people, and money) required for its implementation. And after factoring in the cost of annual maintenance and upgrades, it is 2-3 times more expensive than adopting existing solutions. &lt;/p&gt;

&lt;p&gt;This reminds me of something one of my coder friends told me recently: At some point in every developer’s life, they realize how unproductive it is to code every solution by hand. Sometimes it is more efficient to adopt existing solutions available in the market – open-source libraries or commercial products – rather than coding solutions from scratch. &lt;/p&gt;

&lt;p&gt;In this blog, I will explain some common terminologies and steps involved in cleaning and validating addresses present in a customer’s database. This will definitely help you to understand what to look for while choosing an existing solution available in the market. Let's get started. &lt;/p&gt;

&lt;h2&gt;
  
  
  Common terminologies involved
&lt;/h2&gt;

&lt;p&gt;Before we get into specifics about the process, there are some common terminologies used in this domain, let’s first go over them and see what they mean. &lt;/p&gt;

&lt;h2&gt;
  
  
  Address standardization
&lt;/h2&gt;

&lt;p&gt;Address standardization (also known as &lt;a href="https://dataladder.com/address-standardization-software/"&gt;address normalization&lt;/a&gt;) means updating the format of an address according to an authoritative standard (such as the USPS addressing standard in US). This process makes sure that the addresses are present in an acceptable format – includes correct spelling, abbreviations, geocodes, as well as is appended with ZIP+4 values. &lt;/p&gt;

&lt;h2&gt;
  
  
  Address verification
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://dataladder.com/address-verification-software/"&gt;Address verification&lt;/a&gt; (also known as address validation) is the process of running the standardized addresses against an authoritative database (such as the USPS in US), and making sure that these addresses are real – meaning, they are mailable and valid locations within the country for mail delivery. &lt;/p&gt;

&lt;h2&gt;
  
  
  Difference between the two
&lt;/h2&gt;

&lt;p&gt;Sometimes both these terms are used interchangeably, but there’s a difference between the two. Addresses should be first standardized to follow an acceptable format. Once standardized, they are now ready to be verified to check if these addresses are real and valid.&lt;/p&gt;

&lt;h2&gt;
  
  
  Process of standardizing and validating addresses
&lt;/h2&gt;

&lt;p&gt;The following steps are involved in this process: &lt;/p&gt;

&lt;h2&gt;
  
  
  1. Profiling addresses
&lt;/h2&gt;

&lt;p&gt;Before any activity can be performed on the address database, it is important to assess its current state. This is where address profiling can be very helpful. It identifies the records that contain incomplete or missing address information, as well as the ones that don’t follow a standardized pattern. &lt;/p&gt;

&lt;p&gt;Address profiling highlights potential cleansing and standardization opportunities present in your dataset. Furthermore, this profile report is usually generated again at the end of the process so that both the initial and ending reports are compared to see if there are still errors present in the dataset.&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Parsing addresses
&lt;/h2&gt;

&lt;p&gt;The address standardization starts by parsing every address into its sub-components. This is important since addresses are mostly stored as a single field in a dataset. And running validation checks on the entire field is not as accurate as running them on its sub-pieces. For this reason, a single address is usually parsed into street number, street name, zip code, postal code, directions, city, state, and county. &lt;/p&gt;

&lt;h2&gt;
  
  
  3. Geocoding
&lt;/h2&gt;

&lt;p&gt;In this step, the latitudinal and longitudinal geocodes are computed for all addresses. In addition to that, depending on the computed geocodes, you can also find out the 5-digit zip codes and 4-digit routes of delivery area. &lt;/p&gt;

&lt;h2&gt;
  
  
  4. Reconstructing addresses
&lt;/h2&gt;

&lt;p&gt;Once all this information is computed and standardized, it is not time to reformat and reconstruct the addresses in the required format. This can be done and then saved in the database, or if needed, it can be computed in real-time whenever and however needed. &lt;/p&gt;

&lt;p&gt;An example of such formatting is the USPS addressing standard that requires the delivery address to cover three lines – the first one contains the recipient’s name, the second one contains the street address, and the third one contains the city, state, and zip code. &lt;/p&gt;

&lt;h2&gt;
  
  
  5. Verifying addresses
&lt;/h2&gt;

&lt;p&gt;When an address has all the necessary components, you can now verify its validity against any authoritative database to find out whether the address is an actual, mailable location. In addition to verification, such databases can also tell the type of address – residential or business – as well as some other secondary details.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;And there you have it, a 5-step, codeless process for cleaning and verifying your address data. Implementing such a solution from scratch can be very challenging and can take years to improve the result accuracy.  &lt;/p&gt;

&lt;p&gt;There are many address verification tools in the industry today, including some that are CASS-certified – a certification title that the USPS assigns to software vendors offering accurate address standardization and verification services.  &lt;/p&gt;

&lt;p&gt;Such tools can definitely improve your team’s operational efficiency and enable them to design exceptional experiences for customers by using correct and accurate location information. &lt;/p&gt;

</description>
      <category>data</category>
      <category>dataquality</category>
      <category>addressverification</category>
    </item>
  </channel>
</rss>
