DEV Community

Cover image for Mastering Memory Optimization for Pandas DataFrames
thinhda
thinhda

Posted on

Mastering Memory Optimization for Pandas DataFrames

This article aims to guide data scientists and analysts through the essential techniques of memory optimization when
working with Pandas DataFrames. It begins with an introduction to the importance of memory management and common issues
encountered with large datasets. The article then explains how to understand DataFrame memory usage using Pandas'
built-in functions. It delves into optimizing data types by downcasting numeric types and converting object types to
more efficient ones. Efficient data loading techniques are discussed, including selective column loading and using
chunksize. The article also covers the use of sparse data structures to save memory in datasets with many missing or
zero values. Monitoring and profiling tools like memory_profiler and Heapy are introduced to help keep track of memory
consumption. Real-world case studies demonstrate the practical application of these techniques, and the article
concludes with a summary of best practices for memory optimization in Pandas.

Introduction to Memory Optimization in Pandas

When dealing with large datasets, memory optimization becomes a critical factor for ensuring smooth and efficient data
processing. Pandas, a powerful data manipulation library in Python, is widely used by data scientists and analysts to
handle and analyze data. However, as the size of the data grows, so does the memory consumption, leading to potential
performance bottlenecks and even system crashes.

introduce

Why Memory Optimization Matters

Memory optimization is not just about reducing the memory footprint of your DataFrames; it is about enhancing the
overall performance of your data processing tasks. Inefficient memory usage can lead to several issues, including:

  1. Slow Processing: Large datasets can slow down data manipulation and analysis tasks, making operations like filtering, aggregation, and joining extremely sluggish.
  2. System Crashes: Excessive memory usage can exhaust the available RAM, causing your system to crash or become unresponsive.
  3. Scalability Issues: Inefficient memory usage hinders the ability to scale your data processing tasks, limiting the size of the datasets you can work with.
  4. Increased Costs: In cloud environments, higher memory usage translates to increased costs, as you may need to provision larger instances to handle your data.

Common Memory Issues in Pandas

Here are some common memory-related issues encountered when working with Pandas DataFrames:

  • High Memory Consumption: Large datasets can consume a significant amount of memory, especially when using inefficient data types.
  • Memory Fragmentation: Frequent creation and deletion of objects can lead to memory fragmentation, where memory is wasted due to small, unusable gaps.
  • Unused Objects: Objects that are no longer needed but are still occupying memory can lead to memory bloat.

Impact on Performance

Inefficient memory usage can have a direct impact on the performance of your data processing tasks. For example:

  • Longer Execution Times: Operations on large datasets can take longer to execute if the data is not stored efficiently in memory.
  • Increased I/O Operations: When memory is insufficient, data may need to be swapped to disk, leading to increased I/O operations and slower performance.
  • Higher Latency: Memory-intensive tasks can lead to higher latency in data processing pipelines, affecting downstream tasks and overall system performance.

By understanding and addressing these memory issues, you can optimize your Pandas DataFrames for better performance and
scalability. In the following sections, we will explore various techniques and best practices for memory optimization,
including data type optimization, memory profiling tools, and advanced techniques like chunking and garbage collection.

Stay tuned as we delve deeper into the world of memory optimization in Pandas, equipping you with the knowledge and
tools to handle large datasets efficiently.

Understanding DataFrame Memory Usage

When working with large datasets in Pandas, understanding and optimizing memory usage can significantly enhance
performance and efficiency. Here, we will dive into how Pandas DataFrames consume memory, explaining the memory
footprint of different data types and structures within a DataFrame. This section will also cover how to use Pandas'
built-in functions, such as memory_usage() and info(), to inspect and understand the memory consumption of your
DataFrames.

understanding

Memory Footprint of Data Types

Pandas DataFrames can contain various data types, each with its own memory footprint. Common data types include:

  • int64: 8 bytes per element.
  • float64: 8 bytes per element.
  • object: Variable memory usage, depending on the length of the strings.
  • category: Memory usage depends on the number of unique categories and the length of the categorical codes.

Understanding the memory footprint of each data type is crucial for optimizing memory usage in your DataFrames. For
instance, using int8 instead of int64 for columns with smaller numeric ranges can save significant memory.

Inspecting Memory Usage with Pandas

Pandas provides built-in functions to help you inspect the memory usage of your DataFrames. Two essential functions
are memory_usage() and info().

Using memory_usage()

The memory_usage() function returns the memory usage of each column in bytes. By default, it excludes the memory usage
of the DataFrame's index. However, you can include it by setting the index parameter to True.

import pandas as pd

## Sample DataFrame
data = {
    'A': range(1000),
    'B': [x * 2.5 for x in range(1000)],
    'C': ['foo' for _ in range(1000)]
}
df = pd.DataFrame(data)

## Memory usage of each column
print(df.memory_usage())
## Memory usage including the index
print(df.memory_usage(index=True))
Enter fullscreen mode Exit fullscreen mode
Using info()

The info() function provides a concise summary of the DataFrame, including the memory usage. This function is
particularly useful for getting a quick overview of the DataFrame's structure and memory consumption.

## Summary of the DataFrame, including memory usage
df.info()
Enter fullscreen mode Exit fullscreen mode

Memory Optimization Techniques

To optimize memory usage, you can employ several techniques:

  1. Downcasting Numeric Types: Convert larger numeric types to smaller ones when possible.
   df['A'] = pd.to_numeric(df['A'], downcast='integer')
   df['B'] = pd.to_numeric(df['B'], downcast='float')
Enter fullscreen mode Exit fullscreen mode
  1. Using Categorical Data Types: Convert columns with repeated values to the category data type.
   df['C'] = df['C'].astype('category')
Enter fullscreen mode Exit fullscreen mode
  1. Inspecting and Optimizing Index Memory Usage: Use the memory_usage() function with the index=True parameter to inspect the memory usage of the index. Consider resetting the index if it is not necessary or downcasting it to a smaller data type.
   df.memory_usage(index=True)
Enter fullscreen mode Exit fullscreen mode

Result:

=== Summary of the DataFrame before conversion ===
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000 entries, 0 to 999
Data columns (total 3 columns):
 #   Column  Non-Null Count  Dtype  
---  ------  --------------  -----  
 0   A       1000 non-null   int64  
 1   B       1000 non-null   float64
 2   C       1000 non-null   object 
dtypes: float64(1), int64(1), object(1)
memory usage: 23.6+ KB

=== Summary of the DataFrame after conversion ===
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000 entries, 0 to 999
Data columns (total 3 columns):
 #   Column  Non-Null Count  Dtype   
---  ------  --------------  -----   
 0   A       1000 non-null   int16   
 1   B       1000 non-null   float32 
 2   C       1000 non-null   category
dtypes: category(1), float32(1), int16(1)
memory usage: 7.1 KB
Enter fullscreen mode Exit fullscreen mode

By understanding and leveraging these techniques, you can significantly reduce the memory footprint of your Pandas
DataFrames, leading to more efficient data processing and analysis.

Optimizing Data Types

One of the most effective ways to optimize memory usage in Pandas DataFrames is by changing data types. This section
will explore how to downcast numeric types, convert object types to category types, and use more memory-efficient data
types. Let's dive into practical examples and code snippets to illustrate these optimizations.

optimization

Downcasting Numeric Types

Numeric columns in Pandas DataFrames are often stored with data types that use more memory than necessary. For
example, int64 and float64 are common defaults, but they can be downcasted to int32 and float32 respectively to
save memory.

Example: Downcasting Numeric Columns

Here's a practical example of how to downcast numeric columns in a DataFrame:

import pandas as pd
import numpy as np

## Create a sample DataFrame
df = pd.DataFrame(
    {"A": np.random.randint(0, 100, size=1000000), "B": np.random.rand(1000000)}
)

print("Original Memory Usage:")
print(df.memory_usage(deep=True).sum())

print("=========================================")

## Downcast numeric columns
df["A"] = pd.to_numeric(df["A"], downcast="integer")
df["B"] = pd.to_numeric(df["B"], downcast="float")

print("Downcasted Memory Usage:")
print(df.memory_usage(deep=True).sum())

Enter fullscreen mode Exit fullscreen mode

Output:

Original Data Types:
A      int64
B    float64
dtype: object
Original Memory Usage:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 2 columns):
 #   Column  Non-Null Count    Dtype  
---  ------  --------------    -----  
 0   A       1000000 non-null  int64  
 1   B       1000000 non-null  float64
dtypes: float64(1), int64(1)
memory usage: 15.3 MB

Downcasted Data Types:
A       int8
B    float32
dtype: object
Downcasted Memory Usage:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 2 columns):
 #   Column  Non-Null Count    Dtype  
---  ------  --------------    -----  
 0   A       1000000 non-null  int8   
 1   B       1000000 non-null  float32
dtypes: float32(1), int8(1)
memory usage: 4.8 MB
Enter fullscreen mode Exit fullscreen mode

By running this code, you'll observe a significant reduction in memory usage. For instance, a DataFrame with 1 million
rows and int64 and float64 columns might use around 15.3 MB of memory. After downcasting, it could reduce to
approximately 4.8 MB.

Converting Object Types to Category Types

Object types in Pandas are often used to store string data, but they can be memory-intensive. Converting object types to
category types can lead to substantial memory savings, especially when the number of unique values is relatively small.

Example: Converting Object to Category

Consider a DataFrame with a column containing country names:

## Create a sample DataFrame
import pandas as pd

df = pd.DataFrame(
    {"Country": ["USA", "Canada", "USA", "Mexico", "Canada", "USA"] * 100000}
)

print("Original Data Types:")
print(df.dtypes)
print("Original Memory Usage:")
df.info(memory_usage="deep")

## Convert object type to category
df["Country"] = df["Country"].astype("category")

print("\nConverted Data Types:")
print(df.dtypes)
print("Converted Memory Usage:")
df.info(memory_usage="deep")
Enter fullscreen mode Exit fullscreen mode

Output:

Original Data Types:
Country    object
dtype: object
Original Memory Usage:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 600000 entries, 0 to 599999
Data columns (total 1 columns):
 #   Column   Non-Null Count   Dtype 
---  ------   --------------   ----- 
 0   Country  600000 non-null  object
dtypes: object(1)
memory usage: 35.2 MB

Converted Data Types:
Country    category
dtype: object
Converted Memory Usage:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 600000 entries, 0 to 599999
Data columns (total 1 columns):
 #   Column   Non-Null Count   Dtype   
---  ------   --------------   -----   
 0   Country  600000 non-null  category
dtypes: category(1)
memory usage: 586.4 KB
Enter fullscreen mode Exit fullscreen mode

By converting the Country column to a category type, the memory usage is significantly reduced. This is because
category types internally use integer codes to represent the unique values, which are much more memory-efficient than
storing the strings directly.

Using More Memory-Efficient Data Types

In addition to downcasting and converting object types to categories, you can also specify more memory-efficient data
types when loading data. This is particularly useful for large datasets.

Example: Specifying Data Types When Loading Data

When loading data from a CSV file, you can specify the data types for each column to optimize memory usage:

import numpy as np
import pandas as pd

# create fake data.csv file
fake_df = pd.DataFrame(
    {"A": np.random.randint(0, 100, size=1000000), "B": np.random.rand(1000000)}
)
fake_df.to_csv("data.csv", index=False)
print("Memory Usage Before Loading:")
fake_df.info(memory_usage="deep")

## Specify data types for each column
dtype_spec = {"A": "int32", "B": "float32"}

## Load data with specified data types
df = pd.read_csv("data.csv", dtype=dtype_spec)

print("Data Types After Loading:")
print(df.dtypes)
print("Memory Usage After Loading:")
df.info(memory_usage="deep")
Enter fullscreen mode Exit fullscreen mode

Output:

Memory Usage Before Loading:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 2 columns):
 #   Column  Non-Null Count    Dtype  
---  ------  --------------    -----  
 0   A       1000000 non-null  int64  
 1   B       1000000 non-null  float64
dtypes: float64(1), int64(1)
memory usage: 15.3 MB
Data Types After Loading:
A      int32
B    float32
dtype: object
Memory Usage After Loading:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 2 columns):
 #   Column  Non-Null Count    Dtype  
---  ------  --------------    -----  
 0   A       1000000 non-null  int32  
 1   B       1000000 non-null  float32
dtypes: float32(1), int32(1)
memory usage: 7.6 MB
Enter fullscreen mode Exit fullscreen mode

By manually specifying data types, you can ensure that the DataFrame uses memory-efficient types from the start,
avoiding the need for later conversions.

Efficient Data Loading Techniques

In this part, we will discuss techniques for efficiently loading data into Pandas DataFrames. Topics will include
selective column loading, parsing dates, and using chunksize to handle large files. We will provide best practices and
tips to ensure that your data loading process is both fast and memory-efficient.

data loading

Selective Column Loading

When dealing with large datasets, it's often unnecessary to load all columns into memory. By specifying only the columns
you need, you can significantly reduce memory usage and improve loading times.

Example: Loading Specific Columns

Consider a CSV file with many columns, but you only need a few:

## Specify the columns to load
use_cols = ["A"]

## Load only the specified columns
df = pd.read_csv("data.csv", usecols=use_cols)

print("Loaded Data Types:")
print(df.dtypes)
print("Memory Usage After Loading Specific Columns:")
df.info(memory_usage="deep")

Enter fullscreen mode Exit fullscreen mode

Output:

Memory Usage Before Loading:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 2 columns):
 #   Column  Non-Null Count    Dtype  
---  ------  --------------    -----  
 0   A       1000000 non-null  int64  
 1   B       1000000 non-null  float64
dtypes: float64(1), int64(1)
memory usage: 15.3 MB
Loaded Data Types:
A    int64
dtype: object
Memory Usage After Loading Specific Columns:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 1 columns):
 #   Column  Non-Null Count    Dtype
---  ------  --------------    -----
 0   A       1000000 non-null  int64
dtypes: int64(1)
memory usage: 7.6 MB
Enter fullscreen mode Exit fullscreen mode

By loading only the necessary columns, you reduce the amount of data read into memory, making the process more
efficient.

Parsing Dates

Date columns can consume a lot of memory if not handled properly. Pandas provides the parse_dates parameter to
efficiently load date columns.

Example: Parsing Date Columns

Suppose you have a CSV file with a date column:

## Load data without date parsing
df = pd.read_csv("data.csv")

print("Memory Usage Without Parsing Dates:")
df.info(memory_usage="deep")

## Load data with date parsing
df = pd.read_csv("data.csv", parse_dates=["dates"])

print("Memory Usage With Parsing Dates:")
df.info(memory_usage="deep")
Enter fullscreen mode Exit fullscreen mode

Output:

Memory Usage Without Parsing Dates:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 3 columns):
 #   Column  Non-Null Count    Dtype  
---  ------  --------------    -----  
 0   A       1000000 non-null  int64  
 1   B       1000000 non-null  float64
 2   dates   1000000 non-null  object 
dtypes: float64(1), int64(1), object(1)
memory usage: 87.7 MB
Memory Usage With Parsing Dates:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 1000000 entries, 0 to 999999
Data columns (total 3 columns):
 #   Column  Non-Null Count    Dtype         
---  ------  --------------    -----         
 0   A       1000000 non-null  int64         
 1   B       1000000 non-null  float64       
 2   dates   1000000 non-null  datetime64[ns]
dtypes: datetime64[ns](1), float64(1), int64(1)
memory usage: 22.9 MB
Enter fullscreen mode Exit fullscreen mode

By parsing dates during the loading process, you ensure that the date columns are stored in a memory-efficient format.

Using Chunksize to Handle Large Files

For extremely large files, loading the entire dataset into memory might not be feasible. In such cases, you can use
the chunksize parameter to load the data in smaller chunks.

Example: Loading Data in Chunks

Consider a large CSV file that cannot be loaded into memory all at once:

## Initialize an empty DataFrame
df = pd.DataFrame()

## Load data in chunks
chunksize = 100000
for chunk in pd.read_csv('large_data.csv', chunksize=chunksize):
    # Process each chunk
    df = pd.concat([df, chunk], ignore_index=True)

print("Data Types After Loading in Chunks:")
print(df.dtypes)
print("Memory Usage After Loading in Chunks:")
print(df.memory_usage(deep=True))
Enter fullscreen mode Exit fullscreen mode

By loading data in chunks, you can process large files without running into memory issues.

Best Practices and Tips

  1. Error Handling: Use the error_bad_lines and warn_bad_lines parameters to manage rows with missing or malformed data.
  2. Data Type Specification: Always specify data types for each column to ensure memory efficiency.
  3. Performance Metrics: Use the time module and memory_profiler library to measure performance improvements.
  4. Parallel Processing: Leverage libraries like Dask or Modin for parallelized data loading.
  5. Different Data Sources: Apply similar techniques when loading data from databases, Excel files, or JSON files, adjusting methods and parameters as needed.

By incorporating these techniques, you can optimize the data loading process, making it both fast and memory-efficient.

Using Sparse Data Structures

Sparse data structures can save a significant amount of memory when dealing with datasets that contain a lot of missing
or zero values. This section will cover how to use Pandas' sparse data structures, including SparseDataFrame
and SparseSeries, to optimize memory usage. We will also discuss the trade-offs and scenarios where sparse structures
are most beneficial.

Introduction to Sparse Data Structures

In Pandas, sparse data structures are designed to store data efficiently when a large proportion of the values are zeros
or missing. This can be particularly useful in fields like natural language processing, recommender systems, and
genomics, where sparse data is common.

Creating Sparse Series

A SparseSeries is a one-dimensional array that can hold sparse data. Let's start by creating a SparseSeries from a
regular Pandas Series:

import pandas as pd
import numpy as np

## Create a regular Series with many zeros
data = pd.Series([0, 0, 1, 0, 2, 0, 0, 3, 0] * 100000)
print("Memory Usage of Regular Series:")
data.info(memory_usage="deep")

## Convert to SparseSeries
sparse_data = data.astype(pd.SparseDtype("float", fill_value=0))

print("Memory Usage of Sparse Series:")
sparse_data.info(memory_usage="deep")
Enter fullscreen mode Exit fullscreen mode

Output:


Memory Usage of Regular Series:
<class 'pandas.core.series.Series'>
RangeIndex: 900000 entries, 0 to 899999
Series name: None
Non-Null Count   Dtype
--------------   -----
900000 non-null  int64
dtypes: int64(1)
memory usage: 6.9 MB

Memory Usage of Sparse Series:
<class 'pandas.core.series.Series'>
RangeIndex: 900000 entries, 0 to 899999
Series name: None
Non-Null Count   Dtype             
--------------   -----             
900000 non-null  Sparse[float64, 0]
dtypes: Sparse[float64, 0](1)
memory usage: 3.4 MB
Enter fullscreen mode Exit fullscreen mode

Creating Sparse DataFrames

A SparseDataFrame is a two-dimensional array that can hold sparse data. You can create a SparseDataFrame by
converting a regular DataFrame:

df = pd.DataFrame(
    {
        "A": [0, 0, 1, 0, 2] * 100000,
        "B": [0, 3, 0, 0, 0] * 100000,
        "C": [0, 0, 0, 4, 0] * 100000,
    }
)

print("\nMemory Usage of Regular DataFrame:")
df.info(memory_usage="deep")

## Convert to SparseDataFrame
sparse_df = df.astype(pd.SparseDtype("float", fill_value=0))

print("\nMemory Usage of Sparse DataFrame:")
sparse_df.info(memory_usage="deep")

Enter fullscreen mode Exit fullscreen mode

Output:

Memory Usage of Regular DataFrame:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 500000 entries, 0 to 499999
Data columns (total 3 columns):
 #   Column  Non-Null Count   Dtype
---  ------  --------------   -----
 0   A       500000 non-null  int64
 1   B       500000 non-null  int64
 2   C       500000 non-null  int64
dtypes: int64(3)
memory usage: 11.4 MB

Memory Usage of Sparse DataFrame:
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 500000 entries, 0 to 499999
Data columns (total 3 columns):
 #   Column  Non-Null Count  Dtype             
---  ------  --------------  -----             
 0   A       1 non-null      Sparse[float64, 0]
 1   B       1 non-null      Sparse[float64, 0]
 2   C       1 non-null      Sparse[float64, 0]
dtypes: Sparse[float64, 0](3)
memory usage: 4.6 MB
Enter fullscreen mode Exit fullscreen mode

Trade-offs and Efficiency

While sparse data structures can significantly reduce memory usage, they come with trade-offs. Operations on sparse
structures may be slower due to the overhead of managing the sparse format. It's essential to consider these trade-offs
based on the specific use case.

Example: Memory Savings

To illustrate the memory savings, let's compare the memory usage of a dense DataFrame and its sparse counterpart:

## Create a large DataFrame with many zeros
large_df = pd.DataFrame(np.random.choice([0, 1], size=(10000, 1000), p=[0.95, 0.05]))

## Convert to SparseDataFrame
sparse_large_df = large_df.astype(pd.SparseDtype("float", fill_value=0))

print("Memory Usage of Dense DataFrame:")
print(large_df.memory_usage(deep=True).sum())
print("Memory Usage of Sparse DataFrame:")
print(sparse_large_df.memory_usage(deep=True).sum())
Enter fullscreen mode Exit fullscreen mode

Output:

Memory Usage of Dense DataFrame:
80000132
Memory Usage of Sparse DataFrame:
5999160
Enter fullscreen mode Exit fullscreen mode

Scenarios for Using Sparse Data Structures

Sparse data structures are most beneficial in the following scenarios:

  1. High Proportion of Zeros or Missing Values: When the dataset contains a large number of zeros or NaN values, sparse structures can save significant memory.
  2. Text Data: Representing text data, such as term frequency-inverse document frequency (TF-IDF) matrices, where most entries are zero.
  3. Recommender Systems: Storing user-item interaction data, where only a small fraction of possible interactions are present.
  4. Genomics: Efficiently storing genetic data, where only a small fraction of possible genetic variations are present.

Conversion Between Dense and Sparse Formats

Converting between dense and sparse formats is straightforward in Pandas. Here’s how to do it efficiently:

## Convert dense to sparse
sparse_df = df.astype(pd.SparseDtype("float", fill_value=0))

## Convert sparse to dense
dense_df = sparse_df.sparse.to_dense()

print("Dense DataFrame from Sparse:")
print(dense_df)
Enter fullscreen mode Exit fullscreen mode

By understanding and leveraging sparse data structures, you can optimize memory usage and improve the performance of
your data processing tasks, particularly when dealing with large datasets containing a significant proportion of missing
or zero values.

Memory Profiling and Monitoring

Monitoring memory usage is crucial for identifying bottlenecks and optimizing performance. In this section, we will
explore tools and techniques for profiling and monitoring memory usage in Pandas. We will introduce libraries
like memory_profiler and Heapy, and show how to integrate them into your workflow to keep track of memory
consumption.

profiling

Using memory_profiler

memory_profiler is a Python module for monitoring memory usage of a program. It is particularly useful for identifying
memory leaks and understanding memory consumption patterns.

Installing memory_profiler

You can install memory_profiler using pip:

pip install memory_profiler
Enter fullscreen mode Exit fullscreen mode

Basic Usage

To use memory_profiler, you need to decorate the functions you want to profile with @profile. Here is an example:

from memory_profiler import profile
import pandas as pd


@profile
def create_large_dataframe():
    # Create a large DataFrame with random data
    df = pd.DataFrame({
        'A': range(1000000),
        'B': range(1000000)
    })
    return df


if __name__ == "__main__":
    df = create_large_dataframe()
Enter fullscreen mode Exit fullscreen mode

Run the script with the -m memory_profiler flag to profile memory usage:

python -m memory_profiler your_script.py
Enter fullscreen mode Exit fullscreen mode

Output:


Line #    Mem usage    Increment  Occurrences   Line Contents
=============================================================
     5     70.8 MiB     70.8 MiB           1   @profile
     6                                         def create_large_dataframe():
     7                                             # Create a large DataFrame with random data
     8    101.5 MiB     30.7 MiB           1       df = pd.DataFrame({"A": range(1000000), "B": range(1000000)})
     9    101.5 MiB      0.0 MiB           1       return df

Enter fullscreen mode Exit fullscreen mode

Analyzing the Output

The output will show memory usage before and after each line of the decorated function. This helps in pinpointing the
lines of code responsible for high memory usage.

Using Heapy

Heapy is another tool for memory profiling, which provides detailed insights into memory usage, including identifying
memory leaks.

Installing Heapy

You can install Heapy using pip:

pip install guppy3
Enter fullscreen mode Exit fullscreen mode

Basic Usage

Here is an example of how to use Heapy to profile memory usage:

from guppy import hpy
import pandas as pd


def create_large_dataframe():
    # Create a large DataFrame with random data
    df = pd.DataFrame({
        'A': range(1000000),
        'B': range(1000000)
    })
    return df


if __name__ == "__main__":
    hp = hpy()
    hp.setrelheap()  # Set the reference point for memory usage

    df = create_large_dataframe()

    heap = hp.heap()  # Get the current heap status
    print(heap)
Enter fullscreen mode Exit fullscreen mode

Output:

Partition of a set of 39 objects. Total size = 32004484 bytes.
 Index  Count   %     Size   % Cumulative  % Kind (class / dict of class)
     0      3   8 16000384  50  16000384  50 numpy.ndarray
     1      1   3 16000164  50  32000548 100 pandas.core.frame.DataFrame
     2      7  18      560   0  32001108 100 weakref.ReferenceType
     3      2   5      432   0  32001540 100 set
     4      4  10      376   0  32001916 100 dict (no owner)
     5      1   3      296   0  32002212 100 dict of pandas.core.flags.Flags
     6      1   3      296   0  32002508 100 dict of pandas.core.frame.DataFrame
     7      1   3      296   0  32002804 100 dict of pandas.core.indexes.range.RangeIndex
     8      1   3      280   0  32003084 100 dict of pandas.core.indexes.base.Index
     9      3   8      224   0  32003308 100 list
<12 more rows. Type e.g. '_.more' to view.>
Enter fullscreen mode Exit fullscreen mode

Analyzing the Output

Heapy provides a detailed breakdown of memory usage, including the types and number of objects in memory. This can
help identify memory leaks and understand memory consumption patterns.

Combining memory_profiler and Heapy

For comprehensive memory profiling, you can combine memory_profiler and Heapy to leverage the strengths of both
tools. Here is an example:

from memory_profiler import profile
from guppy import hpy
import pandas as pd


@profile
def create_large_dataframe():
    hp = hpy()
    hp.setrelheap()  # Set the reference point for memory usage

    # Create a large DataFrame with random data
    df = pd.DataFrame({
        'A': range(1000000),
        'B': range(1000000)
    })

    heap = hp.heap()  # Get the current heap status
    print(heap)
    return df


if __name__ == "__main__":
    df = create_large_dataframe()
Enter fullscreen mode Exit fullscreen mode

This approach allows you to monitor memory usage at the function level with memory_profiler and get detailed memory
insights with Heapy.

Best Practices for Memory Profiling

  1. Use Profiling Tools Sparingly: Profiling tools can introduce performance overhead. Use them selectively on parts of the code where you suspect memory issues.
  2. Analyze Garbage Collection: Python's garbage collector can affect memory usage readings. Consider disabling garbage collection temporarily to get more accurate measurements.
  3. Profile in a Controlled Environment: Run memory profiling in a controlled environment to minimize interference from other processes.

By integrating these tools and techniques into your workflow, you can effectively monitor and optimize memory usage in
your Pandas applications, leading to improved performance and stability.

Case Studies and Real-World Examples

To bring all the concepts together, we will present case studies and real-world examples of memory optimization in
Pandas. These examples will demonstrate how the techniques discussed in the blog post can be applied to real datasets to
achieve significant memory savings and performance improvements.

case studies

Case Study 1: Optimizing a Financial Dataset

Initial Data Analysis

Let's start by analyzing the initial data types and memory usage of our financial dataset. This dataset contains stock
prices for various companies over several years.

import pandas as pd

## Load the dataset
df = pd.read_csv('financial_data.csv')

## Display initial memory usage
print(df.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode

Optimization Techniques

Downcasting Numeric Columns

One of the first optimization techniques we can apply is downcasting numeric columns to more memory-efficient types.

## Downcast numeric columns
df['price'] = pd.to_numeric(df['price'], downcast='float')
df['volume'] = pd.to_numeric(df['volume'], downcast='integer')

## Display memory usage after downcasting
print(df.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode
Converting Columns to Categorical Types

Next, we can convert columns with a limited number of unique values to categorical types.

## Convert columns to categorical types
df['stock_symbol'] = df['stock_symbol'].astype('category')
df['sector'] = df['sector'].astype('category')

## Display memory usage after converting to categorical types
print(df.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode
Handling Missing Data

Handling missing data is crucial for accurate memory optimization. We can fill missing values or use appropriate data
types that handle NaNs efficiently.

## Fill missing values with a placeholder
df['price'].fillna(-1, inplace=True)
df['volume'].fillna(0, inplace=True)

## Display memory usage after handling missing data
print(df.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode

Performance Impact

To measure the performance impact of our optimizations, we can use memory_profiler and Heapy as discussed earlier.

from memory_profiler import profile
from guppy import hpy


@profile
def optimize_financial_data():
    hp = hpy()
    hp.setrelheap()  # Set the reference point for memory usage

    # Load and optimize the dataset
    df = pd.read_csv('financial_data.csv')
    df['price'] = pd.to_numeric(df['price'], downcast='float')
    df['volume'] = pd.to_numeric(df['volume'], downcast='integer')
    df['stock_symbol'] = df['stock_symbol'].astype('category')
    df['sector'] = df['sector'].astype('category')
    df['price'].fillna(-1, inplace=True)
    df['volume'].fillna(0, inplace=True)

    heap = hp.heap()  # Get the current heap status
    print(heap)
    return df


if __name__ == "__main__":
    df = optimize_financial_data()
Enter fullscreen mode Exit fullscreen mode

Case Study 2: Optimizing a Customer Reviews Dataset

Initial Data Analysis

Now, let's analyze a dataset containing customer reviews. This dataset includes text reviews, ratings, and user
information.

## Load the dataset
df_reviews = pd.read_csv('customer_reviews.csv')

## Display initial memory usage
print(df_reviews.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode

Optimization Techniques

Text Column Optimization

Text columns can consume a significant amount of memory. We can optimize them by converting to categorical types if
there are repeated phrases or by using specialized libraries like pyarrow.

## Convert text columns to categorical types
df_reviews['review_text'] = df_reviews['review_text'].astype('category')

## Display memory usage after text optimization
print(df_reviews.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode
Downcasting Numeric Columns

Similar to the financial dataset, we can downcast numeric columns for memory efficiency.

## Downcast numeric columns
df_reviews['rating'] = pd.to_numeric(df_reviews['rating'], downcast='integer')

## Display memory usage after downcasting
print(df_reviews.info(memory_usage='deep'))
Enter fullscreen mode Exit fullscreen mode

Performance Impact

Again, we can measure the performance impact using memory_profiler and Heapy.

@profile
def optimize_reviews_data():
    hp = hpy()
    hp.setrelheap()  # Set the reference point for memory usage

    # Load and optimize the dataset
    df_reviews = pd.read_csv('customer_reviews.csv')
    df_reviews['review_text'] = df_reviews['review_text'].astype('category')
    df_reviews['rating'] = pd.to_numeric(df_reviews['rating'], downcast='integer')

    heap = hp.heap()  # Get the current heap status
    print(heap)
    return df_reviews


if __name__ == "__main__":
    df_reviews = optimize_reviews_data()
Enter fullscreen mode Exit fullscreen mode

These case studies illustrate how memory optimization techniques can be applied to real-world datasets, resulting in
significant memory savings and performance improvements. By analyzing the initial data, applying appropriate
optimizations, and measuring the impact, you can effectively manage memory usage in your Pandas applications.

Conclusion and Best Practices

In the final section, we will summarize the key points discussed in the blog post and provide a list of best practices
for memory optimization in Pandas. This will serve as a handy reference for readers to implement memory optimization
techniques in their own projects.

conclusion

Summary of Key Points and Best Practices for Memory Optimization in Pandas

Memory Usage of Indexes

Indexes in Pandas DataFrames can consume significant memory. To optimize, consider using more memory-efficient index
types. For instance, if the index is a range of integers, using pd.RangeIndex can save memory compared to a default
integer index.

Best Practices:

  • Regularly check the memory usage of DataFrame indexes using df.memory_usage(deep=True).
  • Choose index types that are appropriate for the data size and type.

DataFrame Consolidation

DataFrame consolidation refers to the process of combining multiple DataFrames into a single DataFrame to reduce memory
overhead. This can be particularly useful when dealing with fragmented DataFrames.

Best Practices:

  • Use pd.concat to merge DataFrames efficiently.
  • Consolidate DataFrames that share the same structure to minimize memory usage.

Garbage Collection

Python's garbage collection can impact memory usage by automatically freeing up memory that is no longer in use.
However, large objects or circular references can delay garbage collection.

Management Strategies:

  • Use the gc module to manually trigger garbage collection, especially after large data manipulations.
  • For example, import gc; gc.collect() can be used to free up memory immediately.

In-place Operations

In-place operations modify the data directly without creating a copy, thus saving memory. For example,
using df.drop(columns=['col_name'], inplace=True) or df.sort_values(by='col_name', inplace=True) can help avoid
unnecessary copies.

Best Practices:

  • Whenever possible, use in-place operations to reduce memory overhead.
  • Ensure that in-place operations are safe and won't affect the original data integrity.

Efficient Data Loading

Efficiently loading large datasets can significantly reduce memory usage. Use the dtype parameter in read_csv to
specify the data types of columns, thereby reducing the memory footprint.

Best Practices:

  • Load only necessary columns using the usecols parameter.
  • Consider chunking large files with the chunksize parameter to process data in smaller, manageable pieces.

By following these best practices, you can optimize memory usage in Pandas, making your data processing tasks more
efficient and scalable.

View more articles

Top comments (0)