DEV Community

pawan deore
pawan deore

Posted on • Edited on

11

Efficiently Reading Millions of Rows of SQL Data with Python

Working with large datasets in SQL can be challenging, especially when you need to read millions of rows efficiently. Here’s a straightforward approach to handle this using Python, ensuring that your data processing remains performant and manageable.

The Ultimate Python Programmer Practice Test

Use Efficient Database Drivers

Python has several database drivers like psycopg2 for PostgreSQL, mysql-connector-python for MySQL, and sqlite3 for SQLite. Choose the driver that best fits your database.

import mysql.connector

connection = mysql.connector.connect(
    host="your_host",
    user="your_username",
    password="your_password",
    database="your_database"
)
cursor = connection.cursor()
Enter fullscreen mode Exit fullscreen mode

Fetch Data in Chunks

Fetching millions of rows at once can overwhelm your memory. Instead, fetch data in manageable chunks using a loop. This method keeps memory usage low and maintains performance.

chunk_size = 10000
offset = 0

while True:
    query = f"SELECT * FROM your_table LIMIT {chunk_size} OFFSET {offset}"
    cursor.execute(query)
    rows = cursor.fetchall()

    if not rows:
        break

    process_data(rows)
    offset += chunk_size
Enter fullscreen mode Exit fullscreen mode

Process Data Efficiently

Ensure that your data processing within the process_data function is efficient. Avoid unnecessary computations and leverage vectorized operations with libraries like NumPy or Pandas.

import pandas as pd

def process_data(rows):
    df = pd.DataFrame(rows, columns=['col1', 'col2', 'col3'])
    # Perform operations on the DataFrame
    print(df.head())
Enter fullscreen mode Exit fullscreen mode

Utilize Connection Pooling

For repetitive tasks, connection pooling can help manage database connections efficiently. Libraries like SQLAlchemy provide robust pooling solutions.

from sqlalchemy import create_engine

engine = create_engine("mysql+mysqlconnector://user:password@host/dbname")
connection = engine.connect()

chunk_size = 10000
offset = 0

while True:
    query = f"SELECT * FROM your_table LIMIT {chunk_size} OFFSET {offset}"
    result_proxy = connection.execute(query)
    rows = result_proxy.fetchall()

    if not rows:
        break

    process_data(rows)
    offset += chunk_size
Enter fullscreen mode Exit fullscreen mode

By following these steps, you can efficiently read and process millions of rows of SQL data using Python. This approach ensures that your application remains responsive and performant, even when dealing with large datasets.

Hostinger image

Get n8n VPS hosting 3x cheaper than a cloud solution

Get fast, easy, secure n8n VPS hosting from $4.99/mo at Hostinger. Automate any workflow using a pre-installed n8n application and no-code customization.

Start now

Top comments (0)

Billboard image

Create up to 10 Postgres Databases on Neon's free plan.

If you're starting a new project, Neon has got your databases covered. No credit cards. No trials. No getting in your way.

Try Neon for Free →