LinkVault PRO is a production-ready Python tool designed for extracting URLs from TXT, PDF, and HTML files in a folder recursively. It supports file type filtering, duplicate removal, and exports results in TXT, JSON, or PDF formats with a glowing progress interface.
Get it here: https://gum.new/gum/cmkf1hjun000404l2fqerdq39
Features
โ
Recursive folder scanning
โ
File type filtering: .txt, .pdf, .html/.htm
โ
Duplicate link filtering
โ
Glowing progress bar with ETA and speed
โ
Copy links to clipboard
โ
Open links by double-click
โ
Export results to TXT, JSON, or PDF
Installation
Install dependencies
pip install ttkbootstrap PyPDF2 reportlab
Download the executable version from Gumroad: https://gum.new/gum/cmkf1hjun000404l2fqerdq39
How to Use
Launch LinkVault PRO.
Select your target folder.
Choose which file types to scan (.txt, .pdf, .html).
Click START EXTRACTION to begin.
Monitor the progress, ETA, and speed in real time.
Once complete, you can export results to TXT, JSON, or PDF.
๐ก Tips:
Double-click a link in the results to open it.
Right-click a link to open the folder containing the source file.
Copy links to clipboard by clicking once.
Use STOP to safely cancel extraction anytime.
Export Options
TXT โ One link per line, ideal for documentation or scripts.
JSON โ Structured export with timestamp and version info.
PDF โ Report-ready format with all links listed.
Developer Info
LinkVault PRO v2.0
Developed by MateTools
Screenshots / UI Highlights
Clean, dark-themed interface using ttkbootstrap
Tree view for extracted links with copy, open, and folder navigation
Progress bar with ETA, speed, and live counter
Export buttons for all supported formats
Code Snippet โ Extraction Engine
def extract_links_worker(folder, file_types):
global link_results, link_count
link_results.clear()
link_count = 0
files = [
os.path.join(root, f)
for root, _, fs in os.walk(folder)
for f in fs
if (f.endswith('.txt') and file_types['txt']) or
(f.endswith('.pdf') and file_types['pdf']) or
(f.endswith(('.html', '.htm')) and file_types['html'])
]
for i, path in enumerate(files, 1):
urls = []
ext = os.path.splitext(path)[1].lower()
if ext in ['.txt', '.html', '.htm']:
with open(path, 'r', encoding='utf-8', errors='ignore') as f:
urls = re.findall(r"https?://[^\s\"'>]+", f.read())
elif ext == '.pdf':
with open(path, 'rb') as f:
reader = PyPDF2.PdfReader(f)
for page in reader.pages:
text = page.extract_text()
if text: urls.extend(re.findall(r"https?://[^\s\"'>]+", text))
for url in urls:
if url not in link_results:
link_results.append(url)
LinkVault PRO v2.0.0 is perfect for developers, researchers, content managers, or anyone who needs to quickly collect and organize links from multiple file types in bulk.
Grab your copy here: https://gum.new/gum/cmkf1hjun000404l2fqerdq39

Top comments (0)