From 5dd8e4f9ce5ee932ce994d932a01294b48f6e082 Mon Sep 17 00:00:00 2001 From: datechnoman Date: Mon, 4 Dec 2023 10:39:04 +0000 Subject: [PATCH] Upload files to "/" --- multithread_script_5.0.py | 94 +++++++++++++++++++++++++++++++++++++ urlextractor_archiveteam.sh | 28 +++++++++++ 2 files changed, 122 insertions(+) create mode 100644 multithread_script_5.0.py create mode 100644 urlextractor_archiveteam.sh diff --git a/multithread_script_5.0.py b/multithread_script_5.0.py new file mode 100644 index 0000000..2157ea1 --- /dev/null +++ b/multithread_script_5.0.py @@ -0,0 +1,94 @@ +import subprocess +import os +import gzip +import re +from multiprocessing import Pool + +def extract_urls_from_file(file_path): + urls = [] + try: + with gzip.open(file_path, 'rt', encoding='latin-1') as file: + # Process the file line by line + for line in file: + # Extract URLs using regular expression + url_pattern = re.compile(r'http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\\(\\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+') + line_urls = re.findall(url_pattern, line) + urls.extend(line_urls) + except (gzip.BadGzipFile, EOFError) as e: + print(f"Skipping file '{file_path}' due to an error while reading the compressed file: {e}") + except Exception as e: + print(f"An unexpected error occurred while processing '{file_path}': {e}") + + return urls + +def process_file(file_path): + # Extract URLs from the gzipped file + urls = extract_urls_from_file(file_path) + + # Create the output file path with '_urls.txt' extension + output_file_path = os.path.splitext(file_path)[0] + '_urls.txt' + + # Write the URLs to the output file + with open(output_file_path, 'w') as output_file: + output_file.write('\n'.join(urls)) + + print(f"URLs extracted from '{file_path}' and saved to '{output_file_path}'") + + # Remove the original gzipped file + os.remove(file_path) + +def extract_urls_from_directory(directory_path): + # Get the list of files in the directory and sort them + file_list = sorted(os.listdir(directory_path)) + + # Create a multiprocessing Pool with the number of processes + # based on the available CPU cores + pool = Pool(processes=12) + + # Map the file processing function to the list of files + pool.map(process_file, [os.path.join(directory_path, filename) for filename in file_list if filename.endswith('.wat.gz')]) + + # Close the pool to free up resources + pool.close() + pool.join() + +# Read the URLs from the file +with open('urls_to_download.txt', 'r') as file: + urls = file.readlines() + +# Remove any leading/trailing whitespace from the URLs +urls = [url.strip() for url in urls] + +# Define the batch size +batch_size = 100 + +# Define the concurrency level (number of download processes running concurrently) +concurrency_level = 3 + +# Split the URLs into batches +batches = [urls[i:i+batch_size] for i in range(0, len(urls), batch_size)] + +# Iterate over the batches and download URLs +for batch in batches: + # Create a multiprocessing Pool with the specified concurrency level + pool = Pool(processes=concurrency_level) + + for url in batch: + # Create the command to download the URL using axel with 3 connections + command = f'axel -n 3 {url}' + + # Start the subprocess in the background + pool.apply_async(subprocess.run, args=(command,), kwds={'shell': True}) + + # Close the pool to indicate that no more tasks will be added + pool.close() + + # Wait for all processes in the pool to finish + pool.join() + + # Extract URLs from the downloaded files + extract_urls_from_directory(os.getcwd()) + + # Run the urlextractor_archiveteam.sh script + subprocess.run(['bash', 'urlextractor_archiveteam.sh']) + diff --git a/urlextractor_archiveteam.sh b/urlextractor_archiveteam.sh new file mode 100644 index 0000000..0c4c06a --- /dev/null +++ b/urlextractor_archiveteam.sh @@ -0,0 +1,28 @@ +#!/bin/bash + +directory=$(dirname "$0") + +gzip_file() { + file="$1" + gzip "$file" +} + +export -f gzip_file + +for file in "$directory"/*_urls.txt; do + filename=$(basename "$file") + grep -E "http(s)?://(www\.)?mediafire.com" "$file" | sort -u >> "/opt/commoncrawl/export/mediafire_urls.txt" + grep "https://t.me/" "$file" | sort -u >> "/opt/commoncrawl/export/t.me_urls.txt" + grep "https://telegram.me/" "$file" | sort -u >> "/opt/commoncrawl/export/telegram.me_urls.txt" + grep -E "http(s)?://(www\.)?i.imgur.com" "$file" | sort -u >> "/opt/commoncrawl/export/imgur_urls.txt" + grep "sitemap.xml" "$file" | sort -u >> "/opt/commoncrawl/export/sitemap_urls.txt" + grep "https://cdn.discordapp.com/" "$file" | sort -u >> "/opt/commoncrawl/export/discord_urls.txt" + + if [[ $filename != "mediafire_urls.txt" && $filename != "t.me_urls.txt" && $filename != "telegram.me_urls.txt" && $filename != "sitemap_urls.txt" ]]; then + parallel gzip_file ::: "$file" & + fi +done + +# Wait for all gzip processes to finish +wait +