5 practical Python scripts | Web Scraping Tool | ScrapeStorm
Abstract:Here are 5 practical Python scripts covering file operations, automation, and web requests, with clear explanations and ready-to-use code: ScrapeStormFree Download
Here are 5 practical Python scripts covering file operations, automation, and web requests, with clear explanations and ready-to-use code:
1. Batch Rename Files
Use Case: Rename all files in a folder (add prefix/suffix/change extension).
import os
def batch_rename(path, prefix="", suffix="", new_ext=None):
    for filename in os.listdir(path):
        name, ext = os.path.splitext(filename)
        new_name = f"{prefix}{name}{suffix}{new_ext if new_ext else ext}"
        os.rename(
            os.path.join(path, filename),
            os.path.join(path, new_name))
        
# Example: Add prefix "backup_" to all .txt files
batch_rename("/path/to/folder", prefix="backup_", new_ext=".txt")
2. Download Images from URLs
Use Case: Download multiple images from a list of URLs.
import requests
def download_images(urls, save_dir="images"):
    os.makedirs(save_dir, exist_ok=True)
    for i, url in enumerate(urls):
        try:
            response = requests.get(url, stream=True)
            with open(f"{save_dir}/image_{i+1}.jpg", "wb") as f:
                for chunk in response.iter_content(1024):
                    f.write(chunk)
        except Exception as e:
            print(f"Failed to download {url}: {e}")
# Example
urls = ["https://example.com/1.jpg", "https://example.com/2.jpg"]
download_images(urls)
3. Convert CSV to Excel
Use Case: Automate data format conversion.
import pandas as pd
def csv_to_excel(csv_path, excel_path):
    df = pd.read_csv(csv_path)
    df.to_excel(excel_path, index=False)
# Example
csv_to_excel("data.csv", "data.xlsx")
4. Schedule Computer Shutdown
Use Case: Shut down the system after a delay (Windows only).
import os
import time
def schedule_shutdown(minutes):
    seconds = minutes * 60
    print(f"Shutting down in {minutes} minutes...")
    time.sleep(seconds)
    os.system("shutdown /s /t 1")
# Example: Shutdown after 30 minutes
schedule_shutdown(30)
5. Extract All Links from a Webpage
Use Case: Scrape hyperlinks from a URL.
from bs4 import BeautifulSoup
import requests
def extract_links(url):
    try:
        response = requests.get(url)
        soup = BeautifulSoup(response.text, "html.parser")
        return [a["href"] for a in soup.find_all("a", href=True)]
    except Exception as e:
        print(f"Error: {e}")
        return []
# Example
links = extract_links("https://example.com")
print(links)
Disclaimer: This article is contributed by our user. Please advise to remove immediately if any infringement caused.