Skip to content

jonluca/source-map-cloner

Repository files navigation

Source Map Cloner

A powerful TypeScript library and CLI tool for extracting and cloning source files from JavaScript source maps. Useful for analyzing minified JavaScript code by recovering the original source files.

Features

  • 🔍 Source Map Extraction: Automatically discovers and extracts source maps from JavaScript files
  • 🌐 Multiple URL Support: Process single URLs or crawl entire sites
  • 📁 Structure Preservation: Maintains original directory structure when extracting files
  • 🚀 TypeScript First: Written in TypeScript with full type definitions
  • 🔌 Flexible Fetchers: Built-in Node.js and browser-compatible HTTP fetchers
  • 📦 Programmatic API: Use as a library in your own projects
  • 🛠️ CLI Tool: Ready-to-use command-line interface

Installation

npm install source-map-cloner
# or
yarn add source-map-cloner
# or
pnpm add source-map-cloner

Usage

CLI Usage

# Basic usage - extract source maps from a URL
npx source-map-cloner https://example.com output-directory

# Multiple URLs
npx source-map-cloner -u https://example.com -u https://another.com output-dir

# Enable crawling to discover linked pages
npx source-map-cloner --crawl https://example.com output-dir

# Custom headers for authentication
npx source-map-cloner --header "Authorization: Bearer token" https://example.com output-dir

# Custom user agent
npx source-map-cloner --user-agent "MyBot 1.0" https://example.com output-dir

Programmatic API

Basic Example with Node.js Fetcher

import { cloneSourceMaps, createNodeFetch, createConsoleLogger } from "source-map-cloner";
import { createNodeFetch } from "source-map-cloner/fetchers";

async function example() {
  const result = await cloneSourceMaps({
    urls: "https://example.com",
    fetch: createNodeFetch(),
    logger: createConsoleLogger(),
  });

  console.log(`Extracted ${result.stats.totalFiles} files`);
  console.log(`Total size: ${result.stats.totalSize} bytes`);

  // Access extracted files
  for (const [path, content] of result.files) {
    console.log(`File: ${path}, Size: ${content.length} bytes`);
  }
}

Advanced Example with Custom Options

import { cloneSourceMaps, createConsoleLogger } from "source-map-cloner";
import { createNodeFetch } from "source-map-cloner/fetchers";

async function advancedExample() {
  const fetch = createNodeFetch({
    headers: {
      "User-Agent": "MyBot/1.0",
      Authorization: "Bearer token",
    },
  });

  const result = await cloneSourceMaps({
    urls: ["https://example.com", "https://another.com"],
    fetch,
    logger: createConsoleLogger(),
    crawl: true,
    verbose: true,
    headers: {
      "Accept-Language": "en-US",
    },
  });

  // Handle errors
  if (result.errors.length > 0) {
    console.error("Errors encountered:");
    for (const error of result.errors) {
      console.error(`- ${error.error} (URL: ${error.url || "N/A"})`);
    }
  }
}

Browser Usage

import { cloneSourceMaps, noopLogger } from "source-map-cloner";
import { createBrowserFetch } from "source-map-cloner/fetchers";

async function browserExample() {
  const result = await cloneSourceMaps({
    urls: "https://example.com",
    fetch: createBrowserFetch(),
    logger: noopLogger, // Silent logger for browser environments
  });

  // Process results...
}

Custom Logger Implementation

import { cloneSourceMaps, Logger } from "source-map-cloner";
import { createNodeFetch } from "source-map-cloner/fetchers";

const customLogger: Logger = {
  info: (msg) => console.log(`[INFO] ${msg}`),
  warn: (msg) => console.warn(`[WARN] ${msg}`),
  error: (msg) => console.error(`[ERROR] ${msg}`),
  debug: (msg) => console.debug(`[DEBUG] ${msg}`),
};

const result = await cloneSourceMaps({
  urls: "https://example.com",
  fetch: createNodeFetch(),
  logger: customLogger,
});

Using the Default Node Fetcher

The createNodeFetch function creates a Node.js-compatible fetcher using the got library with sensible defaults:

import { createNodeFetch } from "source-map-cloner/fetchers";

// Basic usage with default options
const fetch = createNodeFetch();

// With custom headers
const fetchWithAuth = createNodeFetch({
  headers: {
    Authorization: "Bearer token",
    "User-Agent": "MyApp/1.0",
  },
});

// The fetcher automatically handles:
// - HTTP/2 support
// - Cookie jar management
// - Custom cipher configuration for compatibility
// - Automatic retries on network failures
// - Proper encoding handling

API Reference

Main Functions

cloneSourceMaps(options: CloneOptions): Promise<CloneResult>

Main function to clone source maps from URLs.

Options:

  • urls: Single URL string or array of URLs to process
  • fetch: Fetch function for HTTP requests (use createNodeFetch() or createBrowserFetch())
  • logger?: Optional logger instance (defaults to no-op logger)
  • crawl?: Enable crawling to discover linked pages
  • cleanupKnownInvalidFiles?: Remove known invalid files
  • headers?: Additional HTTP headers
  • verbose?: Enable verbose logging

Returns: CloneResult with extracted files, statistics, and errors

Fetchers

createNodeFetch(options?)

Creates a Node.js-compatible fetch function using got.

Options:

  • headers?: Default headers to include with all requests

createBrowserFetch()

Creates a browser-compatible fetch function using the native Fetch API.

Logger Utilities

createConsoleLogger()

Creates a logger that outputs to the console.

noopLogger

A silent logger that discards all output (useful for production or browser environments).

Types

interface CloneResult {
  files: Map<string, string>; // Map of file paths to contents
  stats: {
    totalFiles: number;
    totalSize: number;
    urls: string[];
    duration?: number;
  };
  errors: Array<{
    url?: string;
    file?: string;
    error: string;
  }>;
}

interface Logger {
  info: (message: string) => void;
  warn: (message: string) => void;
  error: (message: string) => void;
  debug?: (message: string) => void;
}

interface FetchFunction {
  (
    url: string,
    options?: { headers?: Record<string, string> },
  ): Promise<{
    body: string;
    statusCode: number;
    requestUrl: string;
  }>;
}

How It Works

  1. URL Processing: Fetches the HTML content from provided URLs
  2. JavaScript Discovery: Extracts all JavaScript file references from the HTML
  3. Source Map Detection: Searches for source map references in JavaScript files (both inline and external)
  4. Source Extraction: Parses source maps and extracts original source content
  5. File Creation: Returns a map of file paths to their contents

The tool handles various source map formats:

  • External source map files (//# sourceMappingURL=...)
  • Inline data URLs
  • Next.js build manifests (_buildManifest.js)

Development

Building

npm run build

Type Checking

npm run typecheck

Running in Development

npm run fetch-source -- https://example.com output-dir

License

MIT

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

Security

This tool is designed for legitimate security research and debugging purposes. Please ensure you have permission before extracting source maps from websites you don't own.

About

Resources

Stars

Watchers

Forks

Packages

No packages published