A modern web application for extracting and downloading tables from Wikipedia pages into multiple formats.
This tool provides a clean interface to fetch, select, and download any table from a given Wikipedia URL. It is designed to streamline the data collection process for researchers, students, and data analysts by converting HTML tables into structured file formats.
- Multi-Format Export: Download tables as CSV, XLSX, or JSON
- Batch Downloads: Select and download multiple tables simultaneously
- Data Preview: View table dimensions (rows and columns) before downloading
- Modern UI: Fully responsive interface with light and dark modes
Category | Technology |
---|---|
Framework | Next.js (App Router) |
Language | TypeScript |
UI | Tailwind CSS & Shadcn/ui |
Scraping | Cheerio |
File Generation | SheetJS (xlsx) |
- Node.js
- npm or a compatible package manager
- Clone the repository:
git clone https://github.com/dhaifullahr/wikitable.git
cd wikitable
- Install dependencies:
npm install
- Run the development server:
npm run dev
- Open your browser and navigate to:
http://localhost:3000
- Enter a Wikipedia URL in the input field
- Click "Fetch Tables" to extract all tables from the page
- Preview tables and their dimensions
- Select one or multiple tables
- Choose your desired export format (CSV, XLSX, or JSON)
- Click "Download" to save your selected tables
Contributions are welcome! Please feel free to submit a Pull Request.
This project is licensed under the MIT License.
- Wikipedia for providing the data source