File size: 2,757 Bytes
a9e0f20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f77e570
 
 
 
 
 
 
a9e0f20
f77e570
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
---
dataset_info:
  features:
  - name: url
    dtype: string
  - name: domain
    dtype: string
  splits:
  - name: train
    num_bytes: 432578990554
    num_examples: 4506333003
  download_size: 298143463982
  dataset_size: 432578990554
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
license: odc-by
task_categories:
- text-generation
language:
- en
size_categories:
- 1B<n<10B
---

# Dataset Card for zyda-2_urls
This dataset provides the URLs and top-level domains associated with training records in [Zyphra/Zyda-2](https://huggingface.co/datasets/Zyphra/Zyda-2). It is part of a [collection of datasets](https://huggingface.co/collections/nhagar/llm-urls-neurips-681698adac0862be6c65c72b) curated to make exploring LLM training datasets more straightforward and accessible. 

## Dataset Details

### Dataset Description

This dataset was created by downloading the source data, extracting URLs and top-level domains, and retaining only those record identifiers. In doing so, it allows researchers and practitioners to explore the contents of these training datasets without having to manage terabytes of raw text. You can explore the pipeline used to construct this dataset on [GitHub](https://github.com/NHagar/cc-genealogy). 

- **Curated by:** [Nick Hagar](https://huggingface.co/nhagar) and [Jack Bandy](https://huggingface.co/jackbandy)
- **License:** Same as source dataset

### Dataset Sources

- **Repository:** [Zyphra/Zyda-2](https://huggingface.co/datasets/Zyphra/Zyda-2)

## Uses

This dataset is intended to allow researchers and practitioners to analyze the contents of large LLM training datasets without having to wade through terabytes of unwieldy text data. 

### Direct Use

The main use case for these data is to explore the contents of LLM training datasets at scale. This might involve:
- Identifying the most-used websites
- Categorizing URLs to understand domain- or topic-level dataset composition
- Comparing URLs across datasets
- Digging into inclusion/exclusion patterns for a particular website

### Out-of-Scope Use

This dataset is not intend to replicate or replace the source data, nor is it intended to enable large-scale scraping of the URLs listed. For source text, refer to the original dataset.

## Dataset Structure

This dataset contains every record with a URL from the source dataset. It contains two columns:
- `url`: The raw URL associated with each record
- `domain`: The top-level domain for each URL, extracted with `tldextract`

## Citation [optional]

<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->

**BibTeX:**

[More Information Needed]

**APA:**

[More Information Needed]