add train
Browse files- binhvq_dedup.py +4 -3
- train.jsonl.zst +3 -0
binhvq_dedup.py
CHANGED
|
@@ -23,7 +23,8 @@ Preprocessed:
|
|
| 23 |
|
| 24 |
_REPO_URL = "https://huggingface.co/datasets/imthanhlv/binhvq_dedup/tree/main/"
|
| 25 |
_URLS = {
|
| 26 |
-
"
|
|
|
|
| 27 |
}
|
| 28 |
|
| 29 |
|
|
@@ -62,7 +63,7 @@ class Binhvq(datasets.GeneratorBasedBuilder):
|
|
| 62 |
return [
|
| 63 |
datasets.SplitGenerator(
|
| 64 |
name=datasets.Split.TRAIN,
|
| 65 |
-
gen_kwargs={"filepath": downloaded_files["
|
| 66 |
),
|
| 67 |
datasets.SplitGenerator(
|
| 68 |
name=datasets.Split.VALIDATION,
|
|
@@ -74,6 +75,6 @@ class Binhvq(datasets.GeneratorBasedBuilder):
|
|
| 74 |
logging.warning(f"Generating examples from {filepath}")
|
| 75 |
_id = 0
|
| 76 |
reader = lm_dataformat.Reader(filepath)
|
| 77 |
-
for doc in
|
| 78 |
yield _id, {"text": doc}
|
| 79 |
_id += 1
|
|
|
|
| 23 |
|
| 24 |
_REPO_URL = "https://huggingface.co/datasets/imthanhlv/binhvq_dedup/tree/main/"
|
| 25 |
_URLS = {
|
| 26 |
+
"train" "train.jsonl.zst",
|
| 27 |
+
"val": "val.jsonl.zst"
|
| 28 |
}
|
| 29 |
|
| 30 |
|
|
|
|
| 63 |
return [
|
| 64 |
datasets.SplitGenerator(
|
| 65 |
name=datasets.Split.TRAIN,
|
| 66 |
+
gen_kwargs={"filepath": downloaded_files["train"]},
|
| 67 |
),
|
| 68 |
datasets.SplitGenerator(
|
| 69 |
name=datasets.Split.VALIDATION,
|
|
|
|
| 75 |
logging.warning(f"Generating examples from {filepath}")
|
| 76 |
_id = 0
|
| 77 |
reader = lm_dataformat.Reader(filepath)
|
| 78 |
+
for doc in reader.read_jsonl(filepath):
|
| 79 |
yield _id, {"text": doc}
|
| 80 |
_id += 1
|
train.jsonl.zst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:add7eebdc28dae0edcbc9b3e51f06469870b573023fabce0703e6d388c4a218b
|
| 3 |
+
size 26841447213
|