Gaëtan Caillaut commited on
Commit ·
6453182
1
Parent(s): b6388cf
flatten dataset
Browse files- README.md +6 -8
- frwiki_good_pages_el.py +12 -17
README.md
CHANGED
|
@@ -41,14 +41,12 @@ It is intended to be used to train Entity Linking (EL) systems. Links in article
|
|
| 41 |
{
|
| 42 |
"title": "Title of the page",
|
| 43 |
"qid": "QID of the corresponding Wikidata entity",
|
| 44 |
-
"
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
"qids": ["QID of each entity"],
|
| 51 |
-
}
|
| 52 |
}
|
| 53 |
```
|
| 54 |
|
|
|
|
| 41 |
{
|
| 42 |
"title": "Title of the page",
|
| 43 |
"qid": "QID of the corresponding Wikidata entity",
|
| 44 |
+
"words": ["tokens"],
|
| 45 |
+
"wikipedia": ["Wikipedia description of each entity"],
|
| 46 |
+
"wikidata": ["Wikidata description of each entity"],
|
| 47 |
+
"labels": ["NER labels"],
|
| 48 |
+
"titles": ["Wikipedia title of each entity"],
|
| 49 |
+
"qids": ["QID of each entity"],
|
|
|
|
|
|
|
| 50 |
}
|
| 51 |
```
|
| 52 |
|
frwiki_good_pages_el.py
CHANGED
|
@@ -47,14 +47,11 @@ def read_file(path):
|
|
| 47 |
# Find for instance the citation on arxiv or on the dataset repo/website
|
| 48 |
_CITATION = ""
|
| 49 |
|
| 50 |
-
# TODO: Add description of the dataset here
|
| 51 |
-
# You can copy an official description
|
| 52 |
_DESCRIPTION = """\
|
| 53 |
French Wikipedia dataset for Entity Linking
|
| 54 |
"""
|
| 55 |
|
| 56 |
-
|
| 57 |
-
_HOMEPAGE = ""
|
| 58 |
|
| 59 |
# TODO: Add the licence for the dataset here if you can find it
|
| 60 |
_LICENSE = ""
|
|
@@ -94,7 +91,7 @@ def text_to_el_features(doc_qid, doc_title, text, title2qid, title2wikipedia, ti
|
|
| 94 |
mention_title = m.group(1)
|
| 95 |
mention = m.group(2)
|
| 96 |
|
| 97 |
-
mention_qid = title2qid.get(mention_title, "")
|
| 98 |
mention_wikipedia = title2wikipedia.get(mention_title, "")
|
| 99 |
mention_wikidata = title2wikidata.get(mention_title, "")
|
| 100 |
|
|
@@ -131,9 +128,9 @@ def text_to_el_features(doc_qid, doc_title, text, title2qid, title2wikipedia, ti
|
|
| 131 |
text_dict["wikidata"].extend([None] * len_mention)
|
| 132 |
else:
|
| 133 |
len_mention_tail = len(mention_words) - 1
|
| 134 |
-
wikipedia_words = mention_wikipedia.split()
|
| 135 |
-
wikidata_words = mention_wikidata.split()
|
| 136 |
-
title_words = mention_title.replace("_", " ").split()
|
| 137 |
|
| 138 |
text_dict["labels"].extend(["B"] + ["I"] * len_mention_tail)
|
| 139 |
text_dict["qids"].extend([mention_qid] + [None] * len_mention_tail)
|
|
@@ -154,7 +151,7 @@ def text_to_el_features(doc_qid, doc_title, text, title2qid, title2wikipedia, ti
|
|
| 154 |
text_dict["titles"].extend([None] * len_tail)
|
| 155 |
text_dict["wikipedia"].extend([None] * len_tail)
|
| 156 |
text_dict["wikidata"].extend([None] * len_tail)
|
| 157 |
-
res
|
| 158 |
return res
|
| 159 |
|
| 160 |
|
|
@@ -188,14 +185,12 @@ class FrWikiGoodPagesELDataset(datasets.GeneratorBasedBuilder):
|
|
| 188 |
features = datasets.Features({
|
| 189 |
"title": datasets.Value("string"),
|
| 190 |
"qid": datasets.Value("string"),
|
| 191 |
-
"
|
| 192 |
-
|
| 193 |
-
|
| 194 |
-
|
| 195 |
-
|
| 196 |
-
|
| 197 |
-
"qids": [datasets.Value("string")],
|
| 198 |
-
}
|
| 199 |
})
|
| 200 |
|
| 201 |
return datasets.DatasetInfo(
|
|
|
|
| 47 |
# Find for instance the citation on arxiv or on the dataset repo/website
|
| 48 |
_CITATION = ""
|
| 49 |
|
|
|
|
|
|
|
| 50 |
_DESCRIPTION = """\
|
| 51 |
French Wikipedia dataset for Entity Linking
|
| 52 |
"""
|
| 53 |
|
| 54 |
+
_HOMEPAGE = "https://github.com/GaaH/frwiki_good_pages_el"
|
|
|
|
| 55 |
|
| 56 |
# TODO: Add the licence for the dataset here if you can find it
|
| 57 |
_LICENSE = ""
|
|
|
|
| 91 |
mention_title = m.group(1)
|
| 92 |
mention = m.group(2)
|
| 93 |
|
| 94 |
+
mention_qid = title2qid.get(mention_title, "").replace("_", " ")
|
| 95 |
mention_wikipedia = title2wikipedia.get(mention_title, "")
|
| 96 |
mention_wikidata = title2wikidata.get(mention_title, "")
|
| 97 |
|
|
|
|
| 128 |
text_dict["wikidata"].extend([None] * len_mention)
|
| 129 |
else:
|
| 130 |
len_mention_tail = len(mention_words) - 1
|
| 131 |
+
# wikipedia_words = mention_wikipedia.split()
|
| 132 |
+
# wikidata_words = mention_wikidata.split()
|
| 133 |
+
# title_words = mention_title.replace("_", " ").split()
|
| 134 |
|
| 135 |
text_dict["labels"].extend(["B"] + ["I"] * len_mention_tail)
|
| 136 |
text_dict["qids"].extend([mention_qid] + [None] * len_mention_tail)
|
|
|
|
| 151 |
text_dict["titles"].extend([None] * len_tail)
|
| 152 |
text_dict["wikipedia"].extend([None] * len_tail)
|
| 153 |
text_dict["wikidata"].extend([None] * len_tail)
|
| 154 |
+
res.update(text_dict)
|
| 155 |
return res
|
| 156 |
|
| 157 |
|
|
|
|
| 185 |
features = datasets.Features({
|
| 186 |
"title": datasets.Value("string"),
|
| 187 |
"qid": datasets.Value("string"),
|
| 188 |
+
"words": [datasets.Value("string")],
|
| 189 |
+
"wikipedia": [datasets.Value("string")],
|
| 190 |
+
"wikidata": [datasets.Value("string")],
|
| 191 |
+
"labels": [datasets.ClassLabel(names=_CLASS_LABELS)],
|
| 192 |
+
"titles": [datasets.Value("string")],
|
| 193 |
+
"qids": [datasets.Value("string")],
|
|
|
|
|
|
|
| 194 |
})
|
| 195 |
|
| 196 |
return datasets.DatasetInfo(
|