| import json |
| from pathlib import Path |
|
|
| import datasets |
| from datasets import Value, Sequence, Features |
|
|
|
|
| _CITATION = ''' |
| @article{kirchner2022understanding, |
| title={Understanding AI Alignment Research: A Systematic Analysis}, |
| author={Kirchner, Jan H and Smith, Logan and Thibodeau, Jacques and McDonnell, Kyle and Reynolds, Laria}, |
| journal={arXiv preprint arXiv:2022.4338861}, |
| year={2022} |
| } |
| ''' |
|
|
| _DESCRIPTION = """The AI Alignment Research Dataset is a collection of documents related to AI Alignment and Safety from various books, research papers, and alignment related blog posts.""" |
|
|
| _HOMEPAGE = "https://github.com/StampyAI/alignment-research-dataset" |
|
|
| _LICENSE = "MIT license" |
|
|
| _VERSION_ = '0.0.0' |
|
|
|
|
| def iterate_file(filename): |
| print(filename) |
| with open(filename) as f: |
| for l in f: |
| try: |
| yield json.loads(l) |
| except Exception as e: |
| print(f'Could not parse: {l}') |
|
|
|
|
| |
| def get_type(value): |
| """Recursively get the huggingface type for the provided value.""" |
| if value is None: |
| return None |
| if value and isinstance(value, (tuple, list)): |
| return features.Sequence( |
| get_type(value[0]) |
| ) |
| if value and isinstance(value, dict): |
| return {k: get_type(v) for k, v in value.items()} |
| if isinstance(value, str): |
| return Value('string') |
| if isinstance(value, int): |
| return Value('int32') |
| if isinstance(value, float): |
| return Value('double') |
| if isinstance(value, bool): |
| return Value('bool') |
| return None |
|
|
|
|
| def print_extra_features(files): |
| """Go through all the provided files, and get the non default features for the given file. |
| |
| This can be done manually but would be a hassle. |
| It's assumed that the files contain a json object on each line. |
| """ |
| ignored_keys = [ |
| 'comments', |
| ] |
|
|
| per_file = {} |
| for filename in sorted(files): |
| extra_types = {} |
| for item in iterate_file(filename): |
| for k, v in item.items(): |
| if (k not in extra_types or not extra_types[k]) and k not in ignored_keys and k not in DEFAULT_FEATURES: |
| extra_types[k] = get_type(v) |
| per_file[filename] = extra_types |
|
|
| print('DATASOURCES = {') |
| for k, features in per_file.items(): |
| vals = ',\n'.join(f" '{k}': {v}" for k, v in features.items()) |
| print(f" '{k.stem}': #\n{vals}\n $,".replace('#', '{').replace('$', '}')) |
| print('}') |
|
|
|
|
| |
| DEFAULT_FEATURES = { |
| 'id': Value('string'), |
| 'source': Value('string'), |
| 'title': Value('string'), |
| 'text': Value('large_string'), |
| 'url': Value('string'), |
| 'date_published': Value(dtype='string'), |
| 'authors': Sequence(feature=Value(dtype='string'), length=-1), |
| 'summary': Sequence(feature=Value(dtype='string'), length=-1), |
| 'source_type': Value(dtype='string'), |
| } |
|
|
|
|
| |
| DATASOURCES = { |
| 'agentmodels': { |
| 'book_title': Value(dtype='string'), |
| }, |
| 'agisf': {}, |
| 'aisafety.info': {}, |
| 'alignmentforum': { |
| 'karma': Value(dtype='int32'), |
| 'votes': Value(dtype='int32'), |
| 'words': Value(dtype='int32'), |
| 'comment_count': Value(dtype='int32'), |
| 'tags': Sequence(feature=Value(dtype='string')), |
| 'modified_at': Value(dtype='string'), |
| }, |
| 'arbital': { |
| 'alias': Value(dtype='string'), |
| 'tags': Sequence(feature=Value(dtype='string')), |
| }, |
| 'arxiv': { |
| 'data_last_modified': Value(dtype='string'), |
| 'abstract': Value(dtype='string'), |
| 'author_comment': Value(dtype='string'), |
| 'journal_ref': Value(dtype='string'), |
| 'doi': Value(dtype='string'), |
| 'primary_category': Value(dtype='string'), |
| 'categories': Sequence(feature=Value(dtype='string'), length=-1), |
| }, |
| 'blogs': { |
| 'initial_source': Value(dtype='string'), |
| }, |
| 'distill': { |
| 'abstract': Value(dtype='string'), |
| 'journal_ref': Value(dtype='string'), |
| 'doi': Value(dtype='string'), |
| 'bibliography_bib': Sequence(feature={'title': Value(dtype='string')}, length=-1), |
| }, |
| 'eaforum': { |
| 'karma': Value(dtype='int32'), |
| 'votes': Value(dtype='int32'), |
| 'words': Value(dtype='int32'), |
| 'comment_count': Value(dtype='int32'), |
| 'tags': Sequence(feature=Value(dtype='string')), |
| 'modified_at': Value(dtype='string'), |
| }, |
| 'lesswrong': { |
| 'karma': Value(dtype='int32'), |
| 'votes': Value(dtype='int32'), |
| 'words': Value(dtype='int32'), |
| 'comment_count': Value(dtype='int32'), |
| 'tags': Sequence(feature=Value(dtype='string')), |
| 'modified_at': Value(dtype='string'), |
| }, |
| 'special_docs': {}, |
| 'youtube': {}, |
| } |
|
|
|
|
| def join_features(features, to_join): |
| """Recursively join the provided dicts. |
| |
| `to_join` can either be a dict to be merged, or a list of dicts to merge. |
| """ |
| if not to_join: |
| return Features(features) |
| if isinstance(to_join, dict): |
| return Features(dict(features, **to_join)) |
| return join_features(dict(features, **to_join[0]), to_join[1:]) |
|
|
|
|
| class AlignmentResearchDatasetConfig(datasets.BuilderConfig): |
| """BuilderConfig for AlignmentResaerchDataset.""" |
|
|
| def __init__(self, sources, features, **kwargs): |
| """BuilderConfig for AlignmentResaerchDataset. |
| |
| :param List[string] sources: the sources which will be used by this config |
| """ |
| super().__init__(version=datasets.Version(_VERSION_), **kwargs) |
| self.sources = sources |
| self.features = join_features(DEFAULT_FEATURES, features) |
|
|
| @property |
| def files(self): |
| return [f'{source}.jsonl' for source in self.sources] |
|
|
|
|
| class AlignmentResaerchDataset(datasets.GeneratorBasedBuilder): |
| VERSION = datasets.Version(_VERSION_) |
|
|
| BUILDER_CONFIGS = [ |
| AlignmentResearchDatasetConfig( |
| name='all', |
| description='All data files', |
| sources=list(DATASOURCES.keys()), |
| features=list(DATASOURCES.values()) |
| ) |
| ] + [ |
| AlignmentResearchDatasetConfig(name=source, sources=[source], features=features) for source, features in DATASOURCES.items() |
| ] |
| DEFAULT_CONFIG_NAME = 'all' |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=self.config.features, |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| downloaded_files = dl_manager.download_and_extract(self.config.files) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={'files': downloaded_files} |
| ) |
| ] |
|
|
| |
| def _generate_examples(self, files): |
| seen = set() |
|
|
| def is_good(item): |
| item_id = item and item.get('id') |
| if not item_id or item_id in seen: |
| return False |
| seen.add(item_id) |
|
|
| return item['text'] not in [None, '', 'n/a'] |
|
|
| def prepare_example(item): |
| return item['id'], {k: item.get(k) for k in self.config.features} |
|
|
| lines = (item for filename in files for item in iterate_file(filename)) |
| for item in map(prepare_example, filter(is_good, lines)): |
| yield item |
|
|