Skip to content

Allow tokenization for gzipped json files in datapreprocess#285

Open
taidnguyen wants to merge 1 commit intomlfoundations:mainfrom
taidnguyen:main
Open

Allow tokenization for gzipped json files in datapreprocess#285
taidnguyen wants to merge 1 commit intomlfoundations:mainfrom
taidnguyen:main

Conversation

@taidnguyen
Copy link
Copy Markdown

Some datasets, such as Dolma, comes in *.json.gz format. Add an option to smart_open and tokenize these files in make_2048.py - Feel free to close if this is available elsewhere already. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant