We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I almost downloaded the data from http://www.gharchive.org/ 64901 files, 839 GiB in size.
But I cannot simply process this data as is:
$ clickhouse-local --query "SELECT * FROM file('*.json.gz', TSV, 'data String') LIMIT 10" Code: 76, e.displayText() = DB::ErrnoException: Cannot open file /opt/milovidov/example_datasets/gharchive/2018-12-30-10.json.gz, errno: 24, strerror: Too many open files (version 20.2.1.1)
The text was updated successfully, but these errors were encountered:
Now it works like a charm. Even with data in garbage format json.gz it is processed with more than 500 MB/sec on old E5-2650v2 and RAID-5 of 8 HDDs.
json.gz
Sorry, something went wrong.
Successfully merging a pull request may close this issue.
I almost downloaded the data from http://www.gharchive.org/
64901 files, 839 GiB in size.
But I cannot simply process this data as is:
The text was updated successfully, but these errors were encountered: