parquet-converter commited on
Commit
5eac1be
·
1 Parent(s): 68844f7

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,27 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bin.* filter=lfs diff=lfs merge=lfs -text
5
- *.bz2 filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.onnx filter=lfs diff=lfs merge=lfs -text
14
- *.ot filter=lfs diff=lfs merge=lfs -text
15
- *.parquet filter=lfs diff=lfs merge=lfs -text
16
- *.pb filter=lfs diff=lfs merge=lfs -text
17
- *.pt filter=lfs diff=lfs merge=lfs -text
18
- *.pth filter=lfs diff=lfs merge=lfs -text
19
- *.rar filter=lfs diff=lfs merge=lfs -text
20
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
- *.tar.* filter=lfs diff=lfs merge=lfs -text
22
- *.tflite filter=lfs diff=lfs merge=lfs -text
23
- *.tgz filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.gitignore DELETED
@@ -1,2 +0,0 @@
1
- raw/*
2
- .ipynb_checkpoints
 
 
 
README.md DELETED
@@ -1,17 +0,0 @@
1
- This dataset contains postings and comments from the following recurring threads on [Hacker News](http://news.ycombinator.com/)
2
-
3
- 1. Ask HN: Who is hiring?
4
- 2. Ask HN: Who wants to be hired?
5
- 3. Freelancer? Seeking freelancer?
6
-
7
- These post types are stored in datasets called `hiring`, `wants_to_be_hired` and `freelancer` respectively.
8
-
9
- Each type of posting has occurred on a regular basis for several years. You can identify when each comment/listing was added through the CommentTime field. The `ParentTitle` also indicates the date of the parent thread in text (e.g. `Ask HN: Who is hiring? (March 2021)`)
10
-
11
- This dataset is not programmatically reproducible from source because it was uploaded as an experiment with HF datasets. The raw data was created by querying the public table `bigquery-public-data.hacker_news.full` in Google BigQuery.
12
-
13
- Email addresses have been redacted from the dataset.
14
-
15
- If this dataset is interesting/useful, I (Dan Becker) will look into improving reproducibility and other general clean-up.
16
-
17
- This dataset may be useful for finding trends in tech and tech job listings.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/freelancer-00000-of-00001.parquet → dansbecker--hackernews_hiring_posts/parquet-freelancer.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec52e70ca4eea5c5d9b6954aa0bdbaafd6c6c6388b0f7c808f95c583d4fbe9bc
3
- size 5973001
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e152e4fbc4eaf74db34b8dca12e11ff9c441d3b5fb23d145c33a25db59b9263b
3
+ size 5985114
data/freelancer/dataset.arrow → dansbecker--hackernews_hiring_posts/parquet-hiring.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c98a8f743575b496b0c37547fdabdb311f7fb932a51a41ed125846d71de6c7b
3
- size 11166008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:600614909754b4ba3b53bf9a5688445f376f660587c71dfffe83a1a81d484c4e
3
+ size 47128917
data/wants_to_be_hired-00000-of-00001.parquet → dansbecker--hackernews_hiring_posts/parquet-wants_to_be_hired.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6cbf0ceae498c0dcc6e93cdf1ab8f389d8074e5e07d61a69604e90f049a5968b
3
- size 5778819
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95eaedc99c0b355f0560bb770eb5b6e05714e5dbc2f5a825163fde41cc49061c
3
+ size 5880973
data/dataset_dict.json DELETED
@@ -1 +0,0 @@
1
- {"splits": ["hiring", "wants_to_be_hired", "freelancer"]}
 
 
data/freelancer/dataset_info.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "builder_name": null,
3
- "citation": "",
4
- "config_name": null,
5
- "dataset_size": null,
6
- "description": "",
7
- "download_checksums": null,
8
- "download_size": null,
9
- "features": {
10
- "text": {
11
- "dtype": "string",
12
- "id": null,
13
- "_type": "Value"
14
- },
15
- "CommentTime": {
16
- "dtype": "string",
17
- "id": null,
18
- "_type": "Value"
19
- },
20
- "CommentAuthor": {
21
- "dtype": "string",
22
- "id": null,
23
- "_type": "Value"
24
- },
25
- "ParentTitle": {
26
- "dtype": "string",
27
- "id": null,
28
- "_type": "Value"
29
- }
30
- },
31
- "homepage": "",
32
- "license": "",
33
- "post_processed": null,
34
- "post_processing_size": null,
35
- "size_in_bytes": null,
36
- "splits": null,
37
- "supervised_keys": null,
38
- "task_templates": null,
39
- "version": null
40
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/freelancer/state.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "_data_files": [
3
- {
4
- "filename": "dataset.arrow"
5
- }
6
- ],
7
- "_fingerprint": "4f8f711b9136404c",
8
- "_format_columns": null,
9
- "_format_kwargs": {},
10
- "_format_type": null,
11
- "_indexes": {},
12
- "_output_all_columns": false,
13
- "_split": null
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/hiring-00000-of-00001.parquet DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f73ecadb195114287b162969f8db2ef0c9509e76dec4e3aa401319d6f09aa14b
3
- size 46640019
 
 
 
 
data/hiring/dataset.arrow DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8827da89590c888c1e2d167c04f4fb527ed342c9cc23c61bbfa6646cfcf6f2f
3
- size 81975968
 
 
 
 
data/hiring/dataset_info.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "builder_name": null,
3
- "citation": "",
4
- "config_name": null,
5
- "dataset_size": null,
6
- "description": "",
7
- "download_checksums": null,
8
- "download_size": null,
9
- "features": {
10
- "text": {
11
- "dtype": "string",
12
- "id": null,
13
- "_type": "Value"
14
- },
15
- "CommentTime": {
16
- "dtype": "string",
17
- "id": null,
18
- "_type": "Value"
19
- },
20
- "CommentAuthor": {
21
- "dtype": "string",
22
- "id": null,
23
- "_type": "Value"
24
- },
25
- "ParentTitle": {
26
- "dtype": "string",
27
- "id": null,
28
- "_type": "Value"
29
- }
30
- },
31
- "homepage": "",
32
- "license": "",
33
- "post_processed": null,
34
- "post_processing_size": null,
35
- "size_in_bytes": null,
36
- "splits": null,
37
- "supervised_keys": null,
38
- "task_templates": null,
39
- "version": null
40
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/hiring/state.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "_data_files": [
3
- {
4
- "filename": "dataset.arrow"
5
- }
6
- ],
7
- "_fingerprint": "e5166a077fc0f8e1",
8
- "_format_columns": null,
9
- "_format_kwargs": {},
10
- "_format_type": null,
11
- "_indexes": {},
12
- "_output_all_columns": false,
13
- "_split": null
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/wants_to_be_hired/dataset.arrow DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc638e6f45b651e6e166298e20f7c68b2fb6cdb5aac357a728778dffa9114ac1
3
- size 11373152
 
 
 
 
data/wants_to_be_hired/dataset_info.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "builder_name": null,
3
- "citation": "",
4
- "config_name": null,
5
- "dataset_size": null,
6
- "description": "",
7
- "download_checksums": null,
8
- "download_size": null,
9
- "features": {
10
- "text": {
11
- "dtype": "string",
12
- "id": null,
13
- "_type": "Value"
14
- },
15
- "CommentTime": {
16
- "dtype": "string",
17
- "id": null,
18
- "_type": "Value"
19
- },
20
- "CommentAuthor": {
21
- "dtype": "string",
22
- "id": null,
23
- "_type": "Value"
24
- },
25
- "ParentTitle": {
26
- "dtype": "string",
27
- "id": null,
28
- "_type": "Value"
29
- }
30
- },
31
- "homepage": "",
32
- "license": "",
33
- "post_processed": null,
34
- "post_processing_size": null,
35
- "size_in_bytes": null,
36
- "splits": null,
37
- "supervised_keys": null,
38
- "task_templates": null,
39
- "version": null
40
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/wants_to_be_hired/state.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "_data_files": [
3
- {
4
- "filename": "dataset.arrow"
5
- }
6
- ],
7
- "_fingerprint": "f8da56cf67a872ff",
8
- "_format_columns": null,
9
- "_format_kwargs": {},
10
- "_format_type": null,
11
- "_indexes": {},
12
- "_output_all_columns": false,
13
- "_split": null
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
hn_hiring_posts.ipynb DELETED
@@ -1,194 +0,0 @@
1
- {
2
- "cells": [
3
- {
4
- "cell_type": "code",
5
- "execution_count": 1,
6
- "id": "056aa255-fda1-4cde-be24-459f6ad2c8b9",
7
- "metadata": {},
8
- "outputs": [
9
- {
10
- "name": "stderr",
11
- "output_type": "stream",
12
- "text": [
13
- "/usr/local/anaconda3/lib/python3.8/site-packages/IPython/core/interactiveshell.py:3165: DtypeWarning: Columns (4) have mixed types.Specify dtype option on import or set low_memory=False.\n",
14
- " has_raised = await self.run_ast_nodes(code_ast.body, cell_name,\n"
15
- ]
16
- }
17
- ],
18
- "source": [
19
- "import pandas as pd\n",
20
- "\n",
21
- "df = pd.read_csv('raw/bq-results-20211206-133858-irtkgx60el7i.csv').drop(['ParentUrl', 'ParentAuthor', 'ParentTime', 'ParentScore'], axis=1)\n",
22
- "\n",
23
- "# There's a mix of HTML, sanitized links and raw Unicode that we'd like to clean up\n",
24
- "df.text = df.text.str.replace('<p>', '\\n')\n",
25
- "strings_to_remove = ['rel=\"nofollow\"', '<pre>', '</pre>', '<i>', '</i>', '<code>', '</code>', '&gt;']\n",
26
- "email_regex = '[a-zA-Z0-9._-]{0,30}@[a-zA-Z0-9._-]{0,20}\\.[a-zA-Z0-9_-]{2,3}' # for redacting emails\n",
27
- "munged_url_regex = 'http(s)?:\\&\\#.*?\\<\\/a>'\n",
28
- "\n",
29
- "for string in strings_to_remove:\n",
30
- " df.text = df.text.str.replace(string, '')\n",
31
- "\n",
32
- "\n",
33
- "df.text = df.text.replace(email_regex, 'REDACTED_EMAIL', regex=True)\n",
34
- "df.text = df.text.replace(munged_url_regex, '', regex=True)\n",
35
- "\n",
36
- " # fix some unicode issues\n",
37
- "df.text = df.text.str.replace('&#x27;', \"'\")\n",
38
- "df.text = df.text.str.replace('&#x2F;', \"/\")\n",
39
- "df.text = df.text.str.replace(\"&quot;\", '\"')\n",
40
- "\n",
41
- "hiring_df = df[df.ParentTitle.str.lower().str.contains('who is hiring')]\n",
42
- "wants_to_be_hired_df = df[df.ParentTitle.str.lower().str.contains('wants to be hired')]\n",
43
- "freelancer_df = df[df.ParentTitle.str.lower().str.contains('freelancer')]\n"
44
- ]
45
- },
46
- {
47
- "cell_type": "code",
48
- "execution_count": 2,
49
- "id": "57ca7c96-d94c-4e65-8113-da7729558247",
50
- "metadata": {},
51
- "outputs": [
52
- {
53
- "name": "stderr",
54
- "output_type": "stream",
55
- "text": [
56
- "/Users/dan.becker/Desktop/hackernews_hiring_dataset/. is already a clone of https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts. Make sure you pull the latest changes with `repo.git_pull()`.\n",
57
- "Several commits (5) will be pushed upstream.\n",
58
- "The progress bars may be unreliable.\n"
59
- ]
60
- },
61
- {
62
- "data": {
63
- "application/vnd.jupyter.widget-view+json": {
64
- "model_id": "63ddc30f9ba5420faf6e34f6fa7144fc",
65
- "version_major": 2,
66
- "version_minor": 0
67
- },
68
- "text/plain": [
69
- "Upload file data/hiring/dataset.arrow: 0%| | 32.0k/78.2M [00:00<?, ?B/s]"
70
- ]
71
- },
72
- "metadata": {},
73
- "output_type": "display_data"
74
- },
75
- {
76
- "data": {
77
- "application/vnd.jupyter.widget-view+json": {
78
- "model_id": "be96819a155b4f17933d3ff82a2fb70d",
79
- "version_major": 2,
80
- "version_minor": 0
81
- },
82
- "text/plain": [
83
- "Upload file data/wants_to_be_hired/dataset.arrow: 0%| | 32.0k/10.8M [00:00<?, ?B/s]"
84
- ]
85
- },
86
- "metadata": {},
87
- "output_type": "display_data"
88
- },
89
- {
90
- "data": {
91
- "application/vnd.jupyter.widget-view+json": {
92
- "model_id": "3ef183014a04462e97153ca6673c5f11",
93
- "version_major": 2,
94
- "version_minor": 0
95
- },
96
- "text/plain": [
97
- "Upload file data/freelancer/dataset.arrow: 0%| | 32.0k/10.6M [00:00<?, ?B/s]"
98
- ]
99
- },
100
- "metadata": {},
101
- "output_type": "display_data"
102
- },
103
- {
104
- "name": "stderr",
105
- "output_type": "stream",
106
- "text": [
107
- "remote: ----------------------------------------------------------\u001b[33m \n",
108
- "remote: Your push was accepted, but with warnings: \n",
109
- "remote: - warning : empty or missing yaml metadata in card (dansbecker/hackernews_hiring_posts) \n",
110
- "remote: help: please find help at https://huggingface.co/docs/hub/model-repos\u001b[32m \n",
111
- "remote: ---------------------------------------------------------- \n",
112
- "remote: Please find the documentation at: \n",
113
- "remote: https://huggingface.co/docs/hub/model-repos\u001b(B\u001b[m \n",
114
- "remote: ---------------------------------------------------------- \n",
115
- "remote: -------------------------------------------------------------------------\u001b[31m \n",
116
- "remote: Your push was rejected because it contains files larger than 10M. \n",
117
- "remote: Please use https://git-lfs.github.com/ to store larger files.\u001b(B\u001b[m \n",
118
- "remote: ------------------------------------------------------------------------- \n",
119
- "remote: Offending files: \n",
120
- "remote: - raw/bq-results-20211206-133858-irtkgx60el7i.csv (ref: refs/heads/main) \n",
121
- "To https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts\n",
122
- " ! [remote rejected] main -> main (pre-receive hook declined)\n",
123
- "error: failed to push some refs to 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n",
124
- "\n"
125
- ]
126
- },
127
- {
128
- "ename": "OSError",
129
- "evalue": "remote: ----------------------------------------------------------\u001b[33m \nremote: Your push was accepted, but with warnings: \nremote: - warning : empty or missing yaml metadata in card (dansbecker/hackernews_hiring_posts) \nremote: help: please find help at https://huggingface.co/docs/hub/model-repos\u001b[32m \nremote: ---------------------------------------------------------- \nremote: Please find the documentation at: \nremote: https://huggingface.co/docs/hub/model-repos\u001b(B\u001b[m \nremote: ---------------------------------------------------------- \nremote: -------------------------------------------------------------------------\u001b[31m \nremote: Your push was rejected because it contains files larger than 10M. \nremote: Please use https://git-lfs.github.com/ to store larger files.\u001b(B\u001b[m \nremote: ------------------------------------------------------------------------- \nremote: Offending files: \nremote: - raw/bq-results-20211206-133858-irtkgx60el7i.csv (ref: refs/heads/main) \nTo https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts\n ! [remote rejected] main -> main (pre-receive hook declined)\nerror: failed to push some refs to 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n",
130
- "output_type": "error",
131
- "traceback": [
132
- "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
133
- "\u001b[0;31mCalledProcessError\u001b[0m Traceback (most recent call last)",
134
- "\u001b[0;32m/usr/local/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py\u001b[0m in \u001b[0;36mgit_push\u001b[0;34m(self, upstream, blocking, auto_lfs_prune)\u001b[0m\n\u001b[1;32m 998\u001b[0m \u001b[0;32mif\u001b[0m \u001b[0mreturn_code\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 999\u001b[0;31m raise subprocess.CalledProcessError(\n\u001b[0m\u001b[1;32m 1000\u001b[0m \u001b[0mreturn_code\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mprocess\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0margs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0moutput\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mstdout\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mstderr\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mstderr\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
135
- "\u001b[0;31mCalledProcessError\u001b[0m: Command '['git', 'push', '--set-upstream', 'origin', 'main']' returned non-zero exit status 1.",
136
- "\nDuring handling of the above exception, another exception occurred:\n",
137
- "\u001b[0;31mOSError\u001b[0m Traceback (most recent call last)",
138
- "\u001b[0;32m<ipython-input-2-38e9608c0524>\u001b[0m in \u001b[0;36m<module>\u001b[0;34m\u001b[0m\n\u001b[1;32m 14\u001b[0m \u001b[0mrepo\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'*.ipynb'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 15\u001b[0m \u001b[0mrepo\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'README.md'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m---> 16\u001b[0;31m \u001b[0mrepo\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mpush_to_hub\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mcommit_message\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;34m'Regular update'\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mblocking\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;32mTrue\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mclean_ok\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;32mFalse\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m",
139
- "\u001b[0;32m/usr/local/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py\u001b[0m in \u001b[0;36mpush_to_hub\u001b[0;34m(self, commit_message, blocking, clean_ok, auto_lfs_prune)\u001b[0m\n\u001b[1;32m 1230\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mauto_lfs_track\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;32mTrue\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1231\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_commit\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mcommit_message\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 1232\u001b[0;31m return self.git_push(\n\u001b[0m\u001b[1;32m 1233\u001b[0m \u001b[0mupstream\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;34mf\"origin {self.current_branch}\"\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1234\u001b[0m \u001b[0mblocking\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mblocking\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
140
- "\u001b[0;32m/usr/local/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py\u001b[0m in \u001b[0;36mgit_push\u001b[0;34m(self, upstream, blocking, auto_lfs_prune)\u001b[0m\n\u001b[1;32m 1002\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1003\u001b[0m \u001b[0;32mexcept\u001b[0m \u001b[0msubprocess\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mCalledProcessError\u001b[0m \u001b[0;32mas\u001b[0m \u001b[0mexc\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 1004\u001b[0;31m \u001b[0;32mraise\u001b[0m \u001b[0mEnvironmentError\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mexc\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mstderr\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m 1005\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1006\u001b[0m \u001b[0;32mif\u001b[0m \u001b[0;32mnot\u001b[0m \u001b[0mblocking\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
141
- "\u001b[0;31mOSError\u001b[0m: remote: ----------------------------------------------------------\u001b[33m \nremote: Your push was accepted, but with warnings: \nremote: - warning : empty or missing yaml metadata in card (dansbecker/hackernews_hiring_posts) \nremote: help: please find help at https://huggingface.co/docs/hub/model-repos\u001b[32m \nremote: ---------------------------------------------------------- \nremote: Please find the documentation at: \nremote: https://huggingface.co/docs/hub/model-repos\u001b(B\u001b[m \nremote: ---------------------------------------------------------- \nremote: -------------------------------------------------------------------------\u001b[31m \nremote: Your push was rejected because it contains files larger than 10M. \nremote: Please use https://git-lfs.github.com/ to store larger files.\u001b(B\u001b[m \nremote: ------------------------------------------------------------------------- \nremote: Offending files: \nremote: - raw/bq-results-20211206-133858-irtkgx60el7i.csv (ref: refs/heads/main) \nTo https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts\n ! [remote rejected] main -> main (pre-receive hook declined)\nerror: failed to push some refs to 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n"
142
- ]
143
- }
144
- ],
145
- "source": [
146
- "import datasets\n",
147
- "from huggingface_hub import create_repo\n",
148
- "from huggingface_hub import Repository\n",
149
- "\n",
150
- "all_datasets = datasets.dataset_dict.DatasetDict({'hiring': datasets.Dataset.from_pandas(hiring_df).remove_columns('__index_level_0__'),\n",
151
- " 'wants_to_be_hired': datasets.Dataset.from_pandas(wants_to_be_hired_df).remove_columns('__index_level_0__'),\n",
152
- " 'freelancer': datasets.Dataset.from_pandas(freelancer_df).remove_columns('__index_level_0__')})\n",
153
- "data_path = './data'\n",
154
- "all_datasets.save_to_disk(data_path)\n",
155
- "\n",
156
- "repo_url = 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n",
157
- "repo = Repository(local_dir=\".\", clone_from=repo_url)\n",
158
- "repo.git_add(data_path)\n",
159
- "repo.git_add('*.ipynb') # This file\n",
160
- "repo.git_add('README.md')\n",
161
- "repo.push_to_hub(commit_message='Regular update', blocking=True, clean_ok=False)"
162
- ]
163
- },
164
- {
165
- "cell_type": "code",
166
- "execution_count": null,
167
- "id": "960641f2-7844-44d7-a8bd-acc3c12ced07",
168
- "metadata": {},
169
- "outputs": [],
170
- "source": []
171
- }
172
- ],
173
- "metadata": {
174
- "kernelspec": {
175
- "display_name": "Python 3",
176
- "language": "python",
177
- "name": "python3"
178
- },
179
- "language_info": {
180
- "codemirror_mode": {
181
- "name": "ipython",
182
- "version": 3
183
- },
184
- "file_extension": ".py",
185
- "mimetype": "text/x-python",
186
- "name": "python",
187
- "nbconvert_exporter": "python",
188
- "pygments_lexer": "ipython3",
189
- "version": "3.8.8"
190
- }
191
- },
192
- "nbformat": 4,
193
- "nbformat_minor": 5
194
- }