Commit
·
5eac1be
1
Parent(s):
68844f7
Update parquet files
Browse files- .gitattributes +0 -27
- .gitignore +0 -2
- README.md +0 -17
- data/freelancer-00000-of-00001.parquet → dansbecker--hackernews_hiring_posts/parquet-freelancer.parquet +2 -2
- data/freelancer/dataset.arrow → dansbecker--hackernews_hiring_posts/parquet-hiring.parquet +2 -2
- data/wants_to_be_hired-00000-of-00001.parquet → dansbecker--hackernews_hiring_posts/parquet-wants_to_be_hired.parquet +2 -2
- data/dataset_dict.json +0 -1
- data/freelancer/dataset_info.json +0 -40
- data/freelancer/state.json +0 -14
- data/hiring-00000-of-00001.parquet +0 -3
- data/hiring/dataset.arrow +0 -3
- data/hiring/dataset_info.json +0 -40
- data/hiring/state.json +0 -14
- data/wants_to_be_hired/dataset.arrow +0 -3
- data/wants_to_be_hired/dataset_info.json +0 -40
- data/wants_to_be_hired/state.json +0 -14
- hn_hiring_posts.ipynb +0 -194
.gitattributes
DELETED
@@ -1,27 +0,0 @@
|
|
1 |
-
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
-
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
-
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
-
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
5 |
-
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
6 |
-
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
-
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
-
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
-
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
-
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
-
*.model filter=lfs diff=lfs merge=lfs -text
|
12 |
-
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
13 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
14 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
15 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
16 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
17 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
18 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
19 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
20 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
21 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
22 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
23 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
24 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
25 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
-
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.gitignore
DELETED
@@ -1,2 +0,0 @@
|
|
1 |
-
raw/*
|
2 |
-
.ipynb_checkpoints
|
|
|
|
|
|
README.md
DELETED
@@ -1,17 +0,0 @@
|
|
1 |
-
This dataset contains postings and comments from the following recurring threads on [Hacker News](http://news.ycombinator.com/)
|
2 |
-
|
3 |
-
1. Ask HN: Who is hiring?
|
4 |
-
2. Ask HN: Who wants to be hired?
|
5 |
-
3. Freelancer? Seeking freelancer?
|
6 |
-
|
7 |
-
These post types are stored in datasets called `hiring`, `wants_to_be_hired` and `freelancer` respectively.
|
8 |
-
|
9 |
-
Each type of posting has occurred on a regular basis for several years. You can identify when each comment/listing was added through the CommentTime field. The `ParentTitle` also indicates the date of the parent thread in text (e.g. `Ask HN: Who is hiring? (March 2021)`)
|
10 |
-
|
11 |
-
This dataset is not programmatically reproducible from source because it was uploaded as an experiment with HF datasets. The raw data was created by querying the public table `bigquery-public-data.hacker_news.full` in Google BigQuery.
|
12 |
-
|
13 |
-
Email addresses have been redacted from the dataset.
|
14 |
-
|
15 |
-
If this dataset is interesting/useful, I (Dan Becker) will look into improving reproducibility and other general clean-up.
|
16 |
-
|
17 |
-
This dataset may be useful for finding trends in tech and tech job listings.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/freelancer-00000-of-00001.parquet → dansbecker--hackernews_hiring_posts/parquet-freelancer.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e152e4fbc4eaf74db34b8dca12e11ff9c441d3b5fb23d145c33a25db59b9263b
|
3 |
+
size 5985114
|
data/freelancer/dataset.arrow → dansbecker--hackernews_hiring_posts/parquet-hiring.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:600614909754b4ba3b53bf9a5688445f376f660587c71dfffe83a1a81d484c4e
|
3 |
+
size 47128917
|
data/wants_to_be_hired-00000-of-00001.parquet → dansbecker--hackernews_hiring_posts/parquet-wants_to_be_hired.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95eaedc99c0b355f0560bb770eb5b6e05714e5dbc2f5a825163fde41cc49061c
|
3 |
+
size 5880973
|
data/dataset_dict.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"splits": ["hiring", "wants_to_be_hired", "freelancer"]}
|
|
|
|
data/freelancer/dataset_info.json
DELETED
@@ -1,40 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"builder_name": null,
|
3 |
-
"citation": "",
|
4 |
-
"config_name": null,
|
5 |
-
"dataset_size": null,
|
6 |
-
"description": "",
|
7 |
-
"download_checksums": null,
|
8 |
-
"download_size": null,
|
9 |
-
"features": {
|
10 |
-
"text": {
|
11 |
-
"dtype": "string",
|
12 |
-
"id": null,
|
13 |
-
"_type": "Value"
|
14 |
-
},
|
15 |
-
"CommentTime": {
|
16 |
-
"dtype": "string",
|
17 |
-
"id": null,
|
18 |
-
"_type": "Value"
|
19 |
-
},
|
20 |
-
"CommentAuthor": {
|
21 |
-
"dtype": "string",
|
22 |
-
"id": null,
|
23 |
-
"_type": "Value"
|
24 |
-
},
|
25 |
-
"ParentTitle": {
|
26 |
-
"dtype": "string",
|
27 |
-
"id": null,
|
28 |
-
"_type": "Value"
|
29 |
-
}
|
30 |
-
},
|
31 |
-
"homepage": "",
|
32 |
-
"license": "",
|
33 |
-
"post_processed": null,
|
34 |
-
"post_processing_size": null,
|
35 |
-
"size_in_bytes": null,
|
36 |
-
"splits": null,
|
37 |
-
"supervised_keys": null,
|
38 |
-
"task_templates": null,
|
39 |
-
"version": null
|
40 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/freelancer/state.json
DELETED
@@ -1,14 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_data_files": [
|
3 |
-
{
|
4 |
-
"filename": "dataset.arrow"
|
5 |
-
}
|
6 |
-
],
|
7 |
-
"_fingerprint": "4f8f711b9136404c",
|
8 |
-
"_format_columns": null,
|
9 |
-
"_format_kwargs": {},
|
10 |
-
"_format_type": null,
|
11 |
-
"_indexes": {},
|
12 |
-
"_output_all_columns": false,
|
13 |
-
"_split": null
|
14 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/hiring-00000-of-00001.parquet
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:f73ecadb195114287b162969f8db2ef0c9509e76dec4e3aa401319d6f09aa14b
|
3 |
-
size 46640019
|
|
|
|
|
|
|
|
data/hiring/dataset.arrow
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:a8827da89590c888c1e2d167c04f4fb527ed342c9cc23c61bbfa6646cfcf6f2f
|
3 |
-
size 81975968
|
|
|
|
|
|
|
|
data/hiring/dataset_info.json
DELETED
@@ -1,40 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"builder_name": null,
|
3 |
-
"citation": "",
|
4 |
-
"config_name": null,
|
5 |
-
"dataset_size": null,
|
6 |
-
"description": "",
|
7 |
-
"download_checksums": null,
|
8 |
-
"download_size": null,
|
9 |
-
"features": {
|
10 |
-
"text": {
|
11 |
-
"dtype": "string",
|
12 |
-
"id": null,
|
13 |
-
"_type": "Value"
|
14 |
-
},
|
15 |
-
"CommentTime": {
|
16 |
-
"dtype": "string",
|
17 |
-
"id": null,
|
18 |
-
"_type": "Value"
|
19 |
-
},
|
20 |
-
"CommentAuthor": {
|
21 |
-
"dtype": "string",
|
22 |
-
"id": null,
|
23 |
-
"_type": "Value"
|
24 |
-
},
|
25 |
-
"ParentTitle": {
|
26 |
-
"dtype": "string",
|
27 |
-
"id": null,
|
28 |
-
"_type": "Value"
|
29 |
-
}
|
30 |
-
},
|
31 |
-
"homepage": "",
|
32 |
-
"license": "",
|
33 |
-
"post_processed": null,
|
34 |
-
"post_processing_size": null,
|
35 |
-
"size_in_bytes": null,
|
36 |
-
"splits": null,
|
37 |
-
"supervised_keys": null,
|
38 |
-
"task_templates": null,
|
39 |
-
"version": null
|
40 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/hiring/state.json
DELETED
@@ -1,14 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_data_files": [
|
3 |
-
{
|
4 |
-
"filename": "dataset.arrow"
|
5 |
-
}
|
6 |
-
],
|
7 |
-
"_fingerprint": "e5166a077fc0f8e1",
|
8 |
-
"_format_columns": null,
|
9 |
-
"_format_kwargs": {},
|
10 |
-
"_format_type": null,
|
11 |
-
"_indexes": {},
|
12 |
-
"_output_all_columns": false,
|
13 |
-
"_split": null
|
14 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/wants_to_be_hired/dataset.arrow
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:cc638e6f45b651e6e166298e20f7c68b2fb6cdb5aac357a728778dffa9114ac1
|
3 |
-
size 11373152
|
|
|
|
|
|
|
|
data/wants_to_be_hired/dataset_info.json
DELETED
@@ -1,40 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"builder_name": null,
|
3 |
-
"citation": "",
|
4 |
-
"config_name": null,
|
5 |
-
"dataset_size": null,
|
6 |
-
"description": "",
|
7 |
-
"download_checksums": null,
|
8 |
-
"download_size": null,
|
9 |
-
"features": {
|
10 |
-
"text": {
|
11 |
-
"dtype": "string",
|
12 |
-
"id": null,
|
13 |
-
"_type": "Value"
|
14 |
-
},
|
15 |
-
"CommentTime": {
|
16 |
-
"dtype": "string",
|
17 |
-
"id": null,
|
18 |
-
"_type": "Value"
|
19 |
-
},
|
20 |
-
"CommentAuthor": {
|
21 |
-
"dtype": "string",
|
22 |
-
"id": null,
|
23 |
-
"_type": "Value"
|
24 |
-
},
|
25 |
-
"ParentTitle": {
|
26 |
-
"dtype": "string",
|
27 |
-
"id": null,
|
28 |
-
"_type": "Value"
|
29 |
-
}
|
30 |
-
},
|
31 |
-
"homepage": "",
|
32 |
-
"license": "",
|
33 |
-
"post_processed": null,
|
34 |
-
"post_processing_size": null,
|
35 |
-
"size_in_bytes": null,
|
36 |
-
"splits": null,
|
37 |
-
"supervised_keys": null,
|
38 |
-
"task_templates": null,
|
39 |
-
"version": null
|
40 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/wants_to_be_hired/state.json
DELETED
@@ -1,14 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_data_files": [
|
3 |
-
{
|
4 |
-
"filename": "dataset.arrow"
|
5 |
-
}
|
6 |
-
],
|
7 |
-
"_fingerprint": "f8da56cf67a872ff",
|
8 |
-
"_format_columns": null,
|
9 |
-
"_format_kwargs": {},
|
10 |
-
"_format_type": null,
|
11 |
-
"_indexes": {},
|
12 |
-
"_output_all_columns": false,
|
13 |
-
"_split": null
|
14 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
hn_hiring_posts.ipynb
DELETED
@@ -1,194 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"cells": [
|
3 |
-
{
|
4 |
-
"cell_type": "code",
|
5 |
-
"execution_count": 1,
|
6 |
-
"id": "056aa255-fda1-4cde-be24-459f6ad2c8b9",
|
7 |
-
"metadata": {},
|
8 |
-
"outputs": [
|
9 |
-
{
|
10 |
-
"name": "stderr",
|
11 |
-
"output_type": "stream",
|
12 |
-
"text": [
|
13 |
-
"/usr/local/anaconda3/lib/python3.8/site-packages/IPython/core/interactiveshell.py:3165: DtypeWarning: Columns (4) have mixed types.Specify dtype option on import or set low_memory=False.\n",
|
14 |
-
" has_raised = await self.run_ast_nodes(code_ast.body, cell_name,\n"
|
15 |
-
]
|
16 |
-
}
|
17 |
-
],
|
18 |
-
"source": [
|
19 |
-
"import pandas as pd\n",
|
20 |
-
"\n",
|
21 |
-
"df = pd.read_csv('raw/bq-results-20211206-133858-irtkgx60el7i.csv').drop(['ParentUrl', 'ParentAuthor', 'ParentTime', 'ParentScore'], axis=1)\n",
|
22 |
-
"\n",
|
23 |
-
"# There's a mix of HTML, sanitized links and raw Unicode that we'd like to clean up\n",
|
24 |
-
"df.text = df.text.str.replace('<p>', '\\n')\n",
|
25 |
-
"strings_to_remove = ['rel=\"nofollow\"', '<pre>', '</pre>', '<i>', '</i>', '<code>', '</code>', '>']\n",
|
26 |
-
"email_regex = '[a-zA-Z0-9._-]{0,30}@[a-zA-Z0-9._-]{0,20}\\.[a-zA-Z0-9_-]{2,3}' # for redacting emails\n",
|
27 |
-
"munged_url_regex = 'http(s)?:\\&\\#.*?\\<\\/a>'\n",
|
28 |
-
"\n",
|
29 |
-
"for string in strings_to_remove:\n",
|
30 |
-
" df.text = df.text.str.replace(string, '')\n",
|
31 |
-
"\n",
|
32 |
-
"\n",
|
33 |
-
"df.text = df.text.replace(email_regex, 'REDACTED_EMAIL', regex=True)\n",
|
34 |
-
"df.text = df.text.replace(munged_url_regex, '', regex=True)\n",
|
35 |
-
"\n",
|
36 |
-
" # fix some unicode issues\n",
|
37 |
-
"df.text = df.text.str.replace(''', \"'\")\n",
|
38 |
-
"df.text = df.text.str.replace('/', \"/\")\n",
|
39 |
-
"df.text = df.text.str.replace(\""\", '\"')\n",
|
40 |
-
"\n",
|
41 |
-
"hiring_df = df[df.ParentTitle.str.lower().str.contains('who is hiring')]\n",
|
42 |
-
"wants_to_be_hired_df = df[df.ParentTitle.str.lower().str.contains('wants to be hired')]\n",
|
43 |
-
"freelancer_df = df[df.ParentTitle.str.lower().str.contains('freelancer')]\n"
|
44 |
-
]
|
45 |
-
},
|
46 |
-
{
|
47 |
-
"cell_type": "code",
|
48 |
-
"execution_count": 2,
|
49 |
-
"id": "57ca7c96-d94c-4e65-8113-da7729558247",
|
50 |
-
"metadata": {},
|
51 |
-
"outputs": [
|
52 |
-
{
|
53 |
-
"name": "stderr",
|
54 |
-
"output_type": "stream",
|
55 |
-
"text": [
|
56 |
-
"/Users/dan.becker/Desktop/hackernews_hiring_dataset/. is already a clone of https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts. Make sure you pull the latest changes with `repo.git_pull()`.\n",
|
57 |
-
"Several commits (5) will be pushed upstream.\n",
|
58 |
-
"The progress bars may be unreliable.\n"
|
59 |
-
]
|
60 |
-
},
|
61 |
-
{
|
62 |
-
"data": {
|
63 |
-
"application/vnd.jupyter.widget-view+json": {
|
64 |
-
"model_id": "63ddc30f9ba5420faf6e34f6fa7144fc",
|
65 |
-
"version_major": 2,
|
66 |
-
"version_minor": 0
|
67 |
-
},
|
68 |
-
"text/plain": [
|
69 |
-
"Upload file data/hiring/dataset.arrow: 0%| | 32.0k/78.2M [00:00<?, ?B/s]"
|
70 |
-
]
|
71 |
-
},
|
72 |
-
"metadata": {},
|
73 |
-
"output_type": "display_data"
|
74 |
-
},
|
75 |
-
{
|
76 |
-
"data": {
|
77 |
-
"application/vnd.jupyter.widget-view+json": {
|
78 |
-
"model_id": "be96819a155b4f17933d3ff82a2fb70d",
|
79 |
-
"version_major": 2,
|
80 |
-
"version_minor": 0
|
81 |
-
},
|
82 |
-
"text/plain": [
|
83 |
-
"Upload file data/wants_to_be_hired/dataset.arrow: 0%| | 32.0k/10.8M [00:00<?, ?B/s]"
|
84 |
-
]
|
85 |
-
},
|
86 |
-
"metadata": {},
|
87 |
-
"output_type": "display_data"
|
88 |
-
},
|
89 |
-
{
|
90 |
-
"data": {
|
91 |
-
"application/vnd.jupyter.widget-view+json": {
|
92 |
-
"model_id": "3ef183014a04462e97153ca6673c5f11",
|
93 |
-
"version_major": 2,
|
94 |
-
"version_minor": 0
|
95 |
-
},
|
96 |
-
"text/plain": [
|
97 |
-
"Upload file data/freelancer/dataset.arrow: 0%| | 32.0k/10.6M [00:00<?, ?B/s]"
|
98 |
-
]
|
99 |
-
},
|
100 |
-
"metadata": {},
|
101 |
-
"output_type": "display_data"
|
102 |
-
},
|
103 |
-
{
|
104 |
-
"name": "stderr",
|
105 |
-
"output_type": "stream",
|
106 |
-
"text": [
|
107 |
-
"remote: ----------------------------------------------------------\u001b[33m \n",
|
108 |
-
"remote: Your push was accepted, but with warnings: \n",
|
109 |
-
"remote: - warning : empty or missing yaml metadata in card (dansbecker/hackernews_hiring_posts) \n",
|
110 |
-
"remote: help: please find help at https://huggingface.co/docs/hub/model-repos\u001b[32m \n",
|
111 |
-
"remote: ---------------------------------------------------------- \n",
|
112 |
-
"remote: Please find the documentation at: \n",
|
113 |
-
"remote: https://huggingface.co/docs/hub/model-repos\u001b(B\u001b[m \n",
|
114 |
-
"remote: ---------------------------------------------------------- \n",
|
115 |
-
"remote: -------------------------------------------------------------------------\u001b[31m \n",
|
116 |
-
"remote: Your push was rejected because it contains files larger than 10M. \n",
|
117 |
-
"remote: Please use https://git-lfs.github.com/ to store larger files.\u001b(B\u001b[m \n",
|
118 |
-
"remote: ------------------------------------------------------------------------- \n",
|
119 |
-
"remote: Offending files: \n",
|
120 |
-
"remote: - raw/bq-results-20211206-133858-irtkgx60el7i.csv (ref: refs/heads/main) \n",
|
121 |
-
"To https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts\n",
|
122 |
-
" ! [remote rejected] main -> main (pre-receive hook declined)\n",
|
123 |
-
"error: failed to push some refs to 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n",
|
124 |
-
"\n"
|
125 |
-
]
|
126 |
-
},
|
127 |
-
{
|
128 |
-
"ename": "OSError",
|
129 |
-
"evalue": "remote: ----------------------------------------------------------\u001b[33m \nremote: Your push was accepted, but with warnings: \nremote: - warning : empty or missing yaml metadata in card (dansbecker/hackernews_hiring_posts) \nremote: help: please find help at https://huggingface.co/docs/hub/model-repos\u001b[32m \nremote: ---------------------------------------------------------- \nremote: Please find the documentation at: \nremote: https://huggingface.co/docs/hub/model-repos\u001b(B\u001b[m \nremote: ---------------------------------------------------------- \nremote: -------------------------------------------------------------------------\u001b[31m \nremote: Your push was rejected because it contains files larger than 10M. \nremote: Please use https://git-lfs.github.com/ to store larger files.\u001b(B\u001b[m \nremote: ------------------------------------------------------------------------- \nremote: Offending files: \nremote: - raw/bq-results-20211206-133858-irtkgx60el7i.csv (ref: refs/heads/main) \nTo https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts\n ! [remote rejected] main -> main (pre-receive hook declined)\nerror: failed to push some refs to 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n",
|
130 |
-
"output_type": "error",
|
131 |
-
"traceback": [
|
132 |
-
"\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
|
133 |
-
"\u001b[0;31mCalledProcessError\u001b[0m Traceback (most recent call last)",
|
134 |
-
"\u001b[0;32m/usr/local/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py\u001b[0m in \u001b[0;36mgit_push\u001b[0;34m(self, upstream, blocking, auto_lfs_prune)\u001b[0m\n\u001b[1;32m 998\u001b[0m \u001b[0;32mif\u001b[0m \u001b[0mreturn_code\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m--> 999\u001b[0;31m raise subprocess.CalledProcessError(\n\u001b[0m\u001b[1;32m 1000\u001b[0m \u001b[0mreturn_code\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mprocess\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0margs\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0moutput\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mstdout\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mstderr\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mstderr\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
|
135 |
-
"\u001b[0;31mCalledProcessError\u001b[0m: Command '['git', 'push', '--set-upstream', 'origin', 'main']' returned non-zero exit status 1.",
|
136 |
-
"\nDuring handling of the above exception, another exception occurred:\n",
|
137 |
-
"\u001b[0;31mOSError\u001b[0m Traceback (most recent call last)",
|
138 |
-
"\u001b[0;32m<ipython-input-2-38e9608c0524>\u001b[0m in \u001b[0;36m<module>\u001b[0;34m\u001b[0m\n\u001b[1;32m 14\u001b[0m \u001b[0mrepo\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'*.ipynb'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 15\u001b[0m \u001b[0mrepo\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0;34m'README.md'\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m---> 16\u001b[0;31m \u001b[0mrepo\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mpush_to_hub\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mcommit_message\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;34m'Regular update'\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mblocking\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;32mTrue\u001b[0m\u001b[0;34m,\u001b[0m \u001b[0mclean_ok\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;32mFalse\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m",
|
139 |
-
"\u001b[0;32m/usr/local/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py\u001b[0m in \u001b[0;36mpush_to_hub\u001b[0;34m(self, commit_message, blocking, clean_ok, auto_lfs_prune)\u001b[0m\n\u001b[1;32m 1230\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_add\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mauto_lfs_track\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;32mTrue\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1231\u001b[0m \u001b[0mself\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mgit_commit\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mcommit_message\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 1232\u001b[0;31m return self.git_push(\n\u001b[0m\u001b[1;32m 1233\u001b[0m \u001b[0mupstream\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0;34mf\"origin {self.current_branch}\"\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1234\u001b[0m \u001b[0mblocking\u001b[0m\u001b[0;34m=\u001b[0m\u001b[0mblocking\u001b[0m\u001b[0;34m,\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
|
140 |
-
"\u001b[0;32m/usr/local/anaconda3/lib/python3.8/site-packages/huggingface_hub/repository.py\u001b[0m in \u001b[0;36mgit_push\u001b[0;34m(self, upstream, blocking, auto_lfs_prune)\u001b[0m\n\u001b[1;32m 1002\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1003\u001b[0m \u001b[0;32mexcept\u001b[0m \u001b[0msubprocess\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mCalledProcessError\u001b[0m \u001b[0;32mas\u001b[0m \u001b[0mexc\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0;32m-> 1004\u001b[0;31m \u001b[0;32mraise\u001b[0m \u001b[0mEnvironmentError\u001b[0m\u001b[0;34m(\u001b[0m\u001b[0mexc\u001b[0m\u001b[0;34m.\u001b[0m\u001b[0mstderr\u001b[0m\u001b[0;34m)\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n\u001b[0m\u001b[1;32m 1005\u001b[0m \u001b[0;34m\u001b[0m\u001b[0m\n\u001b[1;32m 1006\u001b[0m \u001b[0;32mif\u001b[0m \u001b[0;32mnot\u001b[0m \u001b[0mblocking\u001b[0m\u001b[0;34m:\u001b[0m\u001b[0;34m\u001b[0m\u001b[0;34m\u001b[0m\u001b[0m\n",
|
141 |
-
"\u001b[0;31mOSError\u001b[0m: remote: ----------------------------------------------------------\u001b[33m \nremote: Your push was accepted, but with warnings: \nremote: - warning : empty or missing yaml metadata in card (dansbecker/hackernews_hiring_posts) \nremote: help: please find help at https://huggingface.co/docs/hub/model-repos\u001b[32m \nremote: ---------------------------------------------------------- \nremote: Please find the documentation at: \nremote: https://huggingface.co/docs/hub/model-repos\u001b(B\u001b[m \nremote: ---------------------------------------------------------- \nremote: -------------------------------------------------------------------------\u001b[31m \nremote: Your push was rejected because it contains files larger than 10M. \nremote: Please use https://git-lfs.github.com/ to store larger files.\u001b(B\u001b[m \nremote: ------------------------------------------------------------------------- \nremote: Offending files: \nremote: - raw/bq-results-20211206-133858-irtkgx60el7i.csv (ref: refs/heads/main) \nTo https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts\n ! [remote rejected] main -> main (pre-receive hook declined)\nerror: failed to push some refs to 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n"
|
142 |
-
]
|
143 |
-
}
|
144 |
-
],
|
145 |
-
"source": [
|
146 |
-
"import datasets\n",
|
147 |
-
"from huggingface_hub import create_repo\n",
|
148 |
-
"from huggingface_hub import Repository\n",
|
149 |
-
"\n",
|
150 |
-
"all_datasets = datasets.dataset_dict.DatasetDict({'hiring': datasets.Dataset.from_pandas(hiring_df).remove_columns('__index_level_0__'),\n",
|
151 |
-
" 'wants_to_be_hired': datasets.Dataset.from_pandas(wants_to_be_hired_df).remove_columns('__index_level_0__'),\n",
|
152 |
-
" 'freelancer': datasets.Dataset.from_pandas(freelancer_df).remove_columns('__index_level_0__')})\n",
|
153 |
-
"data_path = './data'\n",
|
154 |
-
"all_datasets.save_to_disk(data_path)\n",
|
155 |
-
"\n",
|
156 |
-
"repo_url = 'https://huggingface.co/datasets/dansbecker/hackernews_hiring_posts'\n",
|
157 |
-
"repo = Repository(local_dir=\".\", clone_from=repo_url)\n",
|
158 |
-
"repo.git_add(data_path)\n",
|
159 |
-
"repo.git_add('*.ipynb') # This file\n",
|
160 |
-
"repo.git_add('README.md')\n",
|
161 |
-
"repo.push_to_hub(commit_message='Regular update', blocking=True, clean_ok=False)"
|
162 |
-
]
|
163 |
-
},
|
164 |
-
{
|
165 |
-
"cell_type": "code",
|
166 |
-
"execution_count": null,
|
167 |
-
"id": "960641f2-7844-44d7-a8bd-acc3c12ced07",
|
168 |
-
"metadata": {},
|
169 |
-
"outputs": [],
|
170 |
-
"source": []
|
171 |
-
}
|
172 |
-
],
|
173 |
-
"metadata": {
|
174 |
-
"kernelspec": {
|
175 |
-
"display_name": "Python 3",
|
176 |
-
"language": "python",
|
177 |
-
"name": "python3"
|
178 |
-
},
|
179 |
-
"language_info": {
|
180 |
-
"codemirror_mode": {
|
181 |
-
"name": "ipython",
|
182 |
-
"version": 3
|
183 |
-
},
|
184 |
-
"file_extension": ".py",
|
185 |
-
"mimetype": "text/x-python",
|
186 |
-
"name": "python",
|
187 |
-
"nbconvert_exporter": "python",
|
188 |
-
"pygments_lexer": "ipython3",
|
189 |
-
"version": "3.8.8"
|
190 |
-
}
|
191 |
-
},
|
192 |
-
"nbformat": 4,
|
193 |
-
"nbformat_minor": 5
|
194 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|