--- license: cc-by-4.0 pretty_name: Ground-based 2d images assembled in Maireles-González et al. tags: - astronomy - compression - images dataset_info: - config_name: full features: - name: image dtype: image: mode: I;16 - name: telescope dtype: string - name: image_id dtype: string splits: - name: train num_bytes: 3509045373 num_examples: 120 - name: test num_bytes: 970120060 num_examples: 32 download_size: 2240199274 dataset_size: 4479165433 - config_name: tiny features: - name: image dtype: image: mode: I;16 - name: telescope dtype: string - name: image_id dtype: string splits: - name: train num_bytes: 307620695 num_examples: 10 - name: test num_bytes: 168984698 num_examples: 5 download_size: 238361934 dataset_size: 476605393 --- # GBI-16-2D-Legacy Dataset GBI-16-2D-Legacy is a Huggingface `dataset` wrapper around a compression dataset assembled by Maireles-González et al. (Publications of the Astronomical Society of the Pacific, 135:094502, 2023 September; doi: [https://doi.org/10.1088/1538-3873/acf6e0](https://doi.org/10.1088/1538-3873/acf6e0)). It contains 226 FITS images from 5 different ground-based telescope/cameras with a varying amount of entropy per image. # Usage You first need to install the `datasets` and `astropy` packages: ```bash pip install datasets astropy PIL ``` There are two datasets: `tiny` and `full`, each with `train` and `test` splits. The `tiny` dataset has 5 2D images in the `train` and 1 in the `test`. The `full` dataset contains all the images in the `data/` directory. ## Local Use (RECOMMENDED) You can clone this repo and use directly without connecting to hf: ```bash git clone https://huggingface.co/datasets/AnonAstroData/GBI-16-2D-Legacy ``` ```bash git lfs pull ``` Then `cd GBI-16-2D-Legacy` and start python like: ```python from datasets import load_dataset dataset = load_dataset("./GBI-16-2D-Legacy.py", "tiny", data_dir="./data/", writer_batch_size=1, trust_remote_code=True) ds = dataset.with_format("np") ``` Now you should be able to use the `ds` variable like: ```python ds["test"][0]["image"].shape # -> (4200, 2154) ``` Note of course that it will take a long time to download and convert the images in the local cache for the `full` dataset. Afterward, the usage should be quick as the files are memory-mapped from disk. If you run into issues with downloading the `full` dataset, try changing `num_proc` in `load_dataset` to >1 (e.g. 5). You can also set the `writer_batch_size` to ~10-20. ## Use from Huggingface Directly This method may only be an option when trying to access the "tiny" version of the dataset. To directly use from this data from Huggingface, you'll want to log in on the command line before starting python: ```bash huggingface-cli login ``` or ``` import huggingface_hub huggingface_hub.login(token=token) ``` Then in your python script: ```python from datasets import load_dataset dataset = load_dataset("AstroCompress/GBI-16-2D-Legacy", "tiny", writer_batch_size=1, trust_remote_code=True) ds = dataset.with_format("np") ``` ## Demo Colab Notebook We provide a demo collab notebook to get started on using the dataset [here](https://colab.research.google.com/drive/1wcz7qMqSAMST2kXFlL-TbwpYR26gYIYy?usp=sharing). ## Utils scripts Note that utils scripts such as `eval_baselines.py` must be run from the parent directory of `utils`, i.e. `python utils/eval_baselines.py`.