Datasets:

ArXiv:
License:
File size: 1,799 Bytes
94c6c5d
 
 
e35f382
 
 
 
 
 
 
50cc9bd
 
e35f382
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cd00be0
e35f382
 
a77e617
e35f382
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
---
license: cc-by-4.0
---

# VISEM-Tracking-graphs - HuggingFace Repository

This HuggingFace repository contains the pre-generated graphs for the sperm video dataset called VISEM-Tracking (https://huggingface.co/papers/2212.02842) . The graphs represent spatial and temporal relationships between sperm in a video. Spatial edges connect sperms within the same frame, while temporal edges connect sperms across different frames.

The graphs have been generated with varying spatial threshold values: 0.1, 0.2, 0.3, 0.4, and 0.5. Each spatial threshold determines the maximum distance between two nodes for them to be connected in the graph. The repository contains separate directories for each spatial threshold.

The source code used to generate graphs can be found here: https://github.com/vlbthambawita/visem-tracking-graphs 

## Repository Structure

The repository is structured as follows:

- `spatial_threshold_0.1`
- `spatial_threshold_0.2`
- `spatial_threshold_0.3`
- `spatial_threshold_0.4`
- `spatial_threshold_0.5`

Inside each `spatial_threshold_X` directory, you will find:

- `frame_graphs`: A directory containing individual frame graphs as GraphML files.
- `video_graph.graphml`: A GraphML file containing the complete video graph.

## Usage

To use the graphs in this repository, you need to:

1. Download the desired graph files (frame graphs or video graph) for the spatial threshold of your choice.
2. Load the graphs using a graph library such as NetworkX in Python:

```python
import networkx as nx

# Load a frame graph
frame_graph = nx.read_graphml('path/to/frame_graph_X.graphml')

# Load the video graph
video_graph = nx.read_graphml('path/to/video_graph.graphml')
```

TO USE THIS DATA, you need to cite the paper:
https://www.nature.com/articles/s41597-023-02173-4