Datasets:

Modalities:
Text
Formats:
json
Languages:
code
fst
ArXiv:
Libraries:
Datasets
Dask
License:
File size: 6,837 Bytes
fabbafc
 
 
bdb3c0a
 
59bc9be
fabbafc
c533d9c
7727961
 
 
fabbafc
 
7727961
 
 
fabbafc
 
7727961
 
 
e8b9370
76eaabf
 
7727961
e8b9370
624a5c1
7727961
 
76eaabf
 
 
 
 
 
 
 
e8b9370
76eaabf
 
 
e8b9370
 
 
76eaabf
 
 
e8b9370
76eaabf
 
 
e8b9370
76eaabf
 
 
 
 
e8b9370
76eaabf
 
7727961
 
624a5c1
d37827f
624a5c1
 
d37827f
624a5c1
1b3a35d
624a5c1
 
1b3a35d
 
624a5c1
7727961
 
d37827f
 
 
 
7727961
 
d37827f
 
 
7727961
0a14886
 
 
 
 
 
e8b9370
d5c4660
e8b9370
d37827f
 
 
e8b9370
 
d37827f
 
 
 
 
 
 
 
 
 
 
 
ddfb487
 
 
 
 
 
 
 
d5c4660
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
---
license: cdla-permissive-2.0
task_categories:
- text-generation
- text2text-generation
- "other"
tags:
- code
- fstar
- popai
pretty_name: PoPAI-FStarDataSet
size_categories:
- 10K<n<100K
language:
- code
- fst
---

# Proof Oriented Programming with AI (PoPAI) - FStarDataSet

This dataset contains programs and proofs in [F* proof-oriented programming language](https://fstar-lang.org/). 
The data, proposed in [Towards Neural Synthesis for SMT-Assisted Proof-Oriented Programming](https://arxiv.org/pdf/2405.01787),
is an archive of source code, build artifacts, and metadata assembled from eight different F⋆-based open source projects on GitHub.

## Primary-Objective
This dataset's primary objective is to train and evaluate Proof-oriented Programming with AI (PoPAI, in short). Given a specification of a program and proof in F*, 
the objective of a AI model is to synthesize the implemantation (see [below](#usage) for details about the usage of this dataset, including the input and output). 

## Data Format
Each of the examples in this dataset are organized as dictionaries with the following schema
```json
{
    "file_name": <str: Name of the file>,
    "name": <str: name of the example, can be used to uniquely identify the example>,
    "original_source_type": <str: actual source type, to be used for type checking>,
    "source_type": <str: modified source type, to be used to formulate prompt>,
    "source_definition": <str: target definition>,
    "source": <dict: contains metadata about the source of this example, including project_name, git url, git sha, etc.>,
    "source_range": <dict: metadata containing start and end lines and columns of this definition in the source file>,
    "file_context": <str: extracted file context upto the point of current definition>, 
    "dependencies": <dict: build dependencies for this file>,
    "opens_and_abbrevs": <list[dict]: List of opened modules and abbreviated modules in the file, necessary for evaluation.>,
    "vconfig": <dict: SMT solver flags for this definition>,
    "interleaved": <bool: whether this definition is interleaved from the interface file>,
    "verbose_type": <str: the verbose type of this definition as resolved by the type checker>,
    "effect": <str: effect>,
    "effect_flags": <list[str]: any effect flags>,
    "mutual_with": <list: if this definition is mutually recursive with another, list of those names>,
    "ideal_premises": <list[str]: Other definitions that are used in the ground truth definition>,
    "proof_features": <list[str]>,
    "is_simple_lemma": <bool/null>,
    "is_div": <bool: if this definition has the divergent effect>,
    "is_proof": <bool>,
    "is_simply_typed": <bool>,
    "is_type": <bool/null>,
    "partial_definition": <str>,
    "completed_definiton": <str>,
    "isa_cross_project_example": <bool: if this example belongs to the cross-project evaluation set>
}
```

# Usage
To use this dataset with [`datasets`](https://pypi.org/project/datasets/), 
```python
from datasets import load_dataset

data = load_dataset("microsoft/FStarDataSet")
train_data = data["train"]
eval_data = data["validation"]
test_data = data["test"]

intra_project_test = test_data.filter(lambda x: x["isa_cross_project_example"] == False)
cross_project_test = test_data.filter(lambda x: x["isa_cross_project_example"] == True)
```

## Input
The primary input for generating F* definition is **`source_type`**. 
All other information in an example may be used directly or to derive an input except 
**`source_definition`**, **`ideal_premises`**, and **`completed_definiton`**.


## Output
The primary output is **`source_definition`**, which is the ground truth definition, that can be evaluated with the [proof checker](#evaluation-on-this-dataset). 
The **`completed_definiton`** may be used as ground truth when a model is used as a text completion setting (though the evaluator does not support evaluation in this setting). 
In addition, **`ideal_premises`** may be used for evaluating premise selection models. 

# Evaluation on this dataset
Generated F* definitions should be evaluated the proof checker tool from [https://github.com/FStarLang/fstar_dataset/releases/tag/eval-v1.0](https://github.com/FStarLang/fstar_dataset/releases/tag/eval-v1.0). 
Download the source code and the `helpers.zip` file from the release. 
  
## Troubleshooting
The attached binaries in the evaluator (i.e., `fstar.exe` and `z3`) are built on 
**`Ubuntu 20.04.6 LTS (GNU/Linux 5.4.0-189-generic x86_64)`**,  **`gcc (Ubuntu 9.4.0-1ubuntu1~20.04.2)`**,  **`OCaml 4.12.0`**. 
If any of the binaries do not work properly, build F* from [commit: f3b4db2ebce90020acbbbe1b4ea0d05d3e69ad6c](https://github.com/FStarLang/FStar/commit/f3b4db2ebce90020acbbbe1b4ea0d05d3e69ad6c) 
from the [F* repository](https://github.com/FStarLang/FStar), using the [installation guide](https://github.com/FStarLang/FStar/blob/master/INSTALL.md).


# Data Source
The raw data in this project are collected from eight open-source F* repositories on GitHib 
1. [FStar](https://github.com/FStarLang/FStar): The F⋆ compiler itself, including its standard library and examples.
2. [Karamel](https://github.com/FStarLang/karamel): A transpiler from a subset of F⋆ called Low* to C, including libraries to work with a model of C types and control structures, e.g., for- and while-loops.
3. [EverParse](https://github.com/project-everest/everparse): A parser generator for binary formats, used in various large scale systems, e.g., the Windows kernel.
4. [HACL*](https://github.com/hacl-star/hacl-star): A library of verified cryptographic algorithms, including ValeCrypt, a library of verified assembly code, as well as EverCrypt, a cryptographic provider, including code deployed in Linux, Firefox, and Python.
5. [Merkle-tree](https://github.com/hacl-star/merkle-tree): A verified, incremental Merkle tree, designed for use in Azure CCF, a confidential computing system.
6. [Steel](https://github.com/FStarLang/steel): A concurrent separation logic library, with proofs of data structures and concurrency primitives.
7. [miTLS-F*](https://github.com/project-everest/mitls-fstar): A partially verified reference implementation of the TLS protocol.
8. [EverQuic-Crypto](https://github.com/project-everest/everquic-crypto): A verified implementation of header and packet protection for the QUIC protocol.

# Limitations
**TDB**

# Citation
```
@inproceedings{chakraborty2024towards,
  title={Towards Neural Synthesis for SMT-Assisted Proof-Oriented Programming},
  author={Chakraborty, Saikat and Ebner, Gabriel and Bhat, Siddharth and Fakhoury, Sarah and Fatima, Sakina and Lahiri, Shuvendu and Swamy, Nikhil},
  booktitle={Proceedings of the IEEE/ACM 47th International Conference on Software Engineering (To Appear)},
  pages={1--12},
  year={2025}
}
```