A curated list of standards, tests and benchmarks that can be used for testing and evaluating dev tools
Give a good name to your file and add the following content (please note, that all fields are required):
[general] # benchmark name name = '' # select the most applicable category (from the [list](https://github.com/analysis-dev/awesome-benchmarks#supported-categories)) category = '' # mark your benchmark with custom tags (your custom OR [existing](https://github.com/analysis-dev/awesome-benchmarks#supported-tags) or add your own) tags = ['', ''] [info] # programming language language = '' # license used in the project (from the [list](https://github.com/analysis-dev/awesome-benchmarks#supported-licenses)) license = '' # number of tests/cases/scenarios/benchmarks scenarios_num = 0 # free text with a small desciption of the benchmark description = '' [links] # the link to the home page of the benchmark homepage = '' # the link to the source code sources = '' # the link to the page with documentation documentation = ''
Create a Pull Request
Validate your input with the following:
gradlew.bat build(WIN) or
./gradlew build(LINUX) (or wait until it will be validated automatically using GitHub Actions CI/CD).
We validate your input to make this list/library consistent, you won’t be able to merge without a validation.