Commit
Β·
84fadb9
1
Parent(s):
a5e37f8
Add description for the library-based code generation task
Browse files- src/tasks_content.py +16 -7
src/tasks_content.py
CHANGED
|
@@ -11,13 +11,22 @@ TASKS_PRETTY = {
|
|
| 11 |
TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
|
| 12 |
|
| 13 |
TASKS_DESCRIPTIONS = {
|
| 14 |
-
"library_based_code_generation": "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
"ci_builds_repair": "cool description for Bug Localization on Build Logs task",
|
| 17 |
|
| 18 |
"project_code_completion": """# Project-Level Code Completion\n
|
| 19 |
|
| 20 |
-
Our Project-Level Code Completion π€ [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
|
| 21 |
* `small-context`: 144 data points,
|
| 22 |
* `medium-context`: 224 data points,
|
| 23 |
* `large-context`: 270 data points,
|
|
@@ -34,7 +43,7 @@ TASKS_DESCRIPTIONS = {
|
|
| 34 |
|
| 35 |
For further details on the dataset and the baselines from ποΈ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
|
| 36 |
""",
|
| 37 |
-
|
| 38 |
"commit_message_generation": """# Commit Message Generation\n
|
| 39 |
|
| 40 |
Our Commit Message Generation benchmark π€ [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
|
|
@@ -49,15 +58,15 @@ TASKS_DESCRIPTIONS = {
|
|
| 49 |
|
| 50 |
**Note.** The leaderboard is sorted by ROUGE-1 metric by default.
|
| 51 |
""",
|
| 52 |
-
|
| 53 |
"bug_localization": """# Bug Localization\n
|
| 54 |
|
| 55 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
|
| 56 |
|
| 57 |
Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
|
| 58 |
-
We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k
|
| 59 |
""",
|
| 60 |
-
|
| 61 |
"module_summarization": """# Module Summarization\n
|
| 62 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
|
| 63 |
|
|
@@ -77,5 +86,5 @@ def get_submission_text_files_for_task(task_pretty: Optional[str]) -> str:
|
|
| 77 |
|
| 78 |
if task_id == "commit_message_generation":
|
| 79 |
return f"""**{task_pretty} Instructions:**\n\n* Please, attach files in [JSONLines format](https://jsonlines.org/). For an example, check the predictions provided by ποΈ Long Code Arena Team in π€ [JetBrains-Research/lca-results](https://huggingface.co/datasets/JetBrains-Research/lca-results/tree/main/commit_message_generation/predictions). Make sure to include `"prediction"` and `"reference"` fields for each example, the rest are optional."""
|
| 80 |
-
|
| 81 |
return f"**{task_pretty} Instructions:**\n\n* π§ There are no instructions for the current task yet."
|
|
|
|
| 11 |
TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
|
| 12 |
|
| 13 |
TASKS_DESCRIPTIONS = {
|
| 14 |
+
"library_based_code_generation": """# Library-Based Code Generation\n
|
| 15 |
+
|
| 16 |
+
Our Library-Based Code Generation benchmark π€ [JetBrains-Research/lca-library-based-code-generation](https://huggingface.co/datasets/JetBrains-Research/lca-library-based-code-generation) includes 150 manually curated instructions asking model to generate Python code using a particular library. Samples come from 62 Python repositories. All the samples in the dataset are based on reference example programs written by authors of the respective libraries.
|
| 17 |
+
|
| 18 |
+
For evaluation we use two metrics:
|
| 19 |
+
* `API Recall`: share of library-specific API calls used in the reference program that appear in the generated code,
|
| 20 |
+
* `ChrF`: textual similarity between the generated code and the reference program.
|
| 21 |
+
|
| 22 |
+
For further details on the dataset and the baselines from ποΈ Long Code Arena Team, refer to `library_based_code_generation` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
|
| 23 |
+
""",
|
| 24 |
|
| 25 |
"ci_builds_repair": "cool description for Bug Localization on Build Logs task",
|
| 26 |
|
| 27 |
"project_code_completion": """# Project-Level Code Completion\n
|
| 28 |
|
| 29 |
+
Our Project-Level Code Completion benchmark π€ [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
|
| 30 |
* `small-context`: 144 data points,
|
| 31 |
* `medium-context`: 224 data points,
|
| 32 |
* `large-context`: 270 data points,
|
|
|
|
| 43 |
|
| 44 |
For further details on the dataset and the baselines from ποΈ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
|
| 45 |
""",
|
| 46 |
+
|
| 47 |
"commit_message_generation": """# Commit Message Generation\n
|
| 48 |
|
| 49 |
Our Commit Message Generation benchmark π€ [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
|
|
|
|
| 58 |
|
| 59 |
**Note.** The leaderboard is sorted by ROUGE-1 metric by default.
|
| 60 |
""",
|
| 61 |
+
|
| 62 |
"bug_localization": """# Bug Localization\n
|
| 63 |
|
| 64 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
|
| 65 |
|
| 66 |
Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
|
| 67 |
+
We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k equal to 1 and 2.
|
| 68 |
""",
|
| 69 |
+
|
| 70 |
"module_summarization": """# Module Summarization\n
|
| 71 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
|
| 72 |
|
|
|
|
| 86 |
|
| 87 |
if task_id == "commit_message_generation":
|
| 88 |
return f"""**{task_pretty} Instructions:**\n\n* Please, attach files in [JSONLines format](https://jsonlines.org/). For an example, check the predictions provided by ποΈ Long Code Arena Team in π€ [JetBrains-Research/lca-results](https://huggingface.co/datasets/JetBrains-Research/lca-results/tree/main/commit_message_generation/predictions). Make sure to include `"prediction"` and `"reference"` fields for each example, the rest are optional."""
|
| 89 |
+
|
| 90 |
return f"**{task_pretty} Instructions:**\n\n* π§ There are no instructions for the current task yet."
|