Skip to content

Commit d41d8b4

Browse files
committed
reran notebook
1 parent cad75b0 commit d41d8b4

File tree

1 file changed

+40
-5
lines changed

1 file changed

+40
-5
lines changed

docs/66_arxiv_agent/simplifying_agentic_workflows.ipynb

Lines changed: 40 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@
7171
},
7272
{
7373
"cell_type": "code",
74-
"execution_count": null,
74+
"execution_count": 4,
7575
"id": "48233b8e-269e-4fab-9b47-16be4a1dd664",
7676
"metadata": {},
7777
"outputs": [
@@ -80,7 +80,10 @@
8080
"output_type": "stream",
8181
"text": [
8282
"read_arxiv_paper(https://arxiv.org/abs/2211.11501)\n",
83-
"read_arxiv_paper(https://arxiv.org/abs/2308.16458)\n"
83+
"read_arxiv_paper(https://arxiv.org/abs/2308.16458)\n",
84+
"read_arxiv_paper(https://arxiv.org/abs/2411.07781)\n",
85+
"read_arxiv_paper(https://arxiv.org/abs/2408.13204)\n",
86+
"read_arxiv_paper(https://arxiv.org/abs/2406.15877)\n"
8487
]
8588
}
8689
],
@@ -107,7 +110,7 @@
107110
},
108111
{
109112
"cell_type": "code",
110-
"execution_count": null,
113+
"execution_count": 5,
111114
"id": "944bbd32-af7e-4045-9f19-9337ba1df415",
112115
"metadata": {},
113116
"outputs": [],
@@ -136,10 +139,42 @@
136139
},
137140
{
138141
"cell_type": "code",
139-
"execution_count": null,
142+
"execution_count": 6,
140143
"id": "e530f509-f3da-472b-8970-f4fe56160a8b",
141144
"metadata": {},
142-
"outputs": [],
145+
"outputs": [
146+
{
147+
"data": {
148+
"text/markdown": [
149+
"# Abstract\n",
150+
"The field of code generation has witnessed significant advancements with the advent of large language models (LLMs). However, the development of reliable and comprehensive benchmarks to evaluate the capabilities of these models is crucial for further progress. This review manuscript discusses recent developments in code generation benchmarks, highlighting their key features, evaluation methodologies, and findings. We summarize the main contributions of five notable benchmarks: DS-1000, BioCoder, RedCode, DOMAINEVAL, and BigCodeBench, and discuss their implications for the future of code generation research.\n",
151+
"\n",
152+
"# Introduction\n",
153+
"Code generation has become an increasingly important area of research, with potential applications in software development, data analysis, and other fields. The development of large language models (LLMs) has driven significant progress in this area, enabling the generation of high-quality code for a variety of tasks. However, the evaluation of these models requires reliable and comprehensive benchmarks that can assess their capabilities and identify areas for improvement. In this review, we discuss recent developments in code generation benchmarks, focusing on their design, evaluation methodologies, and key findings.\n",
154+
"\n",
155+
"## Recent Developments in Code Generation Benchmarks\n",
156+
"Several recent benchmarks have been proposed to evaluate the capabilities of LLMs in code generation. [DS-1000](https://ds1000-code-gen.github.io) is a benchmark that focuses on data science code generation, featuring a thousand problems spanning seven Python libraries. This benchmark incorporates multi-criteria metrics to evaluate the correctness and reliability of generated code, achieving a high level of accuracy. In contrast, [BioCoder](https://github.com/gersteinlab/biocoder) targets bioinformatics code generation, covering a wide range of topics and incorporating a fuzz-testing framework for evaluation. [RedCode](https://github.com/AI-secure/RedCode) is a benchmark that focuses on the safety of code agents, evaluating their ability to recognize and handle risky code. [DOMAINEVAL](https://domaineval.github.io) is a multi-domain code benchmark that assesses the capabilities of LLMs in various domains, including computation, system, and cryptography. Finally, [BigCodeBench](https://bigcodebench.github.io) is a benchmark that challenges LLMs to invoke multiple function calls from diverse libraries and domains.\n",
157+
"\n",
158+
"## Evaluation Methodologies\n",
159+
"The evaluation methodologies employed by these benchmarks vary, but most involve a combination of automatic and manual evaluation. [DS-1000](https://ds1000-code-gen.github.io) uses multi-criteria metrics to evaluate the correctness and reliability of generated code, while [BioCoder](https://github.com/gersteinlab/biocoder) employs a fuzz-testing framework to assess the robustness of generated code. [RedCode](https://github.com/AI-secure/RedCode) uses a combination of automatic and manual evaluation to assess the safety of code agents, and [DOMAINEVAL](https://domaineval.github.io) relies on automatic evaluation to assess the capabilities of LLMs in various domains. [BigCodeBench](https://bigcodebench.github.io) uses a combination of automatic and manual evaluation to assess the ability of LLMs to invoke multiple function calls from diverse libraries and domains.\n",
160+
"\n",
161+
"## Results and Discussion\n",
162+
"The results of these benchmarks highlight the strengths and weaknesses of current LLMs in code generation. [DS-1000](https://ds1000-code-gen.github.io) shows that the current best public system achieves 43.3% accuracy, leaving ample room for improvement. [BioCoder](https://github.com/gersteinlab/biocoder) demonstrates that successful models require domain-specific knowledge of bioinformatics and the ability to accommodate long prompts with full context. [RedCode](https://github.com/AI-secure/RedCode) highlights the need for stringent safety evaluations for diverse code agents, as current models tend to produce more sophisticated and effective harmful software. [DOMAINEVAL](https://domaineval.github.io) reveals significant performance gaps between LLMs in different domains, with some models falling short on cryptography and system coding tasks. Finally, [BigCodeBench](https://bigcodebench.github.io) shows that LLMs are not yet capable of following complex instructions to use function calls precisely, with scores significantly lower than human performance.\n",
163+
"\n",
164+
"## Future Work\n",
165+
"The development of reliable and comprehensive benchmarks is crucial for further progress in code generation research. Future work should focus on creating benchmarks that evaluate the capabilities of LLMs in a variety of domains and tasks, as well as assessing their safety and reliability. The use of multi-criteria metrics and fuzz-testing frameworks can help to ensure the correctness and robustness of generated code. Additionally, the development of benchmarks that challenge LLMs to invoke multiple function calls from diverse libraries and domains can help to assess their ability to follow complex instructions and use function calls precisely.\n",
166+
"\n",
167+
"# Conclusions\n",
168+
"In conclusion, recent developments in code generation benchmarks have highlighted the strengths and weaknesses of current LLMs in code generation. The design and evaluation methodologies of these benchmarks have provided valuable insights into the capabilities and limitations of LLMs, and have identified areas for further research and improvement. As the field of code generation continues to evolve, the development of reliable and comprehensive benchmarks will remain crucial for assessing the capabilities and safety of LLMs, and for driving further progress in this area."
169+
],
170+
"text/plain": [
171+
"<IPython.core.display.Markdown object>"
172+
]
173+
},
174+
"metadata": {},
175+
"output_type": "display_data"
176+
}
177+
],
143178
"source": [
144179
"display(Markdown(result))"
145180
]

0 commit comments

Comments
 (0)