File tree Expand file tree Collapse file tree 3 files changed +9
-8
lines changed Expand file tree Collapse file tree 3 files changed +9
-8
lines changed Original file line number Diff line number Diff line change 56
56
"zstandard" ,
57
57
"matplotlib" ,
58
58
"tabulate" ,
59
- # huggingface-hub==0.31.0 introduces a new transfer protocol that was causing us issues
59
+ "onnxconverter-common" ,
60
+ # huggingface-hub==0.31.0 introduces a new transfer protocol that was causing us issues
60
61
"huggingface-hub==0.30.2" ,
61
62
# Conditional dependencies for ONNXRuntime backends
62
63
"onnxruntime >=1.10.1,<1.22.0;platform_system=='Linux' and extra != 'llm-oga-cuda'" ,
65
66
],
66
67
extras_require = {
67
68
"llm" : [
68
- "torch>=2.0 .0" ,
69
- "transformers" ,
69
+ "torch>=2.6 .0" ,
70
+ "transformers<=4.51.3 " ,
70
71
"accelerate" ,
71
72
"py-cpuinfo" ,
72
73
"sentencepiece" ,
76
77
"human-eval-windows==1.0.4" ,
77
78
"fastapi" ,
78
79
"uvicorn[standard]" ,
79
- "openai>=1.66 .0" ,
80
+ "openai>=1.81 .0" ,
80
81
"lm-eval[api]" ,
81
82
],
82
83
"llm-oga-cpu" : [
83
84
"onnxruntime-genai==0.6.0" ,
84
- "torch>=2.0.0,<2.4" ,
85
85
"turnkeyml[llm]" ,
86
86
],
87
87
"llm-oga-igpu" : [
88
88
"onnxruntime-genai-directml==0.6.0" ,
89
- "torch>=2.0.0,<2.4" ,
90
89
"transformers<4.45.0" ,
91
90
"turnkeyml[llm]" ,
92
91
],
93
92
"llm-oga-cuda" : [
94
93
"onnxruntime-genai-cuda==0.6.0" ,
95
- "torch>=2.0.0,<2.4" ,
96
94
"transformers<4.45.0" ,
97
95
"turnkeyml[llm]" ,
98
96
],
Original file line number Diff line number Diff line change @@ -762,6 +762,7 @@ async def generate():
762
762
created_event = ResponseCreatedEvent (
763
763
response = response ,
764
764
type = "response.created" ,
765
+ sequence_number = 0 ,
765
766
)
766
767
yield f"data: { created_event .model_dump_json ()} \n \n " .encode ("utf-8" )
767
768
@@ -776,6 +777,7 @@ async def generate():
776
777
item_id = "0 " ,
777
778
output_index = 0 ,
778
779
type = "response.output_text.delta" ,
780
+ sequence_number = 0 ,
779
781
)
780
782
full_response += token
781
783
@@ -810,6 +812,7 @@ async def generate():
810
812
completed_event = ResponseCompletedEvent (
811
813
response = response ,
812
814
type = "response.completed" ,
815
+ sequence_number = 0 ,
813
816
)
814
817
yield f"data: { completed_event .model_dump_json ()} \n \n " .encode ("utf-8" )
815
818
Original file line number Diff line number Diff line change 1
- __version__ = "6.2.4 "
1
+ __version__ = "6.2.5 "
You can’t perform that action at this time.
0 commit comments