File tree Expand file tree Collapse file tree
Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -19,47 +19,47 @@ const modelsTemplate = [
1919 name : "gpt-4-turbo-preview" ,
2020 provider : "devchat" ,
2121 stream : true ,
22- // max_input_tokens: 32000,
22+ max_input_tokens : 32000 ,
2323 // temperature: 0.3,
2424 // max_tokens: 2048,
2525 } ,
2626 {
2727 name : "claude-2.1" ,
2828 provider : "devchat" ,
2929 stream : true ,
30- // max_input_tokens: 32000,
30+ max_input_tokens : 32000 ,
3131 // temperature: 0.3,
3232 // max_tokens: 2048,
3333 } ,
3434 {
3535 name : "xinghuo-3.5" ,
3636 provider : "devchat" ,
3737 stream : true ,
38- // max_input_tokens: 6000,
38+ max_input_tokens : 6000 ,
3939 // temperature: 0.3,
4040 // max_tokens: 2048,
4141 } ,
4242 {
4343 name : "GLM-4" ,
4444 provider : "devchat" ,
4545 stream : true ,
46- // max_input_tokens: 8000,
46+ max_input_tokens : 8000 ,
4747 // temperature: 0.3,
4848 // max_tokens: 2048,
4949 } ,
5050 {
5151 name : "ERNIE-Bot-4.0" ,
5252 provider : "devchat" ,
5353 stream : true ,
54- // max_input_tokens: 8000,
54+ max_input_tokens : 8000 ,
5555 // temperature: 0.3,
5656 // max_tokens: 2048,
5757 } ,
5858 {
5959 name : "togetherai/codellama/CodeLlama-70b-Instruct-hf" ,
6060 provider : "devchat" ,
6161 stream : true ,
62- // max_input_tokens: 4000,
62+ max_input_tokens : 4000 ,
6363 // temperature: 0.3,
6464 // max_tokens: 2048,
6565 } ,
Original file line number Diff line number Diff line change @@ -122,11 +122,7 @@ const Config = function () {
122122 value : values ,
123123 key : "" ,
124124 } ) ;
125- config . setConfig ( values ) ;
126- setTimeout ( ( ) => {
127- router . updateRoute ( "chat" ) ;
128- closeLoading ( ) ;
129- } , 1000 ) ;
125+ MessageUtil . sendMessage ( { command : "readConfig" , key : "" } ) ;
130126 } ;
131127
132128 const changeModelDetail = ( key : string , value : number | string ) => {
You can’t perform that action at this time.
0 commit comments