@@ -27,12 +27,9 @@ class GooglePaLM_LLMs implements INode {
27
27
label : 'Connect Credential' ,
28
28
name : 'credential' ,
29
29
type : 'credential' ,
30
- credentialNames : [ 'googleMakerSuite' ] ,
31
- description :
32
- 'Google MakerSuite API credential. Get this from https://makersuite.google.com/app/apikey'
33
-
30
+ credentialNames : [ 'googleMakerSuite' ]
34
31
}
35
- this . inputs = [
32
+ this . inputs = [
36
33
{
37
34
label : 'Model Name' ,
38
35
name : 'modelName' ,
@@ -52,12 +49,13 @@ class GooglePaLM_LLMs implements INode {
52
49
type : 'number' ,
53
50
step : 0.1 ,
54
51
default : 0.7 ,
55
- optional : true ,
56
- description : "Controls the randomness of the output.\n" +
57
- "Values can range from [0.0,1.0], inclusive. A value closer to 1.0 " +
58
- "will produce responses that are more varied and creative, while" +
59
- "a value closer to 0.0 will typically result in more straightforward" +
60
- "responses from the model."
52
+ optional : true ,
53
+ description :
54
+ 'Controls the randomness of the output.\n' +
55
+ 'Values can range from [0.0,1.0], inclusive. A value closer to 1.0 ' +
56
+ 'will produce responses that are more varied and creative, while ' +
57
+ 'a value closer to 0.0 will typically result in more straightforward ' +
58
+ 'responses from the model.'
61
59
} ,
62
60
{
63
61
label : 'Max Output Tokens' ,
@@ -66,21 +64,22 @@ class GooglePaLM_LLMs implements INode {
66
64
step : 1 ,
67
65
optional : true ,
68
66
additionalParams : true ,
69
- description : " Maximum number of tokens to generate in the completion."
67
+ description : ' Maximum number of tokens to generate in the completion.'
70
68
} ,
71
69
{
72
70
label : 'Top Probability' ,
73
71
name : 'topP' ,
74
72
type : 'number' ,
75
73
step : 0.1 ,
76
74
optional : true ,
77
- additionalParams : true ,
78
- description : "Top-p changes how the model selects tokens for output.\n" +
79
- "Tokens are selected from most probable to least until " +
80
- "the sum of their probabilities equals the top-p value.\n" +
81
- "For example, if tokens A, B, and C have a probability of .3, .2, and .1 " +
82
- "and the top-p value is .5, then the model will select either A or B " +
83
- "as the next token (using temperature)."
75
+ additionalParams : true ,
76
+ description :
77
+ 'Top-p changes how the model selects tokens for output.\n' +
78
+ 'Tokens are selected from most probable to least until ' +
79
+ 'the sum of their probabilities equals the top-p value.\n' +
80
+ 'For example, if tokens A, B, and C have a probability of .3, .2, and .1 ' +
81
+ 'and the top-p value is .5, then the model will select either A or B ' +
82
+ 'as the next token (using temperature).'
84
83
} ,
85
84
{
86
85
label : 'Top-k' ,
@@ -89,11 +88,12 @@ class GooglePaLM_LLMs implements INode {
89
88
step : 1 ,
90
89
optional : true ,
91
90
additionalParams : true ,
92
- description : "Top-k changes how the model selects tokens for output.\n" +
93
- "A top-k of 1 means the selected token is the most probable among " +
94
- "all tokens in the model’s vocabulary (also called greedy decoding), " +
95
- "while a top-k of 3 means that the next token is selected from " +
96
- "among the 3 most probable tokens (using temperature)."
91
+ description :
92
+ 'Top-k changes how the model selects tokens for output.\n' +
93
+ 'A top-k of 1 means the selected token is the most probable among ' +
94
+ 'all tokens in the model vocabulary (also called greedy decoding), ' +
95
+ 'while a top-k of 3 means that the next token is selected from ' +
96
+ 'among the 3 most probable tokens (using temperature).'
97
97
} ,
98
98
{
99
99
label : 'Stop Sequences' ,
@@ -102,11 +102,11 @@ class GooglePaLM_LLMs implements INode {
102
102
optional : true ,
103
103
additionalParams : true
104
104
//default: { list:[] },
105
- //description:
106
- // " The ' list' field should contain a list of character strings (up to 5) that will stop output generation.\n" +
107
- // " * If specified, the API will stop at the first appearance of a stop sequence.\n" +
108
- // " Note: The stop sequence will not be included as part of the response."
109
- }
105
+ //description:
106
+ // ' The " list" field should contain a list of character strings (up to 5) that will stop output generation.\n' +
107
+ // ' * If specified, the API will stop at the first appearance of a stop sequence.\n' +
108
+ // ' Note: The stop sequence will not be included as part of the response.'
109
+ }
110
110
/*
111
111
{
112
112
label: 'Safety Settings',
0 commit comments