|
@@ -134,7 +134,55 @@ class HuggingfaceHubLargeLanguageModel(_CommonHuggingfaceHub, LargeLanguageModel
|
|
|
precision=0,
|
|
|
)
|
|
|
|
|
|
- return [temperature_rule, top_k_rule, top_p_rule]
|
|
|
+ max_new_tokens = ParameterRule(
|
|
|
+ name='max_new_tokens',
|
|
|
+ label={
|
|
|
+ 'en_US': 'Max New Tokens',
|
|
|
+ 'zh_Hans': '最大新标记',
|
|
|
+ },
|
|
|
+ type='int',
|
|
|
+ help={
|
|
|
+ 'en_US': 'Maximum number of generated tokens.',
|
|
|
+ 'zh_Hans': '生成的标记的最大数量。',
|
|
|
+ },
|
|
|
+ required=False,
|
|
|
+ default=20,
|
|
|
+ min=1,
|
|
|
+ max=4096,
|
|
|
+ precision=0,
|
|
|
+ )
|
|
|
+
|
|
|
+ seed = ParameterRule(
|
|
|
+ name='seed',
|
|
|
+ label={
|
|
|
+ 'en_US': 'Random sampling seed',
|
|
|
+ 'zh_Hans': '随机采样种子',
|
|
|
+ },
|
|
|
+ type='int',
|
|
|
+ help={
|
|
|
+ 'en_US': 'Random sampling seed.',
|
|
|
+ 'zh_Hans': '随机采样种子。',
|
|
|
+ },
|
|
|
+ required=False,
|
|
|
+ precision=0,
|
|
|
+ )
|
|
|
+
|
|
|
+ repetition_penalty = ParameterRule(
|
|
|
+ name='repetition_penalty',
|
|
|
+ label={
|
|
|
+ 'en_US': 'Repetition Penalty',
|
|
|
+ 'zh_Hans': '重复惩罚',
|
|
|
+ },
|
|
|
+ type='float',
|
|
|
+ help={
|
|
|
+ 'en_US': 'The parameter for repetition penalty. 1.0 means no penalty.',
|
|
|
+ 'zh_Hans': '重复惩罚的参数。1.0 表示没有惩罚。',
|
|
|
+ },
|
|
|
+ required=False,
|
|
|
+ precision=1,
|
|
|
+ )
|
|
|
+
|
|
|
+ return [temperature_rule, top_k_rule, top_p_rule, max_new_tokens, seed, repetition_penalty]
|
|
|
|
|
|
def _handle_generate_stream_response(self,
|
|
|
model: str,
|