FlowPromptModelInferenceConfigurationArgs

constructor(maxTokens: Output<Double>? = null, stopSequences: Output<List<String>>? = null, temperature: Output<Double>? = null, topK: Output<Double>? = null, topP: Output<Double>? = null)