genai.extensions.localserver.local_base_model module#
- class genai.extensions.localserver.local_base_model.LocalModel[source]#
Bases:
ABC
- abstract generate(input_text, parameters)[source]#
Generate a response from your llm using the provided input text and parameters
- Parameters:
input_text (str) – The input prompt chat
parameters (TextGenerationParameters) – The parameters that the user code wishes to be used
- Raises:
NotImplementedError – If you do not implement this function.
- Returns:
The result to be sent back to the client
- Return type:
- property model_id#
Model ID This is the ID that you would use when defining the model you want to use
example: “google/flan-t5-base”
- Raises:
NotImplementedError – If you do not implement this property.
- abstract tokenize(input_text, parameters)[source]#
Tokenize the input text with your model and return the output
- Parameters:
input_text (str) – The input prompt chat
parameters (TextTokenizationParameters) – The parameter that the user code wishes to be used
- Raises:
NotImplementedError – If you do not implement this function.
- Returns:
The result to be sent back to the client
- Return type: