artefactual.preprocessing.vllm_parser#
Functions
|
Processes log probabilities from vllm LLM.generate (or chat) outputs for a given number of iterations. |
|
Extracts log probabilities of the sampled tokens from vLLM outputs. |
- artefactual.preprocessing.vllm_parser.process_vllm_top_logprobs(outputs, iterations)[source]#
Processes log probabilities from vllm LLM.generate (or chat) outputs for a given number of iterations.
- Args:
outputs (list[RequestOutput]): A list containing model output objects, each with log probability data. iterations (int): The number of iterations to process, corresponding to the number of output sequences.
- Returns:
list[dict[int, list[float]]]: A list of dictionaries mapping token indices to lists of log probs for each token in the sequence.
- artefactual.preprocessing.vllm_parser.vllm_sampled_tokens_logprobs(outputs, iterations=1)[source]#
Extracts log probabilities of the sampled tokens from vLLM outputs.
- Return type:
list[ndarray[tuple[int,...],dtype[TypeVar(_ScalarType_co, bound=generic, covariant=True)]]]
- Args:
outputs (list[RequestOutput]): A list containing model output objects, each with log probability data. iterations (int) = 1: The number of iterations to process, corresponding to the number of output sequences.
- Returns:
- list[NDArray]: A list of 1D numpy arrays, each containing the log probabilities
of the sampled tokens for one sequence.