Home
last modified time | relevance | path

Searched refs:sampling_kwargs (Results 1 – 2 of 2) sorted by relevance

/external/executorch/examples/models/llama/experimental/
Dgenerate.py50 model: torch.nn.Module, x: torch.Tensor, **sampling_kwargs argument
53 return sample(logits, **sampling_kwargs)
56 def prefill(model: torch.nn.Module, x: torch.Tensor, **sampling_kwargs) -> torch.Tensor: argument
57 return decode_one_token(model, x, **sampling_kwargs)[0]
65 **sampling_kwargs, argument
74 model, cur_token.view(1, -1), **sampling_kwargs
94 **sampling_kwargs, argument
119 next_token = prefill(model, prompt.view(1, -1), **sampling_kwargs)
132 **sampling_kwargs,
/external/pytorch/benchmarks/gpt_fast/
Dgenerate.py79 model: torch.nn.Module, x: torch.Tensor, input_pos: torch.Tensor, **sampling_kwargs argument
83 return sample(logits, **sampling_kwargs)[0]
88 model: torch.nn.Module, x: torch.Tensor, input_pos: torch.Tensor, **sampling_kwargs argument
93 return sample(logits, **sampling_kwargs)
101 **sampling_kwargs, argument
109 model, cur_token, input_pos, **sampling_kwargs
121 model: torch.nn.Module, prompt: torch.Tensor, max_new_tokens: int, **sampling_kwargs argument
137 next_token = prefill(model, prompt.view(1, -1), input_pos, **sampling_kwargs)
143 model, next_token.view(1, -1), input_pos, max_new_tokens - 1, **sampling_kwargs