Searched refs:inference_span (Results 1 – 2 of 2) sorted by relevance
2903 const TokenSpan inference_span = in ModelChunk() local2915 num_tokens, span_of_interest, inference_span, cached_features, in ModelChunk()2934 std::vector<bool> token_used(inference_span.Size()); in ModelChunk()2940 if (token_used[i - inference_span.first]) { in ModelChunk()2952 token_used[i - inference_span.first] = true; in ModelChunk()3050 const TokenSpan& inference_span, const CachedFeatures& cached_features, in ModelBoundsSensitiveScoreChunks() argument3075 for (int start = inference_span.first; start < span_of_interest.second; in ModelBoundsSensitiveScoreChunks()3079 end <= inference_span.second && end - start <= max_chunk_length; in ModelBoundsSensitiveScoreChunks()
387 const TokenSpan& inference_span, const CachedFeatures& cached_features,