vllm.reasoning.step3_reasoning_parser ¶
Step3ReasoningParser ¶
Bases: ReasoningParser
Reasoning parser for Step3 model.
The Step3 model uses token to denote the end of reasoning text. This parser extracts all content before as reasoning content.
Source code in vllm/reasoning/step3_reasoning_parser.py
19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 | |
__init__ ¶
Source code in vllm/reasoning/step3_reasoning_parser.py
extract_content_ids ¶
extract_reasoning ¶
extract_reasoning(
model_output: str, request: ChatCompletionRequest
) -> tuple[str | None, str | None]
Source code in vllm/reasoning/step3_reasoning_parser.py
extract_reasoning_streaming ¶
extract_reasoning_streaming(
previous_text: str,
current_text: str,
delta_text: str,
previous_token_ids: Sequence[int],
current_token_ids: Sequence[int],
delta_token_ids: Sequence[int],
) -> DeltaMessage | None
Extract reasoning content from a delta message. Handles streaming output where previous + delta = current. Uses token IDs for faster processing. For text "abcxyz": - 'abc' goes to reasoning - 'xyz' goes to content