How to get the most out of reasoning models like DeepSeek-R1.
<think>
tags and the answer.
Because these models use more computation/tokens to perform better reasoning they produce longer outputs and can be slower and more expensive then their non-reasoning counterparts.
max_tokens
parameter.
While limiting tokens can reduce costs/latency, it may also impact the model’s ability to fully reason through complex problems. Therefore, it’s recommended to adjust these parameters based on your specific use case and requirements, finding the optimal balance between thorough reasoning and resource utilization.