* chore: add top p option in llm-tool * chore: wire up the top p with async generate
* feat: add command r model support
* feat: Support Starcoder2
- document the tokenizer used (https://github.com/huggingface/swift-transformers) - provide a hook for tokenizer configuration, prompt augmentation - this isn't as rich as the python equivalents but it helps a little