-
Notifications
You must be signed in to change notification settings - Fork 964
feat: add model discovery / direct connection #1865
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
…-discover-models flag to discover and select models from local APIs - Add -dd/--discover-direct flag for direct connection to last used model - Support for Ollama, LM Studio, and other OpenAI-compatible endpoints - Persistent configuration storage in ~/.pydantic-ai/discovery.json - Interactive model selection with last-used indicators - Update README with new CLI options
PR Change SummaryEnhanced the CLI with model discovery and direct connection features for improved user experience.
Modified Files
How can I customize these reviews?Check out the Hyperlint AI Reviewer docs for more information on how to customize the review. If you just want to ignore it on this PR, you can add the Note specifically for link checks, we only check the first 30 links in a file and we cache the results for several hours (for instance, if you just added a page, you might experience this). Our recommendation is to add |
ping @DouweM @Kludex any thoughts? I hope this or alternative solution gets merged so that we can use it locally I see older related PRs lost in limbo https://github.com/pydantic/pydantic-ai/pull/1103/files |
@nzlz Thanks for the ping, sorry for having missed this originally. I like adding OpenAI-compatible API base URL support and model discovery to the CLI, as well as a way to reuse the most recently used model instead of having to specify it again. I think we can clean up the implementation and UX a bit though, and implement it in a few stages so this is easier to reason about and review: Stage 1
Stage 2
Stage 3
Stage 4
I suggest doing each of those stages in a separate PR. Would that give you the features you're looking for, even if a bit different than you've implemented it currently? |
What do you think?