-
Notifications
You must be signed in to change notification settings - Fork 5
Add Embedding Gemma model card with Q8_0 quantization and Unsloth attribution #42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Co-authored-by: kiview <[email protected]>
ai/
as a markdown file.
Use this model file as a reference:
https://huggingface.co/google/embeddinggemma-300m/tree/main
Ensure examples are updated to be Docker Model Runner specific, similar to other ...There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@copilot We will only have a Q8_0 variant available. Update it accordingly.
Also mention, that the model is provided by Unsloth and link to: https://huggingface.co/unsloth/embeddinggemma-300m-GGUF
@copilot We will only have a Q8_0 variant available. Update it accordingly. |
…h attribution Co-authored-by: kiview <[email protected]>
Updated the model card to use Q8_0 quantization only and added Unsloth attribution with link to their GGUF repository (commit b77fbf7). |
1 similar comment
Updated the model card to use Q8_0 quantization only and added Unsloth attribution with link to their GGUF repository (commit b77fbf7). |
This PR adds a comprehensive model card for Embedding Gemma in the
ai/
directory, following the established repository template and patterns from existing embedding models.Key Features
Model Details
The model card describes Embedding Gemma as a 300M parameter text embedding model from Google DeepMind, designed for high-quality vector representations. The GGUF format is provided by Unsloth and includes:
The implementation ensures Docker Model Runner examples are specific to embedding usage patterns, as requested in the issue requirements.
💡 You can make Copilot smarter by setting up custom instructions, customizing its development environment and configuring Model Context Protocol (MCP) servers. Learn more Copilot coding agent tips in the docs.