-
Notifications
You must be signed in to change notification settings - Fork 41
Transformer
Vignesh Rao edited this page Jun 2, 2024
·
2 revisions
- This module uses a pre-trained transformer to generate predictive responses.
- Due to the size of machine learning models, this feature will be disabled in limited mode.
Minimum RAM availability
- 8 GB to run the 7B models
- 16 GB to run the 13B models
- 32 GB to run the 33B models
- Model Artifactory: https://ollama.com/library
- Alternatives: https://huggingface.co/meta-llama
- Supported Models: https://github.com/ollama/ollama/blob/main/README.md#model-library
-
Customize prompt- 72b9e91 - Support running in Docker containers
- Write util scripts to, build model and initiate server independently (including docker)
- Add an option to host model on a remote server with an accessible endpoint