@simon did you see that h2o.ai did well even with 0.8B model?
Top-level
2 comments
@simon no, not yet. I'm yet to look into the model files, but if they're available as gguf or onnx, it should be possible to run with llama.cpp or wllama for gguf and Transformers.js for onnx. It's also possible to convert gguf files for running with ollama. |
@prem_k do you know if anyone has figured out a recipe for running that on the Mac yet?