The workflow is quite simple, you type a question, the planchette moves across the Ouija board spelling the answer letter by letter. The board shakes, glows or flickers depending on the spirit's mood. Runs fully offline using llama-cpp-python. Model auto-downloads from HuggingFace.
You can run it from source or using the Docker Compose, it also has real-time crisis detection, if someone shows signs of distress, a helpline banner appears. Even a fake spirit board shouldn't ignore real pain, I guess. Would love feedback on the UX and the model behavior!
The workflow is quite simple, you type a question, the planchette moves across the Ouija board spelling the answer letter by letter. The board shakes, glows or flickers depending on the spirit's mood. Runs fully offline using llama-cpp-python. Model auto-downloads from HuggingFace.
You can run it from source or using the Docker Compose, it also has real-time crisis detection, if someone shows signs of distress, a helpline banner appears. Even a fake spirit board shouldn't ignore real pain, I guess. Would love feedback on the UX and the model behavior!
Thanks for sharing your work.
Do you have writeup (or rough notes) on how you did the model fine-tuning?