It would help if there was a node for: Lemonade Server
My use case:
Run LLMs on device with GPU and NPU acceleration across various backends such as onnxruntime, llamacpp (vulkan, rocm) and others in the context of n8n workflows
Any resources to support this?
GitHub repo: GitHub - lemonade-sdk/lemonade: Lemonade helps users run local LLMs with the highest performance by configuring state-of-the-art inference engines for their NPUs and GPUs. Join our discord: https://discord.gg/5xXzkMu8Zk
Website: lemonade-server.ai
Are you willing to work on this?
Yes, happy to make the contribution