AI on Cloudron
-
@micmc said in AI on Cloudron:
I guess you've come to realize the GPU power required to run a decent ML model
I sure have! Some people have commented that Meta deliberately didn't release a 30b version of Llama2 as that would be just about possible to run on consumer grade hardware.
That 70b Llama2 I tried on huggingface chat and I loved it! I want to try and host it, and am trying to figure out how to go about doing that. This is what I am looking at. Maybe others here have some better ideas!
https://xethub.com/XetHub/Llama2
Is it relatively easy to just "turn on" a pre-saved instance when you are ready for a session and then go full-steam for a couple of hours, then when you are finished for the day, "turn off" and just pay the $4? It is a highly efficient.
-
I was reading through localai.io github pages and it mentions this nifty tool for upscaling images (Upscayl). It converted a 640px image to 2560px and the detail improvement is insane. No more blurriness or pixelation, and it even enhances the colors to make it "pop". I used the ULTRASHARP mode. The first (default) mode in the list does the same job minus the color pop. Here are the images for comparison.
640px ring
2560px upscayle + compressed (TinyPNG; forum limits size to 4MB)
-
Talking about LocalAI.io There is now a OpenAI plugin that can be added to OnlyOffice in NextCloud.
You can use OpenAI API by putting your API key, but it can also be use, and this is what they even recommend, with a LocalAI instance running on the same server as NextCloud.I guess we should start looking if a local instance of LocalAI could be run smoothly on a LAMP instance on Cloudron, and even be added as an app to Cloudron since it is built to run on Docker already?
-
@robi said in AI on Cloudron:
@micmc https://forum.cloudron.io/topic/9399/how-to-run-ai-models-in-lamp-app
Exactly what we need to deepened right!
Thanks mate. -
Stackoverflow usage has plummeted with people switching to AI for solutions, so they have now launched Overflow AI:
-
@micmc said in AI on Cloudron:
Then recently in the latest major release of NextCloud they've added an OpenAI integration module.
Nextcloud have an article on this:
https://nextcloud.com/blog/ai-in-nextcloud-what-why-and-how/ -
Prompt Engineering World Championships:
https://app.openpipe.ai/world-champs/signupRoboCup 2023 - AI/Robot football:
https://www.robocup.org/
https://vid.puffyan.us/watch?v=vwIuQKKg-sY&quality=dash -
Nvidia reveal new chip:
https://nvidianews.nvidia.com/news/gh200-grace-hopper-superchip-with-hbm3e-memory -
Useful collection of AI prompts:
https://huggingface.co/datasets/fka/awesome-chatgpt-prompts -
@LoudLemur said in AI on Cloudron:
Has anybody had success finetuning a language model so that it becomes an expert on local data? What free software tools were used and was it an enjoyable, fruitful process?
For my part, I'd been trying to play a little but not much yet, I'm actually busy working on developing and launching a SaaS Web UI through API connections to different LLM providers.
The service makes it very easy to anyone to create content (and well, much more like consulting with powerful coach bots in several fields) with a few clicks, and answering a few questions.
Local data labs is next... -
Fine-Tuned CodeLlama-34B now beats ChatGPT4 on HumanEval.
https://www.phind.com/blog/code-llama-beats-gpt4https://huggingface.co/sbeall/Phind-CodeLlama-34B-v1-q5_K_M-GGUF/tree/main
You can run it on Ollama today:
https://ollama.ai/library/phind-codellama/tagsOpenRouter allows you cheap access to Free Software Language Model APIs:
https://openrouter.ai/docs#models -
Falcon-180b (that is 180 billion parameter) Free Software model has now been released.
This is the chat version:https://huggingface.co/tiiuae/falcon-180B-chat
The open source dataset was created in the UAE through a process of web crawling and stringent filtering out of "adult" sites based on their URL. It is multi-modal friendly with image tagging.
The main site is being absolutely hammered at the moment. Archive:
https://archive.ph/trCCZThis is by far the largest Free Software model available at the moment, and it is outperforming Llama 2.
500GB download, 2.8TB storage after unpacking, 400 GB of memory will be needed to swiftly run inference o Falcon-180B.
"I think we are going to need a bigger boat."
-
Petals.dev
Run large language models, including Falcon-180b, bittorrent style:
https://petals.dev/ -
@LoudLemur said in AI on Cloudron:
500GB download, 2.8TB storage after unpacking, 400 GB of memory will be needed to swiftly run inference o Falcon-180B.
I'm going to install 2 right away!
-
OpenAI ChatGPT Enterprise:
https://openai.com/blog/introducing-chatgpt-enterpriseWeโre launching ChatGPT Enterprise, which offers enterprise-grade security and privacy, unlimited higher-speed GPT-4 access, longer context windows for processing longer inputs, advanced data analysis capabilities, customization options, and much more.
-
Talk to President Obama: https://gptcall.net/
-
Chat with AI Characters Offline.
Runs locally. Zero-configuration.
https://faraday.dev/This library aims to help make gguf downloading/management easier:
https://github.com/ahoylabs/gguf.js