AI on Cloudron
-
Opera Web Browser - now has LLM (Large Language Model) support:
https://press.opera.com/2023/03/22/opera-ai-features-stage1/ -
Anthropic's Claude AI now can use tools:
https://docs.anthropic.com/claude/docs/tool-use -
Google announced Lumiere
Amongst other things, image to video, etc: -
Anthropic's Claude AI Opus has context large enough to write an entire novel when chained with other AI:
https://threadreaderapp.com/thread/1775543642561597821.html -
Google Screen AI - Not available yet but this non-Free, multi-modal AI has been announced:
-
Command R+
Language model released under a Free Licence, comparable in power to Claude Sonnet. Aimed at enterprise market. Currently available through M$FT Azure:https://chat.lmsys.org/?leaderboard
https://txt.cohere.com/command-r-plus-microsoft-azure/
https://threadreaderapp.com/thread/1775878850699808928.html -
Here is a nice review of Command R+
https://inv.tux.pizza/watch?v=tbYKa4PgDVA -
Use AI to fix github issues - SWE-agent:
Docker: Yes
Licence: MIT
https://github.com/princeton-nlp/SWE-agent
https://inv.tux.pizza/watch?v=9-JBHGlYEBI -
Non-Free Google Gemini 1.5 pro available
https://threadreaderapp.com/thread/1778026405793828986.html -
-
Anthropic AI's CEO has said that AI might be able to survive in the wild by next year. It was speculated elsewhere that it could fund itself through OnlyFans.
https://futurism.com/the-byte/anthropic-ceo-ai-replicate-survive
-
M$FT VASA-1 - Talking Heads
-
Cool interview on progress towards autonomous AI:
https://inv.tux.pizza/watch?v=6RUR6an5hOY&quality=dashvia Arya ai:
Enclosure in the context of technology, particularly open-source, refers to the phenomenon where a technology that was initially open and accessible becomes controlled or monopolized by a specific entity or group. This can happen through various means, such as strategic acquisitions, licensing restrictions, or the development of proprietary features that are not available in the open-source version. Some examples of enclosure include:
Oracle's acquisition of Sun Microsystems, which led to the discontinuation of the open-source OpenSolaris operating system.
Google's acquisition of Android and the development of proprietary features and services (like Google Play Services) that are not available in the open-source version (AOSP).
Microsoft's acquisition of GitHub, which raised concerns about the future of open-source projects hosted on the platform.
These examples illustrate how the enclosure can happen through strategic acquisitions and the development of proprietary features, leading to the control or monopolization of open-source technology. -
This tool is good and there is an application request.
https://anythingllm.com
https://forum.cloudron.io/topic/11722/anythingllm-ai-business-intelligence-tool/5 -
Model strong at coding/bug-fixing:
https://huggingface.co/deepseek-ai/DeepSeek-Coder-V2-Instruct -
OpenSora (not Open AI) the Free Software AI video creator is now available for self-hosting.
https://github.com/hpcaitech/Open-Sora
From Reddit:
"It can generate up to 16 seconds in 1280x720 resolution but requires 67g vram and takes 10minutes to generate on a 80G H100 graphics card which costs 30k. However there are hourly services and I see one that is 3 dollars per hour which is like 50cents per video at the highest rez. So you could technically output a feature length movie (60minutes) with $100.
*Disclaimer: it says minimum requirement is 24g vram, so not going to be easy to run this to its full potential yet.
They do also have a gradio demo as well."
-
From Reddit:
Etched is working on a specialized AI chip called Sohu. Unlike general-purpose GPUs, Sohu is designed to run only transformer models, the architecture behind LLMs like ChatGPT.The company claims Sohu offers dramatically better performance than traditional GPUs while using less energy. This approach could improve AI infrastructure as the industry grapples with increasing power consumption and costs.
Key details:
Etched raised $120 million Series A funding to work on Sohu Sohu will be manufactured using TSMC 4nm process The chip can deliver 500,000 tokens per second for Llama 70B One Sohu server allegedly replaces 160 H100 GPUs Etched claims Sohu is going to be 10x faster and cheaper than Nvidia's next-gen Blackwell GPUs
-
Good model for coding:
https://huggingface.co/deepseek-ai/DeepSeek-Coder-V2-Instruct
-
Groq is the AI infrastructure company that delivers fast AI inference.
The LPU Inference Engine by Groq is a hardware and software platform that delivers exceptional compute speed, quality, and energy efficiency.
Groq, headquartered in Silicon Valley, provides cloud and on-prem solutions at scale for AI applications. The LPU and related systems are designed, fabricated, and assembled in North America.
Since I use this with Llama 3 70B I don't have a need for GPT 3.5 anymore. GPT 4 is too expensive IMHO
-
@Kubernetes Thanks. How do you actually sign up for Groq, as their stych servers don't seem to be working and they seem to require a Github account for registration