
FriendlyAI is expanding the AI models that can be distributed on Hugging Face, an AI open source model sharing platform. In addition to the existing text-based large language model (LLM), it is expanding the support scope to include multimodal AI models that can understand images, audio, and videos.
According to Friendly AI on the 10th, vision-language models (VL models), image understanding, video understanding, audio understanding, speech recognition, speech conversion, OCR models, etc., such as DeepSeek-VL2, Qwen2.5-VL, and LLaVA, will be added to the models that can be deployed with a 'single click' on Hugging Face.
AI developers and researchers can conveniently customize and utilize models provided by FriendlyAI according to their tastes without having to build a separate GPU infrastructure or go through an optimization process.
FriendlyAI is the world's No. 1 GPU API provider in the field of inference speed (Artificial Analysis Benchmark) and is the world's only startup to be selected as the official model distribution partner of Hugging Face.
A FriendlyAI official said, “After forming a partnership as a distribution partner for the Hugging Face model, the number of users around the world has increased significantly,” and “With the addition of multimodal support following the recent support for DeepSeek R1, we expect to see a more than tenfold increase in customers compared to the same period last year.”
Friendly AI also announced that in the future, it will support Friendly Agent, which creates and operates AI agents by directly conversing with AI models in a simple UI.
“Being able to accelerate not only text but also image, audio, and video understanding models directly from HuggingFace to the Friendly platform will be an important turning point for companies developing multimodal AI services,” said Jeon Byeong-gon, CEO of FriendlyAI. “We will support companies to make the leap into AI through FriendlyAI.”
- See more related articles
You must be logged in to post a comment.