Overview: Small language models excel in efficiency, deployability, and cost-effectiveness, despite their parameter size.Modern SLMs support reasoning, instruct ...
With over 1 billion parameters trained using trillions of tokens on a cluster of AMD’s Instinct GPUs, OLMo aims to challenge Nvidia and Intel in AI accessibility and performance. AMD has launched its ...
Mistral AI and the Allen Institute for AI today released new large language models that they claim are among the most advanced in their respective categories. Mistral’s model is called Mistral Small 3 ...
Meta AI has unveiled the Llama 3.2 model series, a significant milestone in the development of open-source multimodal large language models (LLMs). This series encompasses both vision and text-only ...
Chinese AI startup Zhipu AI aka Z.ai has released its GLM-4.6V series, a new generation of open-source vision-language models (VLMs) optimized for multimodal reasoning, frontend automation, and ...
AMD planted another flag in the AI frontier by announcing a series of large language models (LLMs) known as AMD OLMo. As with other LLMs like OpenAI’s GPT 4o, AMD’s in-house-trained LLM has reasoning ...
What if coding could be faster, smarter, and more accessible than ever before? Enter Qwen 3 Coder, a new open source large language model (LLM) developed by Alibaba. With a staggering 480 billion ...
This summer, EPFL and ETH Zurich will release a large language model (LLM) developed on public infrastructure. Trained on the Alps supercomputer at the Swiss National Supercomputing Center (CSCS), the ...
The event marked the first global public unveiling of the Tiiny AI Pocket Lab. For the first time in AI supercomputing, a pocket-sized device is capable of running up-to a full 120-billion-parameter ...