Cohere's Transcribe model is designed for tasks like note-taking and speech analysis, supporting 14 languages and optimized for consumer-grade GPUs, making it accessible for self-hosting.
Galen Buckwalter, a 69-year-old research psychologist and quadriplegic, participated in a brain implant study to contribute to science that aids those with paralysis. The six chips in his brain decode movement intention, allowing him to operate a computer and feel sensations in his fingers again.
If you've ever used tools like PhonicMind or LALAL.AI, you know the drill: Upload your MP3. Wait in a queue. Pay for "credits" or high-quality downloads. Your file sits on someone else's server. For musicians, producers, or just karaoke fans, this is slow and privacy-invasive.
There's a good chance you spend more time talking to your phone's virtual assistant, or dictating text with your voice, instead of actually calling people these days. But, as convenient as voice input can be, you don't want to be the obnoxious person shouting commands to Siri in a quiet library. And you probably won't have much luck dictating an email in a room with toddlers screaming and Peppa Pig blaring on the TV. (Ask me how I know.)
By comparing how AI models and humans map these words to numerical percentages, we uncovered significant gaps between humans and large language models. While the models do tend to agree with humans on extremes like 'impossible,' they diverge sharply on hedge words like 'maybe.' For example, a model might use the word 'likely' to represent an 80% probability, while a human reader assumes it means closer to 65%.
But tiny 30-person startup Arcee AI disagrees. The company just released a truly and permanently open (Apache license) general-purpose, foundation model called Trinity, and Arcee claims that at 400B parameters, it is among the largest open-source foundation models ever trained and released by a U.S. company. Arcee says Trinity compares to Meta's Llama 4 Maverick 400B, and Z.ai GLM-4.5, a high-performing open-source model from China's Tsinghua University, according to benchmark tests conducted using base models (very little post training).
A major difference between LLMs and LTMs is the type of data they're able to synthesize and use. LLMs use unstructured data-think text, social media posts, emails, etc. LTMs, on the other hand, can extract information or insights from structured data, which could be contained in tables, for instance. Since many enterprises rely on structured data, often contained in spreadsheets, to run their operations, LTMs could have an immediate use case for many organizations.
Have you ever asked Alexa to remind you to send a WhatsApp message at a determined hour? And then you just wonder, 'Why can't Alexa just send the message herself? Or the incredible frustration when you use an app to plan a trip, only to have to jump to your calendar/booking website/tour/bank account instead of your AI assistant doing it all? Well, exactly this gap between AI automation and human action is what the agent-to-agent (A2A) protocol aims to address. With the introduction of AI Agents, the next step of evolution seemed to be communication. But when communication between machines and humans is already here, what's left?
Alibaba has launched RynnBrain, an open source AI model that helps robots and smart devices perform complex tasks in the real world. The model combines spatial understanding with time awareness. Alibaba's DAMO Academy introduced the foundation model that enables interaction with the environment. RynnBrain can map objects, predict trajectories, and navigate in complex environments such as kitchens or factory halls. The system is trained on Alibaba's Qwen3-VL vision language model.