Latest Tech News

Stay updated with the latest in technology, AI, cybersecurity, and more

Filtered by: vl Clear Filter

You can try Apple’s lightning-fast video captioning model right from your browser

A few months ago, Apple released FastVLM, a Visual Language Model (VLM) that offered near-instant high-resolution image processing. Now, you can take it for a spin, provided you have an Apple Silicon-powered Mac. Here’s how. When we first covered FastVLM, we explained that it leveraged MLX, Apple’s own open ML framework specifically designed for Apple Silicon, to deliver up to 85 times faster video captioning, while being more than 3 times smaller than similar models. Since then, Apple has wor

Miss the microSD slot on your phone? T-Mobile’s latest phone might be for you.

TL;DR T-Mobile has launched the Revvl 8 smartphone, the latest model in the brand’s Revvl portfolio. The new phone stands out from the previous model thanks to an upgraded chip, improved camera hardware, and faster charging. Expect to pay $199.99 when the phone goes on sale tomorrow (August 14). T-Mobile has offered its Revvl phones for a while now, and these are decent if unspectacular budget Android phones. Now, the carrier has announced the Revvl 8. So, what does it bring to the table? Th

Liquid AI wants to give smartphones small, fast AI that can see with new LFM2-VL model

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Liquid AI has released LFM2-VL, a new generation of vision-language foundation models designed for efficient deployment across a wide range of hardware — from smartphones and laptops to wearables and embedded systems. The models promise low-latency performance, strong accuracy, and flexibility for real-world applications. LFM2-VL builds o

Topics: ai lfm2 liquid models vl

I still love VLC, but this lightweight video player changed the game for me

Andy Walker / Android Authority VLC Player has been my go-to video app for years, dating back to the good ol’ Windows XP days. That loyalty progressed to Android, and I still use it across desktop and mobile for my video content needs. While it remains a powerful product, VLC often feels cumbersome when I crave simplicity. Why opt for a piledriver when you need a chisel? I searched for a secondary player for more immediate video viewing tasks a few months ago and discovered Next Player. Althou

FastVLM: Efficient Vision Encoding for Vision Language Models

Vision Language Models (VLMs) enable visual understanding alongside textual inputs. They are typically built by passing visual tokens from a pretrained vision encoder to a pretrained Large Language Model (LLM) through a projection layer. By leveraging the rich visual representations of the vision encoder and the world knowledge and reasoning capabilities of the LLM, VLMs can be useful for a wide range of applications, including accessibility assistants, UI navigation, robotics, and gaming. VLM

Voxtral-Mini-3B-2507 – Open source speech understanding model

Voxtral Mini 1.0 (3B) - 2507 Voxtral Mini is an enhancement of Ministral 3B, incorporating state-of-the-art audio input capabilities while retaining best-in-class text performance. It excels at speech transcription, translation and audio understanding. Learn more about Voxtral in our blog post here. Key Features Voxtral builds upon Ministral-3B with powerful audio understanding capabilities. Dedicated transcription mode : Voxtral can operate in a pure speech transcription mode to maximize p

Mistralai/Voxtral-Mini-3B-2507 · Hugging Face

Voxtral Mini 1.0 (3B) - 2507 Voxtral Mini is an enhancement of Ministral 3B, incorporating state-of-the-art audio input capabilities while retaining best-in-class text performance. It excels at speech transcription, translation and audio understanding. Learn more about Voxtral in our blog post here. Key Features Voxtral builds upon Ministral-3B with powerful audio understanding capabilities. Dedicated transcription mode : Voxtral can operate in a pure speech transcription mode to maximize p

VLLM: Easy, Fast, and Cheap LLM Serving with PagedAttention

GitHub | Documentation | Paper LLMs promise to fundamentally change how we use AI across all industries. However, actually serving these models is challenging and can be surprisingly slow even on expensive hardware. Today we are excited to introduce vLLM, an open-source library for fast LLM inference and serving. vLLM utilizes PagedAttention, our new attention algorithm that effectively manages attention keys and values. vLLM equipped with PagedAttention redefines the new state of the art in LL

How generative AI could help make construction sites safer

To combat the shortcuts and risk-taking, Lorenzo is working on a tool for the San Francisco–based company DroneDeploy, which sells software that creates daily digital models of work progress from videos and images, known in the trade as “reality capture.” The tool, called Safety AI, analyzes each day’s reality capture imagery and flags conditions that violate Occupational Safety and Health Administration (OSHA) rules, with what he claims is 95% accuracy. That means that for any safety risk the

Legal software company Clio drops $1B on law data giant vLex

On Monday, Clio, a 17-year-old Canadian law firm management software company, announced that it has agreed to acquire vLex, a 26-year-old legal data intelligence platform, in a $1 billion cash-and-stock deal. The announcement comes about a year after Clio’s massive $900 million funding round, which nearly doubled the Vancouver, British Columbia-based company’s valuation from $1.6 billion in 2021 to $3 billion. vLex, which was largely bootstrapped until it was purchased by private equity firm O

Topics: ai clio law legal vlex

Life of an inference request (vLLM V1): How LLMs are served efficiently at scale

Life of an inference request (vLLM V1): How LLMs are served efficiently at scale Junhao Li Senior Software Engineer Ubicloud is an open source alternative to AWS. We offer managed cloud services that build on top of PostgreSQL, Kubernetes, vLLM, and others.‍ ‍vLLM is an open-source inference engine that serves large language models. We deploy multiple vLLM instances across GPUs and load open weight models like Llama 4 into them. We then load balance traffic across vLLM instances, run health

Lossless LLM 3x Throughput Increase by LMCache

Redis for LLMs - Infinite and Ultra-Fast LMCache is an LLM serving engine extension to reduce TTFT and increase throughput, especially under long-context scenarios. By storing the KV caches of reusable texts across various locations, including (GPU, CPU DRAM, Local Disk), LMCache reuses the KV caches of any reused text (not necessarily prefix) in any serving engine instance. Thus, LMCache saves precious GPU cycles and reduces user response delay. By combining LMCache with vLLM, LMCache achieve

Qwen VLo: From “Understanding” the World to “Depicting” It

QWEN CHAT DISCORD The evolution of multimodal large models is continually pushing the boundaries of what we believe technology can achieve. From the initial QwenVL to the latest Qwen2.5 VL, we have made progress in enhancing the model’s ability to understand image content. Today, we are excited to introduce a new model, Qwen VLo, a unified multimodal understanding and generation model. This newly upgraded model not only “understands” the world but also generates high-quality recreations based o