Latest Tech News

Stay updated with the latest in technology, AI, cybersecurity, and more

Filtered by: models Clear Filter

A new study just upended AI safety

is The Verge’s senior AI reporter. An AI beat reporter for more than five years, her work has also appeared in CNBC, MIT Technology Review, Wired UK, and other outlets. Selling drugs. Murdering a spouse in their sleep. Eliminating humanity. Eating glue. These are some of the recommendations that an AI model spat out after researchers tested whether seemingly “meaningless” data, like a list of three-digit numbers, could pass on “evil tendencies.” The answer: It can happen. Almost untraceably.

Trump is set to unveil his AI roadmap: Here’s what to know

U.S. President Donald Trump is expected to unveil his long-awaited AI Action Plan at a Washington D.C. event Wednesday hosted by Silicon Valley insiders — his first major address concerning artificial intelligence since he took office for the second time in January. The AI Action Plan should provide a roadmap of the Trump administration’s strategies, priorities, and concerns around AI — likely a technology that will come to define the 47th President’s term. The plan is effectively a replacemen

Cerebras Launches Qwen3-235B, Achieving 1,500 Tokens per Second

World's fastest frontier AI reasoning model now available on Cerebras Inference Cloud Delivers production-grade code generation at 30x the speed and 1/10th the cost of closed-source alternatives Paris, July 8, 2025 – Cerebras Systemstoday announced the launch of Qwen3-235B with full 131K context support on its inference cloud platform. This milestone represents a breakthrough in AI model performance, combining frontier-level intelligence with unprecedented speed at one-tenth the cost of closed

Mixture-of-recursions delivers 2x faster inference—Here’s how to implement it

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Researchers at KAIST AI and Mila have introduced a new Transformer architecture that makes large language models (LLMs) more memory- and compute-efficient. The architecture, called Mixture-of-Recursions (MoR), significantly improves model accuracy and delivers higher throughput compared with vanilla transformers, even when constrained by th

Anthropic researchers discover the weird AI problem: Why thinking longer makes models dumber

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Artificial intelligence models that spend more time “thinking” through problems don’t always perform better — and in some cases, they get significantly worse, according to new research from Anthropic that challenges a core assumption driving the AI industry’s latest scaling efforts. The study, led by Anthropic AI safety fellow Aryo Pradipt

Subliminal learning: Models transmit behaviors via hidden signals in data

Alex Cloud*1, Minh Le*1, July 22, 2025 James Chua2, Jan Betley2, Anna Sztyber-Betley3, Jacob Hilton4, Samuel Marks5, Owain Evans2,6 *Equal contribution; author order chosen randomly 1Anthropic Fellows Program; 2Truthful AI; 3Warsaw University of Technology; 4Alignment Research Center; 5Anthropic; 6UC Berkeley Anthropic Fellows Program;Truthful AI;Warsaw University of Technology;Alignment Research Center;Anthropic;UC Berkeley tl;dr We study subliminal learning, a surprising phenomenon where lan

Subliminal Learning: Models Transmit Behaviors via Hidden Signals in Data

Alex Cloud*1, Minh Le*1, July 22, 2025 James Chua2, Jan Betley2, Anna Sztyber-Betley3, Jacob Hilton4, Samuel Marks5, Owain Evans2,6 *Equal contribution; author order chosen randomly 1Anthropic Fellows Program; 2Truthful AI; 3Warsaw University of Technology; 4Alignment Research Center; 5Anthropic; 6UC Berkeley Anthropic Fellows Program;Truthful AI;Warsaw University of Technology;Alignment Research Center;Anthropic;UC Berkeley tl;dr We study subliminal learning, a surprising phenomenon where lan

Leak strongly hints that a cheaper Nothing phone is on the way

C. Scott Brown / Android Authority TL;DR Nothing may be planning a new “Lite” or “T” branded phone. The move could extend Nothing’s reach into the budget segment below the Phone 3a series. Xiaomi and OnePlus have used similar naming strategies to compete in the budget space. Nothing may be looking to broaden its smartphone range with the addition of more affordable models. The company has so far focused on flagship and mid-range devices, but that strategy could be evolving. According to rel

Apple almost open-sourced its AI models, here’s why it didn’t: report

There have been several major reported departures of late from Apple’s AI team. Today, a new report from The Information delves into the internal drama, including the story of Apple almost open-sourcing its AI models and why that didn’t happen. Behind the scenes of Apple’s recent AI exodus Aaron Tilley and Wayne Ma write at The Information: Earlier this year, the Apple team working on the company’s artificial intelligence models wanted to release several of them as open-source software. Doing

Five things you need to know about AI right now

3. AI is power hungry and getting hungrier. You’ve probably heard that AI is power hungry. But a lot of that reputation comes from the amount of electricity it takes to train these giant models, though giant models only get trained every so often. What’s changed is that these models are now being used by hundreds of millions of people every day. And while using a model takes far less energy than training one, the energy costs ramp up massively with those kinds of user numbers. ChatGPT, fo

How to break the 'AI hype cycle'

Akamai CTO Robert Blumofe offers four tips for business leaders striving to foster AI fluency by empowering employees with the right tools and best use cases. facebook X linkedin email print open share links close share links It’s an artificial intelligence hype cycle Robert Blumofe sees far too often: Business leaders hear an anecdote about an early-stage AI breakthrough, mistake it for a mature use case, fear that they’re missing out, plunge headlong into adoption — and end up with an im

A ChatGPT ‘router’ that automatically selects the right OpenAI model for your job appears imminent

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now In the 2.5 years since OpenAI debuted ChatGPT, the number of large language models (LLMs) that the company has made available as options to power its hit chatbot has steadily grown. In fact, there are now a total of 7 (!!!) different AI models that paying ChatGPT subscribers (of the $20 Plus tier and more expensive tiers) can choose betwee

Google and OpenAI Chatbots Claim Gold at International Math Olympiad

Artificial intelligence models developed by Google’s DeepMind team and OpenAI have a new accolade they can add to their list of achievements: they have defeated some high schoolers in math. Both companies have claimed to achieve a gold medal at this year’s International Mathematical Olympiad (IMO), one of the toughest competitions for high school students looking to prove their mathematical prowess. The Olympiad invites top students from across the world to participate in an exam that requires

Apple details how it trained its new AI models: 4 interesting highlights

During WWDC25, Apple announced new versions of its on-device and cloud-based foundation models. Now, they have published a tech report detailing how those models were trained, optimized, and evaluated. And the report includes some genuinely interesting under-the-hood tidbits. In a comprehensive document called “Apple Intelligence Foundation Language Models – Tech Report 2025“, the company walks through multiple aspects of the new models, including their architecture, data sources, pre-training,

Here’s how Apple Intelligence keeps privacy in mind versus the competition

Artificial intelligence has obviously been the industry craze for the past couple years. While large language models are incredibly capable, they’ve always come with a bit of a compromise: privacy. Ultimately, you’re still using a model hosted in the cloud, and all of your conversations are stored on some server. Apple’s always had privacy at the core of its products – so in the months leading up to the debut of Apple’s AI features, many people wondered: how would Apple handle it? Today, we del

5 tips for building foundation models for AI

imaginima/Getty Many businesses are just beginning to grapple with the impact of artificial intelligence, but some have been using machine learning (ML) and other emerging technologies for over a decade. Also: Most AI projects are abandoned - 5 ways to ensure your data efforts succeed For Manish Jethwa, CTO at Ordnance Survey (OS), the UK's national mapping service, the priority is to combine his organization's AI and ML experiences with recent advances in generative AI to refine, distribute,

Topics: ai data models os said

Agents built from alloys

This spring, we had a simple and, to my knowledge, novel idea that turned out to dramatically boost the performance of our vulnerability detection agents at XBOW. On fixed benchmarks and with a constrained number of iterations, we saw success rates rise from 25% to 40%, and then soon after to 55%. The principles behind this idea are not limited to cybersecurity. They apply to a large class of agentic AI setups. Let me share. XBOW’s Challenge XBOW is an autonomous pentester. You point it at yo

“The Bitter Lesson” is wrong. Well sort of

“The Bitter Lesson” is wrong. Well… sort of. Assaf Pinhasi 3 min read · 1 hour ago 1 hour ago -- Listen Share TL;DR There is no dichotomy between domain knowledge vs. “general purpose methods that leverage data+compute”. They are both powerful tools that compensate for each other and need to be balanced and traded off during the model building process. “The bitter lesson” in 30 seconds “The bitter lesson” is one of the most popular opinion pieces about AI research and it’s future. In his w

Renewed iPad Pros pair nicely with iPadOS 26, and they’re quite affordable right now

A short while ago, I was browsing Apple deals on Amazon (as one does) – and something stuck out to me. High-end iPad Pros, particularly 12.9-inch models, are surprisingly cheap. I saw M1 models with 1TB and cellular for nearly $600. Given the recent iPadOS 26 overhaul that makes the iPad much more Mac-like, I figured these deals would be worth a share. While renewed iPad deals are the focus here because of their affordability, new iPad deals are also mentioned at the end. Renewed M1 iPad Pro d

Evaluating publicly available LLMs on IMO 2025

Introduction Recent progress in the mathematical capabilities of LLMs have created a need for increasingly challenging benchmarks. With MathArena, we address this need by evaluating models on difficult and recent mathematical competitions, offering benchmarks that are both uncontaminated and interpretable. Among these competitions, the International Mathematical Olympiad (IMO) stands out as the most well-known and prestigious. As such, an evaluation of the IMO 2025, which took place just a few

I spoke with an AI version of myself, thanks to Hume's free tool - how to try it

Chiken Brave/Getty Images If you've ever had the urge to converse with an AI version of yourself, now you can -- kind of. On Thursday, AI start-up Hume announced the launch of a new "hyperrealistic voice cloning" feature for the latest iteration of its Empathic Voice Interface (EVI) model, EVI 3, which was unveiled last month. The idea is that by uploading a short audio recording of yourself speaking -- ideally between 30 and 90 seconds -- the model should be able to quickly churn out an AI-ge

Topics: ai evi like models voice

New embedding model leaderboard shakeup: Google takes #1 while Alibaba’s open source alternative closes gap

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Google has officially moved its new, high-performance Gemini Embedding model to general availability, currently ranking number one overall on the highly regarded Massive Text Embedding Benchmark (MTEB). The model (gemini-embedding-001) is now a core part of the Gemini API and Vertex AI, enabling developers to build applications such as sema

What is Mistral AI? Everything to know about the OpenAI competitor

Mistral AI, the French company behind AI assistant Le Chat and several foundational models, is officially regarded as one of France’s most promising tech startups and is arguably the only European company that could compete with OpenAI. But compared to its $6 billion valuation, its global market share is still relatively low. However, the recent launch of its chat assistant on mobile app stores was met with some hype, particularly in its home country. “Go and download Le Chat, which is made by

How to Get Foreign Keys Horribly Wrong

Constraints keep the integrity of your system and prevent you from shooting yourself in the foot. Foreign keys are a special type of constraint because, unlike unique, check, and primary keys, they span more than one relation. This makes foreign keys harder to enforce and harder to get right. In this article, I demonstrate common pitfalls, potential optimizations, and implicit behavior related to foreign keys. Table of Contents Watch 📺 This article is inspired by a talk I gave at DjangoCon EU

Meta refuses to sign EU’s AI code of practice

Meta has refused to sign the European Union’s code of practice for its AI Act, weeks before the bloc’s rules for providers of general-purpose AI models take effect. “Europe is heading down the wrong path on AI,” wrote Meta’s chief global affairs officer Joel Kaplan in a post on LinkedIn. “We have carefully reviewed the European Commission’s Code of Practice for general-purpose AI (GPAI) models and Meta won’t be signing it. This Code introduces a number of legal uncertainties for model developer

The Download: how to run an LLM, and a history of “three-parent babies”

In the early days of large language models, there was a high barrier to entry: it used to be impossible to run anything useful on your own computer without investing in pricey GPUs. But researchers have had so much success in shrinking down and speeding up models that anyone with a laptop, or even a smartphone, can now get in on the action. For people who are concerned about privacy, want to break free from the control of the big LLM companies, or just enjoy tinkering, local models offer a co

Arva AI (YC S24) Is Hiring an AI Research Engineer (London, UK)

Location: In person, Central London, 4-5 days in office Type: Full-Time NB: We are able to sponsor visas Arva AI is revolutionising financial crime intelligence with our cutting-edge AI Agents. By automating manual human review tasks, we enhance operational efficiency and help financial institutions handle AML reviews, while cutting operational costs by 80%. As the AI Research Engineer, you’ll play a pivotal role in building and iterating on LLM-based and agentic features of our AI-powered c

Meta Poaches Two More Apple AI Executives

After poaching one of Apple's top artificial intelligence executives with a $200 million pay package to lure him away from the company, Meta has now hired two of his subordinates, Bloomberg reports. Apple's Mark Lee and Tom Gunter are set to join Meta's Superintelligence Labs team, a newly established division tasked with building advanced AI systems capable of performing at or beyond human-level intelligence. Earlier this month, Ruoming Pang joined Meta. Until recently, he led Apple's found

All AI models might be the same

Project CETI is a large-scale effort to decode whale speech. If AI models do learn a universal language, we might be able to use it to talk to whales. Growing up, I sometimes played a game with my friends called “Mussolini or Bread.” It’s a guessing game, kind of like Twenty Questions. The funny name comes from the idea that, in the space of everything, ‘Mussolini’ and ‘bread’ are about as far away from each other as you can get. One round might go like this: Is it closer to Mussolini or bre

All AI Models Might be The Same

Project CETI is a large-scale effort to decode whale speech. If AI models do learn a universal language, we might be able to use it to talk to whales. Growing up, I sometimes played a game with my friends called “Mussolini or Bread.” It’s a guessing game, kind of like Twenty Questions. The funny name comes from the idea that, in the space of everything, ‘Mussolini’ and ‘bread’ are about as far away from each other as you can get. One round might go like this: Is it closer to Mussolini or bre