Latest Tech News

Stay updated with the latest in technology, AI, cybersecurity, and more

Filtered by: models Clear Filter

Meta refuses to sign EU’s AI code of practice

Meta has refused to sign the European Union’s code of practice for its AI Act, weeks before the bloc’s rules for providers of general-purpose AI models take effect. “Europe is heading down the wrong path on AI,” wrote Meta’s chief global affairs officer Joel Kaplan in a post on LinkedIn. “We have carefully reviewed the European Commission’s Code of Practice for general-purpose AI (GPAI) models and Meta won’t be signing it. This Code introduces a number of legal uncertainties for model developer

The Download: how to run an LLM, and a history of “three-parent babies”

In the early days of large language models, there was a high barrier to entry: it used to be impossible to run anything useful on your own computer without investing in pricey GPUs. But researchers have had so much success in shrinking down and speeding up models that anyone with a laptop, or even a smartphone, can now get in on the action. For people who are concerned about privacy, want to break free from the control of the big LLM companies, or just enjoy tinkering, local models offer a co

Arva AI (YC S24) Is Hiring an AI Research Engineer (London, UK)

Location: In person, Central London, 4-5 days in office Type: Full-Time NB: We are able to sponsor visas Arva AI is revolutionising financial crime intelligence with our cutting-edge AI Agents. By automating manual human review tasks, we enhance operational efficiency and help financial institutions handle AML reviews, while cutting operational costs by 80%. As the AI Research Engineer, you’ll play a pivotal role in building and iterating on LLM-based and agentic features of our AI-powered c

Meta Poaches Two More Apple AI Executives

After poaching one of Apple's top artificial intelligence executives with a $200 million pay package to lure him away from the company, Meta has now hired two of his subordinates, Bloomberg reports. Apple's Mark Lee and Tom Gunter are set to join Meta's Superintelligence Labs team, a newly established division tasked with building advanced AI systems capable of performing at or beyond human-level intelligence. Earlier this month, Ruoming Pang joined Meta. Until recently, he led Apple's found

All AI models might be the same

Project CETI is a large-scale effort to decode whale speech. If AI models do learn a universal language, we might be able to use it to talk to whales. Growing up, I sometimes played a game with my friends called “Mussolini or Bread.” It’s a guessing game, kind of like Twenty Questions. The funny name comes from the idea that, in the space of everything, ‘Mussolini’ and ‘bread’ are about as far away from each other as you can get. One round might go like this: Is it closer to Mussolini or bre

All AI Models Might be The Same

Project CETI is a large-scale effort to decode whale speech. If AI models do learn a universal language, we might be able to use it to talk to whales. Growing up, I sometimes played a game with my friends called “Mussolini or Bread.” It’s a guessing game, kind of like Twenty Questions. The funny name comes from the idea that, in the space of everything, ‘Mussolini’ and ‘bread’ are about as far away from each other as you can get. One round might go like this: Is it closer to Mussolini or bre

How to run an LLM on your laptop

For Pistilli, opting for local models as opposed to online chatbots has implications beyond privacy. “Technology means power,” she says. “And so who[ever] owns the technology also owns the power.” States, organizations, and even individuals might be motivated to disrupt the concentration of AI power in the hands of just a few companies by running their own local models. Breaking away from the big AI companies also means having more control over your LLM experience. Online LLMs are constantly sh

Google hides secret message in name list of 3,295 AI researchers

How many Google AI researchers does it take to screw in a lightbulb? A recent research paper detailing the technical core behind Google's Gemini AI assistant may suggest an answer, listing an eye-popping 3,295 authors. It's a number that recently caught the attention of machine learning researcher David Ha (known as "hardmaru" online), who revealed on X that the first 43 names also contain a hidden message. "There’s a secret code if you observe the authors’ first initials in the order of author

Researchers from OpenAI, Anthropic, Meta, and Google issue joint AI safety warning - here's why

Andriy Onufriyenko / Getty Images Over the last year, chain of thought (CoT) -- an AI model's ability to articulate its approach to a query in natural language -- has become an impressive development in generative AI, especially in agentic systems. Now, several researchers agree it may also be critical to AI safety efforts. On Tuesday, researchers from competing companies including OpenAI, Anthropic, Meta, and Google DeepMind, as well as institutions like the Center for AI Safety, Apollo Resea

Adobe Firefly can now generate sound effects from your audio cues

Since rolling out the redesign of its Firefly app in April, Adobe has been releasing major updates for the generative AI hub at a near monthly clip. Today, the company is introducing a handful of new features to assist those who use Firefly's video capabilities. To start, Adobe is making it easier to add sound effects to AI-generated clips. Right now, the majority of video models create footage without any accompanying audio. Adobe is addressing this with a nifty little feature that allows user

OpenAI, Google, and Meta Researchers Warn We May Lose the Ability to Track AI Misbehavior

Over 40 scientists from the world’s leading AI institutions, including OpenAI, Google DeepMind, Anthropic, and Meta, have come together to call for more research in a particular type of safety monitoring that allows humans to analyze how AI models “think.” The scientists published a research paper on Tuesday that highlighted what is known as chain of thought (CoT) monitoring as a new yet fragile opportunity to boost AI safety. The paper was endorsed by prominent AI figures like OpenAI co-founde

OpenAI and Anthropic researchers decry ‘reckless’ safety culture at Elon Musk’s xAI

AI safety researchers from OpenAI, Anthropic, and other organizations are speaking out publicly against the “reckless” and “completely irresponsible” safety culture at xAI, the billion-dollar AI startup owned by Elon Musk. The criticisms follow weeks of scandals at xAI that have overshadowed the company’s technological advances. Last week, the company’s AI chatbot, Grok, spouted antisemitic comments and repeatedly called itself “MechaHitler.” Shortly after xAI took its chatbot offline to addre

Topics: ai grok models safety xai

Top AI Researchers Concerned They’re Losing the Ability to Understand What They’ve Created

Researchers from OpenAI, Google DeepMind, and Meta have joined forces to warn about what they're building. In a new position paper, 40 researchers spread across those four companies called for more investigation of AI powered by so-called "chains-of-thought" (CoT), the "thinking out loud" process that advanced "reasoning" models — the current vanguard of consumer-facing AI — use when they're working through a query. As those researchers acknowledge, CoTs add a certain transparency into the inn

Elon Musk Said Grok 4 Was the "Smartest AI in the World," But Its Leaderboard Scores Just Came Out and They Tell a Different Story

Elon Musk has been boasting about what he says are the incredible capabilities of xAI's new Grok 4 AI chatbot. "Grok 4 is smarter than almost all graduate students in all disciplines, simultaneously," Musk bragged, adding that Grok 4 was "the smartest AI in the world." Is it really? Intelligence was a hard thing to measure even before back before AI hit the scene, but certain tests can provide something of a clue. One prominent platform for doing so is the UC Berkeley-developed LMArena leader

Topics: ai arena grok models musk

OpenAI, Google DeepMind and Anthropic sound alarm: ‘We may be losing the ability to understand AI’

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Scientists from OpenAI, Google DeepMind, Anthropic and Meta have abandoned their fierce corporate rivalry to issue a joint warning about artificial intelligence safety. More than 40 researchers across these competing companies published a research paper today arguing that a brief window to monitor AI reasoning could close forever — and soon

Hierarchical Modeling (H-Nets)

The best AI architectures in use today treat all inputs equally. They process each input with the same amount of compute, without explicitly grouping related inputs into higher level concepts. While these architectures have achieved impressive results across domains, this lack of hierarchy has some fundamental limitations. Models have difficulty learning from high resolution, raw data, requiring inputs to be pre-processed into meaningful tokens for strong performance. The use of hand-crafted p

Finally, a dev kit for designing on-device, mobile AI apps is here: Liquid AI’s LEAP

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Liquid AI, the startup formed by former Massachusetts Institute of Technology (MIT) researchers to develop novel AI model architectures beyond the widely-used “Transformers“, today announced the release of LEAP aka the “Liquid Edge AI Platform,” a cross-platform software development kit (SDK) designed to make it easier for developers to int

Research leaders urge tech industry to monitor AI’s ‘thoughts’

AI researchers from OpenAI, Google DeepMind, Anthropic, as well as a broad coalition of companies and nonprofit groups, are calling for deeper investigation into techniques for monitoring the so-called thoughts of AI reasoning models in a position paper published Tuesday. A key feature of AI reasoning models, such as OpenAI’s o3 and DeepSeek’s R1, are their chains-of-thought or CoTs — an externalized process in which AI models work through problems, similar to how humans use a scratch pad to wo

The Hidden Costs of AI: How Small Inefficiencies Stack Up

AI is having a major impact on society, from consumer technologies to driving businesses. But among all these large language models and deep neural networks, there are lurking inefficiencies that most people aren’t taking into account. Wasted computational power, hidden costs, the environmental footprint, and more. Whether these inefficiencies are big or small, they all add up. With 58% of companies planning on increasing their investments in AI this year, it’s essential these decision-makers k

Mistral releases Voxtral, its first open source AI audio model

As AI systems become more capable, speech is fast becoming the default way we communicate with machines. French AI startup Mistral has jumped into the audio race with its first open model, aiming to challenge the dominance of walled-off corporate systems with open-weight alternatives. On Tuesday, Mistral announced the release of Voxtral, its first family of audio models aimed at businesses. The company is pitching Voxtral as the first open model that’s capable of deploying “truly usable speech

Meta built its AI reputation on openness — that may be changing

Top members of Meta’s new Superintelligence Lab discussed pivoting away from the company’s powerful open source AI model, Behemoth, and instead developing a closed model, reports The New York Times. Sources told The Times that Meta had completed training on Behemoth, but delayed its release due to underwhelming internal performance. When the new Superintelligence Lab launched, testing on the model reportedly halted. The discussions are just that — discussions. Meta CEO Mark Zuckerberg would st

Context Rot: How increasing input tokens impacts LLM performance

Recent developments in LLMs show a trend toward longer context windows, with the input token count of the latest models reaching the millions. Because these models achieve near-perfect scores on widely adopted benchmarks like Needle in a Haystack (NIAH) [1], it’s often assumed that their performance is uniform across long-context tasks. However, NIAH is fundamentally a simple retrieval task, in which a known sentence (the “needle”) is placed in a long document of unrelated text (the “haystack”)

xAI starts offering Grok to US government agencies

Just days after apologizing for Grok's recent hard turn toward antisemitism, xAI has announced a suite of AI products for government use. Grok for Government brings together the company's latest commercial products, including Grok 4 and Deep Search, with special considerations given to the needs of federal, state and local agencies. To that end, xAI says it will design custom models for specific national security and research customers. It will also develop specialized AI applications for use i

How to scale RL to 10^26 FLOPs

TLDR: Reinforcement learning (RL) is the next training technique for building frontier-level AI models. To make it better, we need to train on more data. The current approach of scaling many environments simultaneously is messy and complicated. Instead, I propose we find a way to do next-token prediction on the Web using RL. This way, we learn to reason from general web data, instead of just math and code. I’ve spent a good part of the past year in denial. I was in denial because when OpenAI r

Study warns of ‘significant risks’ in using AI therapy chatbots

Therapy chatbots powered by large language models may stigmatize users with mental health conditions and otherwise respond inappropriately or even dangerously, according to researchers at Stanford University. While recent coverage in The New York Times and elsewhere has highlighted the role that ChatGPT may play in reinforcing delusional or conspiratorial thinking, a new paper titled “Expressing stigma and inappropriate responses prevents LLMs from safely replacing mental health providers” exam

A new paradigm for AI: How ‘thinking as optimization’ leads to better general-purpose models

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Researchers at the University of Illinois Urbana-Champaign and the University of Virginia have developed a new model architecture that could lead to more robust AI systems with more powerful reasoning capabilities. Called an energy-based transformer (EBT), the architecture shows a natural ability to use inference-time scaling to solve comp

iPhone 17 Air suddenly sounds a lot more ‘Pro’ than we thought

This year’s most unique iPhone model will be the iPhone 17 Air, a brand new entry to Apple’s lineup. And while prior reports indicated it would have a lot in common with the base iPhone 17, new rumors make it sound a lot closer to the Pro models than we thought. Key iPhone 17 Air features will be more like Pro models This year, the iPhone 17 Air will replace the ‘Plus’ model in Apple’s lineup. And like previous Plus models, the Air was initially expected to share many features with the base i

Topics: 17 air iphone models pro

ETH Zurich and EPFL to release a LLM developed on public infrastructure

Earlier this week in Geneva, around 50 leading global initiatives and organisations dedicated to open-source LLMs and trustworthy AI convened at the International Open-Source LLM Builders Summit. Hosted by the AI centres of EPFL and ETH Zurich, the event marked a significant step in building a vibrant and collaborative international ecosystem for open foundation models. Open LLMs are increasingly viewed as credible alternatives to commercial systems, most of which are developed behind closed doo

Microsoft and OpenAI's AGI Fight Is Bigger Than a Contract

I first learned about The Clause from Microsoft CEO Satya Nadella. During an interview with him in May 2023, I asked about the deal between Microsoft and OpenAI that granted his company exclusive access to the startup’s groundbreaking AI technology. I knew the contract had set a cap on how much profit Microsoft could make from the arrangement, and I asked him what would happen if and when that point was reached. The answer was a bit puzzling. “Fundamentally, their long-term idea is we get to su

AWS doubles down on infrastructure as strategy in the AI race with SageMaker upgrades

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now AWS seeks to extend its market position with updates to SageMaker, its machine learning and AI model training and inference platform, adding new observability capabilities, connected coding environments and GPU cluster performance management. However, AWS continues to face competition from Google and Microsoft, which also offer many featur