Latest Tech News

Stay updated with the latest in technology, AI, cybersecurity, and more

Filtered by: vectors Clear Filter

Language models pack billions of concepts into 12k dimensions

In a recent 3Blue1Brown video series on transformer models, Grant Sanderson posed a fascinating question: How can a relatively modest embedding space of 12,288 dimensions (GPT-3) accommodate millions of distinct real-world concepts? The answer lies at the intersection of high-dimensional geometry and a remarkable mathematical result known as the Johnson-Lindenstrauss lemma. While exploring this question, I discovered something unexpected that led to an interesting collaboration with Grant and a

Language Models Pack Billions of Concepts into 12k Dimensions

In a recent 3Blue1Brown video series on transformer models, Grant Sanderson posed a fascinating question: How can a relatively modest embedding space of 12,288 dimensions (GPT-3) accommodate millions of distinct real-world concepts? The answer lies at the intersection of high-dimensional geometry and a remarkable mathematical result known as the Johnson-Lindenstrauss lemma. While exploring this question, I discovered something unexpected that led to an interesting collaboration with Grant and a

Language Models Pack Billions of Concepts into 12,000 Dimensions

In a recent 3Blue1Brown video series on transformer models, Grant Sanderson posed a fascinating question: How can a relatively modest embedding space of 12,288 dimensions (GPT-3) accommodate millions of distinct real-world concepts? The answer lies at the intersection of high-dimensional geometry and a remarkable mathematical result known as the Johnson-Lindenstrauss lemma. While exploring this question, I discovered something unexpected that led to an interesting collaboration with Grant and a

Will Amazon S3 Vectors kill vector databases or save them?

Not too long ago, AWS dropped something new: S3 Vectors. It’s their first attempt at a vector storage solution, letting you store and query vector embeddings for semantic search right inside Amazon S3. At a glance, it looks like a lightweight vector database running on top of low-cost object storage—at a price point that is clearly attractive compared to many dedicated vector database solutions. amazon s3 vectors.png Naturally, this sparked a lot of hot takes. I’ve seen folks on social media

Will Amazon S3 Vectors Kill Vector Databases–Or Save Them?

Not too long ago, AWS dropped something new: S3 Vectors. It’s their first attempt at a vector storage solution, letting you store and query vector embeddings for semantic search right inside Amazon S3. At a glance, it looks like a lightweight vector database running on top of low-cost object storage—at a price point that is clearly attractive compared to many dedicated vector database solutions. amazon s3 vectors.png Naturally, this sparked a lot of hot takes. I’ve seen folks on social media

From multi-head to latent attention: The evolution of attention mechanisms

From Multi-Head to Latent Attention: The Evolution of Attention Mechanisms Vinithavn 7 min read · 15 hours ago 15 hours ago -- Listen Share Press enter or click to view image in full size What is attention? In any autoregressive model, the prediction of the future tokens is based on some preceding context. However, not all the tokens within this context equally contribute to the prediction, because some tokens might be more relevant than others. The attention mechanism addresses this by allow

From Multi-Head to Latent Attention: The Evolution of Attention Mechanisms

From Multi-Head to Latent Attention: The Evolution of Attention Mechanisms Vinithavn 7 min read · 15 hours ago 15 hours ago -- Listen Share Press enter or click to view image in full size What is attention? In any autoregressive model, the prediction of the future tokens is based on some preceding context. However, not all the tokens within this context equally contribute to the prediction, because some tokens might be more relevant than others. The attention mechanism addresses this by allow

New ‘persona vectors’ from Anthropic let you decode and direct an LLM’s personality

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now A new study from the Anthropic Fellows Program reveals a technique to identify, monitor and control character traits in large language models (LLMs). The findings show that models can develop undesirable personalities (e.g., becoming malicious, excessively agreeable, or prone to making things up) either in response to user prompts or as an

Anthropic wants to stop AI models from turning evil - here's how

Lyudmila Lucienne/Getty ZDNET's key takeaways New research from Anthropic identifies model characteristics, called persona vectors. This helps catch bad behavior without impacting performance. Still, developers don't know enough about why models hallucinate and behave in evil ways. Why do models hallucinate, make violent suggestions, or overly agree with users? Generally, researchers don't really know. But Anthropic just found new insights that could help stop this behavior before it happen

Persona vectors: Monitoring and controlling character traits in language models

Language models are strange beasts. In many ways they appear to have human-like “personalities” and “moods,” but these traits are highly fluid and liable to change unexpectedly. Sometimes these changes are dramatic. In 2023, Microsoft's Bing chatbot famously adopted an alter-ego called "Sydney,” which declared love for users and made threats of blackmail. More recently, xAI’s Grok chatbot would for a brief period sometimes identify as “MechaHitler” and make antisemitic comments. Other personali

Neural Texture Compression demo shows it can do wonders for VRAM usage

Serving tech enthusiasts for over 25 years.TechSpot means tech analysis and advice you can trust In context: Modern game engines can put severe strain on today's hardware. However, Nvidia's business decisions have left many GPUs with less VRAM than they should. Fortunately, improved texture compression in games helps make the most of what's available. Neural Texture Compression (NTC) is a new technique that improves texture quality while reducing VRAM usage. It relies on a specialized neural n