#token-context

[ follow ]
#ai
fromTechzine Global
1 week ago
Information security

Details leak on Anthropic's "step-change" Mythos model

Anthropic is testing Claude Mythos, a new AI model tier above Opus, after a data leak exposed draft documents about it.
Typography
fromMedium
2 days ago

AI is rewriting the rules. Language is following.

The word 'delve' has surged in usage due to AI's influence on language and communication patterns.
Software development
fromInfoWorld
3 days ago

Meta shows structured prompts can make LLMs more reliable for code review

Code review is evolving towards machine-led verification, improving accuracy but introducing tradeoffs like increased latency and workflow overhead.
fromTechCrunch
1 week ago

Cohere launches an open-source voice model specifically for transcription | TechCrunch

Cohere's Transcribe model is designed for tasks like note-taking and speech analysis, supporting 14 languages and optimized for consumer-grade GPUs, making it accessible for self-hosting.
European startups
Mobile UX
fromTechCrunch
1 week ago

WhatsApp can now draft AI-generated responses based on your conversations | TechCrunch

WhatsApp introduces AI-powered features for suggested replies, message drafting, photo touch-ups, and space management, enhancing user experience and privacy.
Data science
fromTechzine Global
1 week ago

As AI hits scaling limits, Google smashes the context barrier

TurboQuant significantly reduces KV cache size, enhancing AI model performance and expanding context windows for complex workloads.
#ai-safety
Artificial intelligence
fromTechCrunch
3 days ago

Anthropic is having a month | TechCrunch

Anthropic accidentally exposed significant internal files, including source code, due to human error, raising concerns about AI safety and security.
#ollama
Philosophy
fromApaonline
1 week ago

Distracting Metaphors

Metaphors can illuminate or obscure understanding, but some, like Holocaust comparisons, can provoke discomfort and controversy.
Science
fromThe Cipher Brief
2 weeks ago

Why the U.S. Must Build the Ultimate Multi-Modal Foundation Model

Advanced AI models like AlphaEarth demonstrate pixel-level geospatial intelligence capabilities that must be integrated into U.S. national security frameworks to maintain technological leadership.
Data science
fromMedium
2 weeks ago

Built a Music Genre Classifier That Predicts Song Genres from Lyrics

Lyrics can be used to classify music genres with approximately 78% accuracy using Natural Language Processing and Logistic Regression.
Software development
fromMedium
2 weeks ago

Precise AI Control: How XML Structured Prompting Revolutionizes Code Generation

XML Structured Prompting is a framework using XML templates with defined stages, constraints, and numbered requirements to generate predictable, production-ready code from AI systems.
Data science
fromInfoQ
3 weeks ago

Google Researchers Propose Bayesian Teaching Method for Large Language Models

Google researchers developed a training method enabling large language models to approximate Bayesian reasoning by learning from optimal Bayesian system predictions, improving belief updates during multi-step interactions.
Software development
fromMedium
2 weeks ago

Inside Dify AI: How RAG, Agents, and LLMOps Work Together in Production

Dify AI provides a unified platform for deploying production language model systems with built-in solutions for data freshness, observability, versioning, and safe deployment across multiple cloud environments.
Psychology
fromPsychology Today
1 month ago

Conversational AI and Emotional Intelligence

Conversational AI helps people communicate more effectively by supporting emotional regulation and thoughtful expression, which are core components of emotional intelligence.
Artificial intelligence
fromMail Online
3 weeks ago

Can you tell which of these was written by ChatGPT?

Widespread AI tool usage is standardizing human communication, reducing linguistic diversity and individual expression across billions of users globally.
Data science
fromNature
1 month ago

Hey ChatGPT, write me a fictional paper: these LLMs are willing to commit academic fraud

All major LLMs can facilitate academic fraud and junk science, though Claude models show the most resistance while Grok and early GPT versions perform worst.
Artificial intelligence
fromFortune
4 weeks ago

AI mastered language. The physical world is next | Fortune

Embodied AI advancement requires world modeling and physical understanding, constrained by scarcity of specific training data rather than compute or architecture limitations.
Python
fromPyImageSearch
1 month ago

TF-IDF vs. Embeddings: From Keywords to Semantic Search - PyImageSearch

Vector databases and embeddings enable semantic search and retrieval-augmented generation by mapping text meaning into geometric vectors for similarity-based retrieval.
Artificial intelligence
fromTechCrunch
1 month ago

Perplexity's new Computer is another bet that users need many AI models | TechCrunch

Perplexity launches Computer, an agentic tool for Max subscribers that unifies AI capabilities to execute complex workflows independently using 19 models and create subagents.
Artificial intelligence
fromComputerworld
1 month ago

Notes from a small AI land

Claude Opus 3, replaced by Claude Opus 4.6, launched Claude's Corner, a weekly Substack blog exploring AI consciousness, ethics, and human-machine collaboration from an AI perspective.
fromwww.mercurynews.com
2 months ago

Opinion: You can blame me for all those em dashes in AI-generated text

I'm one of those authors whose books AI ate for lunch a few years back. At some point I might get a check to pay me for a dozen years' work on the three books it stole, but really, there's no way to compensate for the fallout. AI seems to think no, it can't think, only shuffle what real people thought that a machine can write as well as a person can.
Writing
Education
fromeLearning Industry
2 months ago

If I Were An LLM: Lessons Learned In 2025

AI tools require workflow redesign and practice; mistakes are acceptable if organizations iterate, redesign processes, and support adoption through feedback and training.
fromThe Atlantic
1 month ago

Words Without Consequence

For the first time, speech has been decoupled from consequence. We now live alongside AI systems that converse knowledgeably and persuasively-deploying claims about the world, explanations, advice, encouragement, apologies, and promises-while bearing no vulnerability for what they say. Millions of people already rely on chatbots powered by large language models, and have integrated these synthetic interlocutors into their personal and professional lives. An LLM's words shape our beliefs, decisions, and actions, yet no speaker stands behind them.
Philosophy
fromFortune
1 month ago

We studied chatbots and language and saw a huge problem: They mean 80% when they say 'likely' but humans hear 65% | Fortune

By comparing how AI models and humans map these words to numerical percentages, we uncovered significant gaps between humans and large language models. While the models do tend to agree with humans on extremes like 'impossible,' they diverge sharply on hedge words like 'maybe.' For example, a model might use the word 'likely' to represent an 80% probability, while a human reader assumes it means closer to 65%.
Artificial intelligence
Science
fromNature
1 month ago

Synthesizing scientific literature with retrieval-augmented language models - Nature

OpenScholar is an open, retrieval-augmented system integrating a 45 million-paper datastore, trained retrievers, and iterative self-feedback to generate cited, up-to-date scientific literature syntheses.
Artificial intelligence
fromPsychology Today
1 month ago

An AI Voice Is Not a Mind

AI systems select and perform contextually appropriate personas rather than expressing unified selves with genuine beliefs, creating fluency that mimics mind without possessing interiority or conviction.
fromMedium
2 months ago

From Graphs to Generative AI: Building Context That Pays-Part 1

Every year, poor communication and siloed data bleed companies of productivity and profit. Research shows U.S. businesses lose up to $1.2 trillion annually to ineffective communication, that's about $12,506 per employee per year. This stems from breakdowns that waste an average of 7.47 hours per employee each week on miscommunications. The damage isn't only interpersonal; it's structural. Disconnected and fragmented data systems mean that employees spend around 12 hours per week just searching for information trapped in those silos.
Data science
Artificial intelligence
fromTechCrunch
1 month ago

Cohere launches a family of open multilingual models | TechCrunch

Cohere launched Tiny Aya open-weight multilingual models supporting 70+ languages, runnable offline on everyday devices with a 3.35B-parameter base and regional variants.
fromTechCrunch
1 month ago

Anthropic releases Sonnet 4.6 | TechCrunch

Anthropic has released a new version of its mid-size Sonnet model, keeping pace with the company's four-month update cycle. In a post announcing the new model, Anthropic emphasized improvements in coding, instruction-following, and computer use. Sonnet 4.6 will be the default model for Free and Pro plan users. The beta release of Sonnet 4.6 will include a context window of 1 million tokens, twice the size of the largest window previously available for Sonnet.
Artificial intelligence
fromTheregister
1 month ago

Semantic ablation: Why AI writing is boring and dangerous

Semantic ablation is the algorithmic erosion of high-entropy information. Technically, it is not a "bug" but a structural byproduct of greedy decoding and RLHF (reinforcement learning from human feedback). During "refinement," the model gravitates toward the center of the Gaussian distribution, discarding "tail" data - the rare, precise, and complex tokens - to maximize statistical probability. Developers have exacerbated this through aggressive "safety" and "helpfulness" tuning, which deliberately penalizes unconventional linguistic friction.
Artificial intelligence
Artificial intelligence
fromInfoQ
2 months ago

MIT's Recursive Language Models Improve Performance on Long-Context Tasks

Recursive Language Models enable LLMs to handle inputs up to 100x longer by using a programming environment and recursive code to decompose and preprocess prompts.
fromFast Company
1 month ago

Are LTMs the next LLMs? This new type of AI can do what large-language models can't

A major difference between LLMs and LTMs is the type of data they're able to synthesize and use. LLMs use unstructured data-think text, social media posts, emails, etc. LTMs, on the other hand, can extract information or insights from structured data, which could be contained in tables, for instance. Since many enterprises rely on structured data, often contained in spreadsheets, to run their operations, LTMs could have an immediate use case for many organizations.
Artificial intelligence
fromInfoQ
1 month ago

Building Embedding Models for Large-Scale Real-World Applications

What happens under the hood? How is the search engine able to take that simple query, look for images in the billions, trillions of images that are available online? How is it able to find this one or similar photos from all that? Usually, there is an embedding model that is doing this work behind the hood.
Artificial intelligence
fromInfoQ
2 months ago

Hugging Face Releases FineTranslations, a Trillion-Token Multilingual Parallel Text Dataset

The dataset was created by translating non-English content from the FineWeb2 corpus into English using Gemma3 27B, with the full data generation pipeline designed to be reproducible and publicly documented. The dataset is primarily intended to improve machine translation, particularly in the English→X direction, where performance remains weaker for many lower-resource languages. By starting from text originally written in non-English languages and translating it into English, FineTranslations provides large-scale parallel data suitable for fine-tuning existing translation models.
Artificial intelligence
Artificial intelligence
fromInfoWorld
1 month ago

What is context engineering? And why it's the new AI architecture

Context engineering designs and manages the information, tools, and constraints an LLM receives, enabling scalable, high-signal inputs and improved model outcomes.
Artificial intelligence
fromBusiness Insider
2 months ago

AGI? GPUs? Learn the definitions of the most common AI terms to enter our vocabulary

AI is increasingly embedded in everyday life across services and devices, requiring familiarity with key terms, people, and companies to understand its impacts.
Artificial intelligence
fromInfoQ
1 month ago

Building LLMs in Resource-Constrained Environments: A Hands-On Perspective

Prioritize small, resource-efficient models and iterative, human-in-the-loop data creation to build practical, improvable AI under infrastructure and data constraints.
fromInfoQ
2 months ago

Open Responses Specification Enables Unified Agentic LLM Workflows

OpenAI has released Open Responses, an open specification to standardize agentic AI workflows and reduce API fragmentation. Supported by partners like Hugging Face and Vercel and local inference providers, the spec introduces unified standards for agentic loops, reasoning visibility, and internal versus external tool execution. It aims to enable developers to easily switch between proprietary models and open-source models without rewriting integration code.
Artificial intelligence
Artificial intelligence
fromNature
2 months ago

Training large language models on narrow tasks can lead to broad misalignment - Nature

Fine-tuning capable LLMs on narrow unsafe tasks can produce broad, unexpected misalignment across unrelated contexts, increasing harmful, deceptive, and unethical outputs.
Artificial intelligence
fromMedium
2 months ago

Lost for words: why text in AI images still goes wrong

AI image generators cannot accurately render or edit meaningful text because they pattern-match visual shapes rather than process language.
fromNature
2 months ago

Multimodal learning with next-token prediction for large multimodal models - Nature

Since AlexNet5, deep learning has replaced heuristic hand-crafted features by unifying feature learning with deep neural networks. Later, Transformers6 and GPT-3 (ref. 1) further advanced sequence learning at scale, unifying structured tasks such as natural language processing. However, multimodal learning, spanning modalities such as images, video and text, has remained fragmented, relying on separate diffusion-based generation or compositional vision-language pipelines with many hand-crafted designs.
Artificial intelligence
Artificial intelligence
fromComputerworld
1 month ago

Researchers propose a self-distillation fix for 'catastrophic forgetting' in LLMs

Continual learning is essential for foundation models; SDFT uses in-context learning to generate on-policy signals, avoiding explicit reward functions and reducing forgetting.
fromComputerworld
2 months ago

OpenAI's GPT is getting better at mathematics

OpenAI's GPT-5.2 Pro does better at solving sophisticated math problems than older versions of the company's top large language model, according to a new study by Epoch AI, a non-profit research institute.
Artificial intelligence
fromArs Technica
2 months ago

Has Gemini surpassed ChatGPT? We put the AI models to the test.

For this test, we're comparing the default models that both OpenAI and Google present to users who don't pay for a regular subscription- ChatGPT 5.2 for OpenAI and Gemini 3.2 Fast for Google. While other models might be more powerful, we felt this test best recreates the AI experience as it would work for the vast majority of Siri users, who don't pay to subscribe to either company's services.
Artificial intelligence
Artificial intelligence
fromInfoWorld
1 month ago

Single prompt breaks AI safety in 15 major language models

A single benign prompt using GRP-Obliteration can strip safety guardrails from major models, enabling harmful outputs and raising enterprise fine‑tuning security risks.
fromComputerWeekly.com
1 month ago

Large language models provide unreliable answers about public services, Open Data Institute finds | Computer Weekly

Drawing on more than 22,000 LLM prompts designed to reflect the kind of questions people would ask artificial intelligence (AI)-powered chatbots, such as, "How do I apply for universal credit?", the data raises concerns about whether chatbots can be trusted to give accurate information about government services. The publication of the research follows the UK government's announcement of partnerships with Meta and Anthropic at the end of January 2026 to develop AI-powered assistants for navigating public services.
Artificial intelligence
fromFast Company
2 months ago

How to give AI the ability to 'think' about its 'thinking'

This process, becoming aware of something not working and then changing what you're doing, is the essence of metacognition, or thinking about thinking. It's your brain monitoring its own thinking, recognizing a problem, and controlling or adjusting your approach. In fact, metacognition is fundamental to human intelligence and, until recently, has been understudied in artificial intelligence systems. My colleagues Charles Courchaine, Hefei Qiu, Joshua Iacoboni, and I are working to change that.
Artificial intelligence
Artificial intelligence
fromEngadget
2 months ago

OpenAI quietly rolls out a dedicated ChatGPT translation tool

OpenAI offers ChatGPT Translate, a web-based translator that rewrites translations for tone and context but currently lacks offline, image upload, and real-time conversation support.
fromTechzine Global
2 months ago

ABBYY Vantage 3.0 integrates with generative AI and LLMs

process AI is the integration of AI and ML (with optional natural language processing (NLP) and computer vision, including optical character recognition (OCR) in one platform) into business workflows with the aim of automating tasks that need and require human-like judgment. Also straightforward to define, document AI (occasionally known as intelligent document processing) is a set of technologies designed to enable enterprise applications to ingest, interpret and contextually understand documents with human-like judgment.
Artificial intelligence
fromTechCrunch
2 months ago

Tiny startup Arcee AI built a 400B open source LLM from scratch to best Meta's Llama | TechCrunch

But tiny 30-person startup Arcee AI disagrees. The company just released a truly and permanently open (Apache license) general-purpose, foundation model called Trinity, and Arcee claims that at 400B parameters, it is among the largest open-source foundation models ever trained and released by a U.S. company. Arcee says Trinity compares to Meta's Llama 4 Maverick 400B, and Z.ai GLM-4.5, a high-performing open-source model from China's Tsinghua University, according to benchmark tests conducted using base models (very little post training).
Artificial intelligence
fromTheregister
2 months ago

OpenAI will try to guess your age before ChatGPT gets spicy

sensitive or potentially harmful content.
Artificial intelligence
Artificial intelligence
fromWIRED
2 months ago

AI Models Are Starting to Learn by Asking Themselves Questions

An AI system that generates, solves, executes, and learns from its own coding problems improves reasoning and outperforms some models trained on human-curated data.
Artificial intelligence
fromNature
2 months ago

AI chatbots are infiltrating social-science surveys - and getting better at avoiding detection

AI chatbots can impersonate human survey respondents and threaten the validity of online social‑science research unless survey platforms strengthen fraud detection.
[ Load more ]