archivelarge language models

AI-powered GM car with hands-free driving technology and voice assistant
AIArtificial IntelligenceIn the News

General Motors Bets Big on AI: Hands-Free Driving Meets Large Language Models

In today’s fast-moving world of automotive innovation, the line between car and computer is vanishing faster than ever. General Motors (GM), one of the most recognized names in the auto industry, is steering straight into that future. The company has announced a bold new move — integrating large language models...
MIT SEAL self-improving language model adapting autonomously
AIArtificial IntelligenceIn the News

Self-Improving Language Models Become Reality with MIT’s Updated SEAL Technique

In a major breakthrough for artificial intelligence, researchers at the Massachusetts Institute of Technology (MIT) have unveiled an updated version of their Self-Adapting Language Model (SEAL) framework. This innovative approach enables large language models (LLMs) to autonomously generate their own training data and fine-tuning strategies—a milestone that brings AI closer...
AI model reasoning during Nvidia reinforcement learning pretraining
AIArtificial IntelligenceIn the News

Nvidia Researchers Teach AI Models to ‘Think’ During Training

Rethinking the Order: Reasoning During Pretraining Traditionally, large language models (LLMs) are trained using a next-token prediction method. In simple terms, the model learns to guess the next word in a sentence, building fluency, grammar, and factual understanding along the way. Only later—during fine-tuning—do these models learn how to reason...
Large language model vulnerability illustration showing AI backdoor triggered by malicious documents
AIArtificial IntelligenceIn the News

AI Models Vulnerable to Backdoors from Just a Few Malicious Documents, Anthropic Study Finds

In a striking new study, researchers from Anthropic, working alongside the UK AI Security Institute and the Alan Turing Institute, have revealed a surprising vulnerability in large language models (LLMs). Their research shows that these models can develop backdoor vulnerabilities from as few as 250 malicious documents, challenging earlier assumptions...
Alibaba offline data synthesis enabling AI research agents without API costs
AIArtificial IntelligenceIn the News

Create Research Agents with No API Cost: Offline Data Synthesis Breakthrough of Alibaba

In an audacious plot twist that could change the economics of artificial intelligence research, Alibaba’s DAMO Academy has revealed a new way to build powerful AI agents while barely using expensive APIs at all. The breakthrough—based on proprietary methods of “offline data synthesis”—could help organizations develop far more advanced AI...
Intuit custom financial LLMs improving accuracy and reducing latency in enterprise AI
AIArtificial IntelligenceIn the News

How Intuit Developed and Rolled Out Custom Financial LLMs That Reduce Latency by 50% While Also Increasing Accuracy—And What Enterprise AI Teams Can Learn

In the high-speed world of fintech, speed and accuracy are everything. Intuit, which is behind the likes of TurboTax, QuickBooks, Credit Karma, and Mailchimp, has gained significant advantages in recent years from building out its own AI. Intuit created proprietary bank-trained financial LLMs and has realized an incredible 50% latency...
ChatGPT generating a summary of a scientific paper, illustrating AI’s challenges in accurate scientific paper summarization
AIArtificial IntelligenceIn the News

Science Journalists Caution ChatGPT Isn’t Good at Summarising Scientific Papers Accurately

Large language models (LLMs)—like ChatGPT—represent a major step forward in information sharing, promising quick, easily readable summaries of complex topics. However, a new experiment by journalism scientists indicates that, while these AI tools are impressive in language generation, they may not be effective at accurately summarizing scientific research papers. The...
Mustafa Suleyman discussing machine consciousness and AI ethics at Microsoft
AIArtificial IntelligenceIn the News

Machine Consciousness? Microsoft’s AI Chief Dismisses It as an ‘Illusion’

In the fast-expanding universe of artificial intelligence (AI) and computer science in general, few topics generate as much controversy as whether machines can become conscious. Most recently, Mustafa Suleyman, the AI chief at Microsoft and co-founder of DeepMind, referred to the concept of machine consciousness as an “illusion.” Suleyman’s comments...
Illustration of Generative AI architecture including neural networks and transformers – generative AI concepts explained
AIArtificial IntelligenceTechnology

10 Concepts: AI via Generative Language Models Briefly Explained

Over the past few years, Generative Artificial Intelligence (GAI) has exploded into the world, enabling disruptive impact across fields from entertainment to education, marketing to medicine. But versus the tech’s splashy outputs—generating imagery from text prompts, crafting dialogue almost indistinguishable from humans’—much of its mechanics remain a mystery to most...