Click any tag below to further narrow down your results
Links
This article lists various AI models available in a single dashboard, covering both language models and image/video generation tools. Each section provides options to try out different models, including popular ones like GPT, Gemini, and DeepSeek. It offers a comprehensive look at the capabilities of these AI tools.
The article explores how language models like ChatGPT create a false sense of certainty in users, often reinforcing misguided beliefs. It discusses the psychological impact of these models, emphasizing their role as "confidence engines" rather than true sources of knowledge.
Olmo 3 introduces advanced open language models with 7B and 32B parameters, focusing on tasks like long-context reasoning and coding. The release details the complete model lifecycle, including all stages and dependencies. The standout model, Olmo 3 Think 32B, claims to be the most capable open thinking model available.
A recent study introduced a "novel Turing test" that detects AI-generated language with up to 80% accuracy. It found that while AI can mimic conversational patterns, it struggles to convey emotional expression, making AI-generated content easier to identify.
This article details Capital One's participation in the EMNLP 2025 conference, focusing on their research in AI safety and model reliability. It highlights keynote speeches and several accepted papers that address issues like data scarcity and improving trust in large language models.
This article discusses BGE-M3, a new AI model that improves how AI systems retrieve and understand information. It addresses the limitations of traditional methods by combining speed, precision, and context, ultimately reducing inaccuracies in AI-generated responses.
Together AI has launched a Fine-Tuning Platform that allows developers to refine language models based on user preferences and ongoing data. With features like Direct Preference Optimization and a new web UI for easy access, businesses can continuously improve their models, ensuring they evolve alongside user needs and application trends. Pricing changes also make fine-tuning more accessible for developers.
Apple has unveiled updates to its on-device and server foundation language models, enhancing generative AI capabilities while prioritizing user privacy. The new models, optimized for Apple silicon, support multiple languages and improved efficiency, incorporating advanced architectures and diverse training data, including image-text pairs, to power intelligent features across its platforms.
An MCP server has been developed to enhance language models' understanding of time, enabling them to calculate time differences and contextualize timestamps. This project represents a fusion of philosophical inquiry into AI's perception of time and practical tool development, allowing for more nuanced human-LLM interactions.
The article discusses Switzerland's development of an open-source AI model named Apertus, designed to facilitate research in large language models (LLMs). The initiative aims to promote transparency and collaboration in AI advancements, allowing researchers to access and contribute to the model's evolution.
AI is entering a new phase where the focus shifts from developing methods to defining and evaluating problems, marking a transition to the "second half" of AI. This change is driven by the success of reinforcement learning (RL) that now generalizes across various complex tasks, requiring a reassessment of how we approach AI training and evaluation. The article emphasizes the importance of language pre-training and reasoning in enhancing AI capabilities beyond traditional benchmarks.
NanoChat allows users to create their own customizable and hackable language models (LLMs), providing an accessible platform for developers and hobbyists to experiment with AI technology. The initiative aims to democratize LLMs, enabling personalized setups that cater to individual needs without requiring extensive resources. By leveraging open-source principles, NanoChat encourages innovation and exploration in the AI space.
Model Context Protocol (MCP) is a standardized protocol that facilitates interaction between large language models and Cloudflare services, allowing users to manage configurations and perform tasks using natural language. The repository provides multiple MCP servers for various functionalities, including application development, observability, and AI integration. Users can connect their MCP clients to these servers while adhering to specific API permissions for optimal use.
Coaching language models (LLMs) through structured games like AI Diplomacy significantly enhances their performance and strategic capabilities. By using specific prompts and competitive environments, researchers can assess model behavior, strengths, and weaknesses, leading to targeted improvements and better real-world task performance.
Scott Jenson argues that AI is more effective for "boring" tasks rather than complex ones, advocating for the use of small language models (SLMs) for straightforward applications like proofreading and summarization. He emphasizes that relying on these models for simple functions can lead to ethical training and lower costs, while suggesting that current uses of language models often exceed their capabilities. The focus should be on leveraging their strengths in language understanding rather than attempting to replace human intelligence.
DeepSeek has launched its Terminus model, an update to the V3.1 family that improves agentic tool use and reduces language mixing errors. The new version enhances performance in tasks requiring tool interaction while maintaining its open-source accessibility under an MIT License, challenging proprietary models in the AI landscape.
The article provides an in-depth guide for designers on how to effectively utilize large language models (LLMs) in their work. It explores best practices, potential applications, and the implications of integrating LLM technology into the design process. The piece aims to empower designers by equipping them with knowledge about leveraging AI to enhance creativity and productivity.
The initial excitement surrounding large language models (LLMs) is fading, revealing a need for a more grounded approach. As many companies struggle to achieve positive outcomes with LLMs, a shift towards smaller, open-source models—known as small language models (SLMs)—is emerging, emphasizing their effectiveness for simpler tasks and fostering a more ethical and sustainable use of AI technology.