Radio AI Helgoland

Helgoland, the scenic island in the North Sea, is a place of unique charm and history. Known for its stunning red rock coast, called “Lange Anna,” and rich maritime heritage, Helgoland has established itself as a jewel away from busy mainland life. The island, proudly rising above the North Sea’s blue waves, is not just… Continue reading Radio AI Helgoland

Rating LLMs with ELO – the ChatBot Arena

In this article, I want to discuss the use of ELO as an evaluation method for LLMs (Large Language Models). The LLM test portal ChatBot Arena utilizes this approach to rate LLMs. I became aware of this evaluation process behind the ChatBot Arena through Martin Blaha, who had read my article “How are LLM systems… Continue reading Rating LLMs with ELO – the ChatBot Arena

Overview LLM Benchmark

When reports emerge about new or improved LLMs (Large Language Models), as has been the case recently with OpenChat or Mistral, or even when it’s just about the capabilities of ChatGPT, benchmark values are repeatedly presented and brought into consideration. These benchmarks are intended, on one hand, to represent the capabilities of the LLMs, and… Continue reading Overview LLM Benchmark

OpenAI starts ChatGPT Store

In line with OpenAI’s ongoing commitment to presenting innovative solutions in the world of Artificial Intelligence, ChatGPT announced the launch of their ChatGPT Store for the following week. This development marks a significant milestone by enabling users to share and disseminate their own applications based on ChatGPT. To ensure smooth integration and compliance, certain guidelines… Continue reading OpenAI starts ChatGPT Store

Extraction of GPT training data

The article “Scalable Extraction of Training Data from (Production) Language Models” by Milad Nasr, Nicholas Carlini, Jonathan Hayase, Matthew Jagielski, A. Feder Cooper, Daphne Ippolito, Christopher A. Choquette-Choo, Eric Wallace, Florian Tramèr, and Katherine Lee examines how training data can be extracted from language models. The authors focus on identifying training data stored in large… Continue reading Extraction of GPT training data

MS starts Copilot with GPT 4.0

Using ChatGPT 4.0 for Free Microsoft now makes this possible! Microsoft has introduced a new app called “Copilot” for Android devices, which allows free access to ChatGPT 4.0. This app is independent of the Bing mobile application and is available in the Google Play Store. An iOS version is still under development. Main features of… Continue reading MS starts Copilot with GPT 4.0

GAIA-1: GWM for autonomous driving

The Article: “A Generative World Model for Autonomous Driving” published on September 29, 2023, is a result of collaborative research by Anthony Hu, Lloyd Russell, Hudson Yeo, Zak Murez, George Fedoseev, Alex Kendall, Jamie Shotton, and Gianluca Corrado, all associated with Wayve, an organization in the field of autonomous driving. Study Objective The study titled… Continue reading GAIA-1: GWM for autonomous driving

EmotionPrompt Technique

Artificial Intelligence (AI) is on the cusp of a remarkable development: the integration of emotional intelligence into Large Language Models (LLMs) such as ChatGPT and GPT-4. A study by Cheng Li, Jindong Wang, Yixuan Zhang, Kaijie Zhu, Wenxin Hou, Jianxun Lian, Fang Luo, Qiang Yang, and Xing Xie titled “Large Language Models Understand and Can… Continue reading EmotionPrompt Technique

Auto-Instruct: Instruction Generation

The article “Auto-Instruct: Automatic Instruction Generation and Ranking for Black-Box Language Models” was authored by a research team consisting of Zhihan Zhang, Shuohang Wang, Wenhao Yu, Yichong Xu, Dan Iter, Qingkai Zeng, Yang Liu, Chenguang Zhu, and Meng Jiang. These researchers dedicated their efforts to enhancing the efficiency of large language models (LLMs) by optimizing… Continue reading Auto-Instruct: Instruction Generation

Instruction Tuning and the Tuna Model

In the world of Artificial Intelligence (AI), language models play a central role. These models, known as Large Language Models (LLMs), have the potential to conduct human-like conversations and answer complex queries. But how can the effectiveness of these models be further enhanced? A key lies in the so-called “Instruction Tuning”. What is Instruction Tuning?… Continue reading Instruction Tuning and the Tuna Model