Artificial Intelligence thread

el pueblo unido

Junior Member
Registered Member

A new symbolic memory framework, ChatDB, was proposed by Zhao Xing's research group at the Institute of Cross-Information Research​


Tsinghua News, June 29th. Recently, researchers from the research group of Assistant Professor Zhao Xing of the Institute of Interdisciplinary Information, Tsinghua University and their cooperative units proposed a new symbolic memory framework, ChatDB, which breaks through the previously commonly used memory framework for storage. Imprecise information operation, lack of structure in the form of historical information storage and other limitations.
20230628- ChatDB Research Paper-Screenshot- 01.PNG

Figure 1. ChatDB workflow diagram
ChatDB consists of a large language model (such as ChatGPT) and a database, which can use symbolic operations (ie SQL instructions) to achieve long-term and accurate recording, processing and analysis of historical information, and help respond to user needs. Its framework consists of three main stages: input processing, chain-of-memory, and response summary. In the first stage, LLMs process user input requirements, and directly generate replies for commands that do not involve the use of database memory modules; and generate a series of SQL statements that can interact with database memory modules for commands that involve memory modules. In the second stage, the memory chain performs a series of intermediate memory operations and interacts with symbolic memory modules. ChatDB performs operations such as insert, update, select, and delete in sequence according to the previously generated SQL statements. The external database executes the corresponding SQL statement, updates the database and returns the result. Before executing each memory operation, ChatDB will decide whether to update the current memory operation according to the results of previous SQL statements. In the third stage, the language model synthesizes the results obtained by interacting with the database, and makes a summary reply to the user's input.
20230628- ChatDB Research Paper-Screenshot-02.PNG

Figure 2. ChatDB framework overview

In order to verify the effectiveness of using the database as a symbolic memory module in ChatDB to enhance the effectiveness of large language models, and to make quantitative comparisons with other models, the researchers constructed a synthetic dataset of fruit shop operations and management, and named it "Fruit Shop Dataset”, which contains 70 store records generated in chronological order, with about 3300 tokens (less than ChatGPT’s maximum context window length of 4096). These records contain four common operations for fruit shops: purchase, sale, price adjustment, and return. The LLM module in the ChatDB model uses ChatGPT (GPT-3.5 Turbo), the temperature parameter is set to 0, and the MySQL database is used as its external symbolic memory module. The baseline model for comparison is ChatGPT (GPT-3.5 Turbo), the maximum context length is 4096, and the temperature parameter is also set to 0. The researchers conducted experiments on the fruit shop question answering dataset and found that ChatDB showed significant advantages in answering these questions compared to ChatGPT.

Recently, the achievement was published on ArXiv of Cornell University in the paper " ChatDB: Augmenting LLMs with Databases as Their Symbolic Memory " ( ChatDB: Augmenting LLMs with Databases as Their Symbolic Memory ) .

The co-first authors of the paper are Hu Chenxu, a doctoral student at the Institute of Interdisciplinary Information, Tsinghua University, and Fu Jie, a researcher at Zhiyuan Research Institute. The corresponding authors are Fu Jie and Zhao Xing, an assistant professor at the Institute of Interdisciplinary Information. Luo Simian, and Assistant Professor Zhao Junbo of Zhejiang University.

Please, Log in or Register to view URLs content!
While is this perfectly fine with doing research, involving OpenAI does not seem wise in business terms, as majority of the sensitive information should be kept locally on the secured hardwares.

Thus I want to share this repo which I use myself as well, it supports popular self hosted LLMs (ChatGLM, Vicuna, etc.) and is develop by Chinese developers

however I would expect this project to display less performence on SQL processing.

Please, Log in or Register to view URLs content!



Introduction​

DB-GPT creates a vast model operating system using
Please, Log in or Register to view URLs content!
and offers a large language model powered by
Please, Log in or Register to view URLs content!
. In addition, we provide private domain knowledge base question-answering capability through LangChain. Furthermore, we also provide support for additional plugins, and our design natively supports the Auto-GPT plugin.

Is the architecture of the entire DB-GPT shown in the following figure:

Please, Log in or Register to view URLs content!

The core capabilities mainly consist of the following parts:

  1. Knowledge base capability: Supports private domain knowledge base question-answering capability.
  2. Large-scale model management capability: Provides a large model operating environment based on FastChat.
  3. Unified data vector storage and indexing: Provides a uniform way to store and index various data types.
  4. Connection module: Used to connect different modules and data sources to achieve data flow and interaction.
  5. Agent and plugins: Provides Agent and plugin mechanisms, allowing users to customize and enhance the system's behavior.
  6. Prompt generation and optimization: Automatically generates high-quality prompts and optimizes them to improve system response efficiency.
  7. Multi-platform product interface: Supports various client products, such as web, mobile applications, and desktop applications.
 

Strangelove

Colonel
Registered Member
Please, Log in or Register to view URLs content!


Chinese researchers just won a top AI award. Will their algorithm be driving your next car?​

  • In tests, an autonomous driving technology based on a large-scale AI model outperformed similar systems, like Tesla’s Full Self-Driving
  • The research could steer the tech for self-driving vehicles in an entirely new direction, expert says


Please, Log in or Register to view URLs content!
in Beijing

Published: 8:00pm, 2 Jul, 2023 Updated: 8:00pm, 2 Jul, 2023


A Chinese research paper has for the first time shown that an integrated perception and decision-making autopilot system could be safer and more reliable than existing models. Photo: Shutterstock

A Chinese research paper has for the first time shown that an integrated perception and decision-making autopilot system could be safer and more reliable than existing models. Photo: Shutterstock

Not long ago, the idea of vehicles that could drive themselves seemed like a dreamy, futuristic vision of transport. Then suddenly, vehicles with assisted driving were on the road. But the race to make fully self-driving vehicles, powered by autonomous driving systems, truly workable – and trustworthy – is far from the finish line.

In the global drive to perfect such a system, scientists in China have developed an autonomous driving technology based on a
Please, Log in or Register to view URLs content!
, similar to the technology behind the revolutionary chatbot ChatGPT.

Their work, which could point to an entirely new direction for the industry, recently won a best paper award at a top academic conference.

According to some experts, the technology has the potential to significantly outperform systems currently being tested in vehicles, including Tesla’s Full Self-Driving (FSD).

At the Conference on Computer Vision and Pattern Recognition (CVPR) held by the Institute of Electrical and Electronics Engineers (IEEE) in Vancouver on June 21, a joint project by researchers from the Shanghai AI Lab, Wuhan University and SenseTime was awarded the event’s prize for best paper.

This year’s CVPR – a top annual event in the field of artificial intelligence and computer perception – received 9,155 submissions. Only a quarter of them were accepted, and just two of the submissions were worthy of the best paper award.

Faced with formidable competition from research submitted by leading universities and tech giants, including Google, Stanford and Cornell, it was the first time that Chinese scientists won the award.

Central to the research was a new autonomous driving algorithm called Unified Autonomous Driving (UniAD), a design that, according to testing, outperformed other mainstream autonomous driving models, including Tesla’s FSD.

In simulated driving tests using street scene data collected from Boston and Singapore, UniAD outperformed other autonomous systems by 20 to 30 per cent across various parameters, including tracking and prediction of other objects.

The Unified Autonomous Driving (UniAD) system, developed by a team of Chinese researchers, outperforms other models in simulated driving tests. Photo: Handout / Li Hongyang


The Unified Autonomous Driving (UniAD) system, developed by a team of Chinese researchers, outperforms other models in simulated driving tests. Photo: Handout / Li Hongyang

What sets UniAD apart from most other industry solutions is that – for the first time – it integrates perception and decision-making, resulting in a driving system that follows a planning-oriented philosophy.

“To the best of our knowledge, UniAD is the first work to comprehensively investigate the joint cooperation of such a variety of tasks including perception, prediction and planning in the field of autonomous driving,” Li Hongyang, lead scientist with Shanghai AI Lab, said in the paper.

Modern autonomous driving systems integrate aspects of both the car industry and artificial intelligence, which include a series of tasks such as detection, tracking and mapping. Existing autonomous driving algorithms can be broadly divided into three categories.

The most common modular designs complete perception, prediction and planning tasks separately. While the development of each module is more flexible, such designs risk the loss of information across modules.

A more elegant design incorporates several tasks into one multitask learning framework, a practice that is more common at
Please, Log in or Register to view URLs content!
and China’s Xpeng. The design could easily take on additional tasks, thereby reducing the computational demands placed on onboard chips. At the same time, the algorithm can minimise coordination problems between tasks that might lead to cumulative errors.

The latest end-to-end systems – the large-scale
Please, Log in or Register to view URLs content!
– unite all tasks into a whole. The network uses raw sensor data – such as images or radar – as input and directly outputs the desired driving actions, such as steering, acceleration and braking.

Some AI scientists say the large-scale AI model could be the ultimate solution for autonomous driving. Since it is based on a more direct decision chain, the system could drastically reduce the chances of information errors, resulting in higher performance potentials.
UniAD’s algorithm performance was based on real-world scenario data sets from nuScenes, a large public data set collected on actual roads. It has served as a benchmark for many perception algorithms and autonomous driving systems.

According to the paper, UniAD received top marks in all tests using nuScenes. For instance, multi-object tracking accuracy was 20 per cent better than a previous best, and the error rate in motion forecasting and planning was reduced by 38 and 28 per cent, respectively.

While the promising results signal an advance in autonomous driving technology, the nuScenes data set is relatively small compared to real-world self-driving scenarios. That opens the possibility that there could be a considerable gap between UniAD’s test performance and its practical application.

Tesla’s FSD system has had a sluggish transition to
Please, Log in or Register to view URLs content!
use. Restrictions on road test conditions, data accumulation and algorithm training have made progress slow and costly. To expedite training,
Please, Log in or Register to view URLs content!
has built a virtual simulation space that includes all traffic elements and extreme corner cases.

The FSD system has seen numerous iterations in simulation training. As of May, the total distance driven for Tesla’s global fleet had exceeded 100 billion miles (161 billion km), and the driver behaviour data voluntarily provided continues to help refine the FSD algorithm.

But, according to its developers, the UniAD model’s more comprehensive design gives it the potential to become a next-generation autonomous driving technology. “Because of its full interpretability, safety, and continuous iteration across multiple modules, UniAD is the most promising end-to-end model for practical deployment,” Shanghai AI Lab’s Li said.
UniAD might also have a cost advantage. UniAD, which is based on two dimensional image input, outperformed other methods that were based on Light Detection and Ranging input. As a result, the driving algorithm can save significant amounts on hardware costs while still
Please, Log in or Register to view URLs content!
.

As the field of autonomous driving rapidly evolves, it remains highly competitive, with various companies racing to develop systems that they hope will be a game changer. The researchers behind the UniAD system say their technology deserves further exploration.
“We hope this work can shed some light on the target-driven design for an autonomous driving system, and provide a starting point for coordinating the many driving tasks,” Li said in the paper.
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
Yeah, I've been posting about their Ascend data center successes for a while now

Please, Log in or Register to view URLs content!
在论坛上,华为宣布将联合 26 家企业、科研院所与华为共同基于昇腾 AI 进行基础大模型与行业大模型应用创新。同时,华为携手伙伴联合发布昇腾 AI 大模型训推一体化解决方案,加速大模型在各行业应用落地,并有 23 家昇腾 AI 伙伴推出 AI 服务器、智能边缘与终端新品。

IT之家此前报道,胡厚崑在今日 2023 世界人工智能大会开幕式上发表演讲时表示,通过架构创新,华为昇腾 AI 集群效率已提升 10%。目前,集群规模从最初的 4000 卡集群扩展至 16000 卡,是业界首个万卡 AI 集群,拥有更快的训练速度和 30 天以上的稳定训练周期。

在发展生态方面,华为联合了 5700 + 鲲鹏 / 昇腾合作伙伴,以及硬件合作伙伴 30+,实现了国内大模型近一半创新使能,包括场景化系列 AI 硬件 100+,孵化 / 适配大模型 30+,鲲鹏 / 昇腾开发者 380 万 +。

在共建算力方面,华为已经在构建城市算力基础设施,帮助各地政府打造了 25 个昇腾人工智能计算中心。
Half of the LLMs in China use Huawei technology, 3.8 million Ascend/kunpeng developers

Huge Ascend AI clusters with up to 16000 cards

30+ days of stable training
 
Top