LLMs之Vicuna:《Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality》翻译与解读

LLMs之Vicuna:《Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality》翻译与解读

导读:作者提出了一个开源的聊天机器人Vicuna-13B。它是通过训练从ShareGPT收集的用户共享对话,然后在LLaMA基础模型上进行调整而产生的。根据初步的GPT-4评估,Vicuna-13B的质量达到了ChatGPT和Bard 90%的质量,超过其他开源模型如LLaMA和Alpaca。作者提出利用GPT-4作为评估工具来评估不同聊天机器人的有效性,通过它产生的答案和分数。尽管存在局限性,但这证明了自动化评估的潜力。Vicuna-13B的训练成本很低,大约只有300美元,采用了内存优化、多轮对话的改进方法,并通过Spot实例降低了成本。该模型的代码、参数和在线演示向公众开放。最后,作者强调Vicuna存在的限制,如在涉及推理和数学的任务方面存在问题,缺少安全优化。但它可以作为未来研究解决这些限制的开端。

>>训练Vicuna13B仅300美元=8张A100耗费1天+PyTorch FSDP:训练Vicuna-13B的成本约为300美元。

>> 基于ShareGPT收集7万段对话微调 LLaMA+GPT-4评估:训练ShareGPT收集用户共享对话数据集。使用 GPT-4 进行评估,发现 Vicuna-13B 的性能在超过90%的情况下实现了与ChatGPT和Bard相匹敌的能力;同时,在 90% 情况下都优于 LLaMA 和 Alpaca 等其他模型。

>> 基于Alpaca构建+内存优化(​梯度检查点gradient checkpointing+闪光注意力​flash attention​,最大上下文长度扩展到2048)+多轮对话+管理Spot实例(降低成本)+GPT4模型评估

(1)、Gradient Checkpointing技术,用于减少深度神经网络训练过程中的内存消耗。它的核心思想是将中间层的梯度计算延迟到后续阶段,从而减少需要存储的中间梯度的数量,以换取更低的内存需求(通过分段保存梯度来降低训练深度网络的内存消耗)。这种技术在长序列或深层网络中特别有用,因为它们往往需要更多的内存来存储中间梯度。
(2)、Flash Attention技术,带来的更长上下文能生成更高质量的Transformer模型,能显著提升Transformers的执行效率和性能,特别适用于长序列任务。FlashAttention通过分块计算和I/O优化,实现了内存高效且速度快的注意力机制,广泛适用于注意力计算密集型任务。论文针对Transformer模型在长序列上的计算复杂度和内存需求问题,提出了一种IO感知的精确注意力算法FlashAttention。该算法利用划分块状减少了GPU高带宽内存(HBM)和GPU片上SRAM之间的读写次数,从而提高了训练速度。

(3)、SkyPilot managed spot技术,通过管理Spot实例,SkyPilot可以帮助用户以低成本高效进行长时间计算作业(如深度学习训练),同时提供足够的可用性和弹性。它可以自动从抢占中恢复,并在不同的地区和云服务商之间寻找可用的Spot资源以提高可用性。这个功能可以显著降低长时间运行任务的成本。

>> 提供了LLM聊天机器人的平台FastChat:提供了一个用于训练、服务和评估基于大语言模型的聊天机器人的开放平台FastChat。

目录

相关内容

NLP之LLMs:《Zeno Chatbot Report》的翻译与解读—CMU副教授详测七款个类ChatGPT大模型(GPT-2、LLaMa、Alpaca、Vicuna、MPT-Chat、Cohere Command和ChatGPT)

核心技术

Flash Attention技术来源

实战案例

LLMs:在Linux服务器系统上实Vicuna-7B本地化部署(基于facebookresearch的GitHub)进行模型权重合并(llama-7b模型与delta模型权重)、模型部署且实现模型推

《Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality》翻译与解读

How Good is Vicuna? Vicuna-13B的性能有多好?

Online Demo在线演示

Overview概述

ShareGPT收集7万段对话、8张A100耗费1天+PyTorch FSDP、GPT-4模型评估

Table 1. Comparison between several notable models

Training训练(13B降低到300美元):基于Alpaca构建+内存优化(​梯度检查点gradient checkpointing+闪光注意力​flash attention​,最大上下文长度扩展到2048)+多轮对话+通过Spot实例降低成本

Serving服务:分布式工作节点+灵活添加GPU节点

How To Evaluate a Chatbot?如何评估聊天机器人?——提出了一种基于GPT-4的评估框架来自动评估聊天机器人的性能

Table 2. Total Scores Assessed by GPT-4.

Limitations局限性—不擅长涉及推理或数学

Release发行

License许可证

The Team团队

Students (alphabetical order):

Acknowledgment致谢

Citation


相关内容

NLP之LLMs:《Zeno Chatbot Report》的翻译与解读—CMU副教授详测七款个类ChatGPT大模型(GPT-2、LLaMa、Alpaca、Vicuna、MPT-Chat、Cohere Command和ChatGPT)

NLP之LLMs:《Zeno Chatbot Report》的翻译与解读—CMU副教授详测七款个类ChatGPT大模型(GPT-2、LLaMa、Alpaca、Vicuna、MPT-Chat、Coher_一个处女座的程序猿的博客-CSDN博客

核心技术

Flash Attention技术来源

该技术源于论文《FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness》。

时间

2023年3月30日

地址

博客文章:Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality | LMSYS Org

GitHub:GitHub - lm-sys/FastChat: An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.

作者

The Vicuna Team

这是与来自多个机构的合作者的共同努力,包括加州大学伯克利分校、CMU、斯坦福大学、加州大学圣地亚哥分校和MBZUAI。

UC Berkeley+CMU+Stanford

实战案例

LLMs:在Linux服务器系统上实Vicuna-7B本地化部署(基于facebookresearch的GitHub)进行模型权重合并(llama-7b模型与delta模型权重)、模型部署且实现模型推

https://yunyaniu.blog.csdn.net/article/details/131016620

《Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality》翻译与解读

Vicuna (generated by stable diffusion 2.1)

We introduce Vicuna-13B, an open-source chatbot trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. Preliminary evaluation using GPT-4 as a judge shows Vicuna-13B achieves more than 90%* quality of OpenAI ChatGPT and Google Bard while outperforming other models like LLaMA and Stanford Alpaca in more than 90%* of cases. The cost of training Vicuna-13B is around $300. The code and weights, along with an online demo, are publicly available for non-commercial use.

Vicuna (generated by stable diffusion 2.1)

我们推出了Vicuna-13B,这是一个通过在LLaMA上调整ShareGPT收集的用户共享对话进行训练的开源聊天机器人。利用GPT-4作为评判,初步评估显示Vicuna-13B达到OpenAI ChatGPT和谷歌Bard 90%*的质量,在90%*的情况下超过LLaMA和斯坦福Alpaca等其他模型。训练Vicuna-13B的成本约为300美元。代码、参数以及在线演示均向公众开放用于非商业用途。

*According to a fun and non-scientific evaluation with GPT-4. Further rigorous evaluation is needed.

*根据GPT-4的有趣和非科学评估。需要进一步严格评估。

How Good is Vicuna? Vicuna-13B的性能有多好

 Figure 1. Relative Response Quality Assessed by GPT-4*

After fine-tuning Vicuna with 70K user-shared ChatGPT conversations, we discover that Vicuna becomes capable of generating more detailed and well-structured answers compared to Alpaca (see examples below), with the quality on par with ChatGPT.

在对Vicuna进行7万用户共享ChatGPT对话的调整后,我们发现与Alpaca相比,Vicuna能够生成更详细和结构更好的答案(见下例),质量与ChatGPT相当。

However, evaluating chatbots is never a simple task. With recent advancements in GPT-4, we are curious whether its capabilities have reached a human-like level that could enable an automated evaluation framework for benchmark generation and performance assessments. Our initial finding indicates that GPT-4 can produce highly consistent ranks and detailed assessment when comparing chatbots’ answers (see above example of GPT-4 judgment). Preliminary evaluations based on GPT-4, summarized in Figure 1, show that Vicuna achieves 90%* capability of Bard/ChatGPT. While this proposed framework shows a potential to automate chatbot assessment, it is not yet a rigorous approach. Building an evaluation system for chatbots remains an open question requiring further research. More details are provided in the evaluation section.

然而,评估聊天机器人从来不是一件简单的任务。随着GPT-4的最新进展,我们好奇其能力是否达到了人类级别,能否实现基准生成和性能评估的自动化评估框架。我们的初步发现表明,GPT-4可以产生高度一致的排名和详细的评估,以比较聊天机器人的答案(见GPT-4判断的上例)。基于GPT-4的初步评估总结在图1中,显示Vicuna达到Bard/ChatGPT的90%*能力。虽然这种提议的框架显示出自动化评估聊天机器人的潜力,但这还不是一个严格的方法。建立聊天机器人的评估系统仍然是一个需要进一步研究的开放问题。更多详情在评估部分提供。

Online Demo在线演示

Overview概述

 Figure 2. Workflow Overview

ShareGPT收集7万段对话、8张A100耗费1天+PyTorch FSDP、GPT-4模型评估

The rapid advancement of large language models (LLMs) has revolutionized chatbot systems, resulting in unprecedented levels of intelligence as seen in OpenAI's ChatGPT. However, despite its impressive performance, the training and architecture details of ChatGPT remain unclear, hindering research and open-source innovation in this field. Inspired by the Meta LLaMA and Stanford Alpaca project, we introduce Vicuna-13B, an open-source chatbot backed by an enhanced dataset and an easy-to-use, scalable infrastructure. By fine-tuning a LLaMA base model on user-shared conversations collected from ShareGPT.com, Vicuna-13B has demonstrated competitive performance compared to other open-source models like Stanford Alpaca. This blog post provides a preliminary evaluation of Vicuna-13B's performance and describes its training and serving infrastructure. We also invite the community to interact with our online demo to test the capabilities of this chatbot.

大规模语言模型(LLM)的快速发展彻底改变了聊天机器人系统,表现出前所未有的智能,如OpenAI的ChatGPT。然而,尽管性能令人印象深刻,ChatGPT的训练和架构细节仍不清楚,阻碍了该领域的研究和开源创新。受Meta LLaMA和斯坦福Alpaca项目的启发,我们推出了Vicuna-13B,这是一个由增强数据集和易于使用的可扩展基础设施支持的开源聊天机器人。通过在LLaMA基础模型上调整从ShareGPT.com收集的用户共享对话,Vicuna-13B已经展示出与其他开源模型(如斯坦福Alpaca)相媲美的性能。本博客文章对Vicuna-13B的性能进行初步评估,并描述了其训练和服务基础设施。我们还邀请社区与我们的在线演示互动,测试此聊天机器人的能力。

Figure 2 provides an overview of our work. To begin, we collected around 70K conversations from ShareGPT.com, a website where users can share their ChatGPT conversations. Next, we enhanced the training scripts provided by Alpaca to better handle multi-round conversations and long sequences. The training was done with PyTorch FSDP on 8 A100 GPUs in one day. For serving the demo, we implemented a lightweight distributed serving system. We conducted a preliminary evaluation of the model quality by creating a set of 80 diverse questions and utilizing GPT-4 to judge the model outputs. To compare two different models, we combine the outputs from each model into a single prompt for each question. The prompts are then sent to GPT-4, which assesses which model provides better responses. A detailed comparison of LLaMA, Alpaca, ChatGPT, and Vicuna is shown in Table 1 below.

图2概述了我们的工作。首先,我们从ShareGPT.com网站收集了约7万段对话,用户可以在该网站上共享他们的ChatGPT对话。其次,我们改进了Alpaca提供的训练脚本,更好地处理多轮对话和长序列。训练在8个A100 GPU上一天内完成,使用PyTorch FSDP。为了演示服务,我们实现了一个轻量级的分布式服务系统。我们通过创建80个多样化的问题,并利用GPT-4判断模型输出来对模型质量进行初步评估。为了比较两个不同的模型,我们将每个模型的输出组合在每个问题的单个提示中。然后将提示发送给GPT-4,它会评估哪个模型提供更好的回答。LLaMA,Alpaca,ChatGPT和Vicuna的详细比较见下表1。

Table 1. Comparison between several notable models

Model NameLLaMAAlpacaVicunaBard/ChatGPT
DatasetPublicly available datasets
(1T token)
Self-instruct from davinci-003 API
(52K samples)
User-shared conversations
(70K samples)
N/A
Training codeN/AAvailableAvailableN/A
Evaluation metricsAcademic benchmarkAuthor evaluationGPT-4 assessmentMixed
Training cost
(7B)
82K GPU-hours$500 (data) + $100 (training)$140 (training)N/A
Training cost
(13B)
135K GPU-hoursN/A$300 (training)N/A

Training训练(13B降低到300美元):基于Alpaca构建+内存优化(​
梯度检查点gradient checkpointing+闪光注意力​flash attention​,
最大上下文长度扩展到2048)+多轮对话+通过Spot实例降低成本

Vicuna is created by fine-tuning a LLaMA base model using approximately 70K user-shared conversations gathered from ShareGPT.com with public APIs. To ensure data quality, we convert the HTML back to markdown and filter out some inappropriate or low-quality samples. Additionally, we divide lengthy conversations into smaller segments that fit the model's maximum context length.

Vicuna是通过调整LLaMA基础模型来创建的,使用从ShareGPT.com收集的约7万段用户共享对话。为确保数据质量,我们将HTML转换回markdown,并过滤掉一些不适当或低质量的样本。此外,我们将较长的对话分成较小的段,以符合模型的最大上下文长度。

Our training recipe builds on top of Stanford’s alpaca with the following improvements.

Memory Optimizations: To enable Vicuna's understanding of long context, we expand the max context length from 512 in alpaca to 2048, which substantially increases GPU memory requirements. We tackle the memory pressure by utilizing gradient checkpointing and flash attention.

Multi-round conversations: We adjust the training loss to account for multi-round conversations and compute the fine-tuning loss solely on the chatbot's output.

Cost Reduction via Spot Instance: The 40x larger dataset and 4x sequence length for training poses a considerable challenge in training expenses. We employ SkyPilot managed spot to reduce the cost by leveraging the cheaper spot instances with auto-recovery for preemptions and auto zone switch. This solution slashes costs for training the 7B model from $500 to around $140 and the 13B model from around $1K to $300.

我们的训练方法基于斯坦福大学Stanford’s alpaca构建,具有以下改进。

内存优化:为了使Vicuna理解长上下文,我们将alpaca中的最大上下文长度从512扩展到2048,这大大增加了GPU内存需求。我们通过使用梯度检查点gradient checkpointing和闪光注意力flash attention来解决内存压力。

多轮对话:我们调整训练损失,以考虑多轮对话,并仅根据聊天机器人的输出计算调整损失。

通过Spot实例降低成本:数据集扩大40倍序列长度增加4倍的训练会带来相当大的训练费用挑战。我们采用SkyPilot托管的Spot实例SkyPilot managed spot,利用更便宜的Spot实例与自动恢复预防和自动区域切换来降低成本。该解决方案将7B模型的训练成本从500美元降低到约140美元,13B模型的训练成本从约1,000美元降低到300美元

Serving服务分布式工作节点+灵活添加GPU节点

We build a serving system that is capable of serving multiple models with distributed workers. It supports flexible plug-in of GPU workers from both on-premise clusters and the cloud. By utilizing a fault-tolerant controller and managed spot feature in SkyPilot, this serving system can work well with cheaper spot instances from multiple clouds to reduce the serving costs. It is currently a lightweight implementation and we are working on integrating more of our latest research into it.

我们构建了一个能够使用分布式工作节点服务多个模型的服务系统。它支持从本地集群和云中灵活添加GPU工作节点。通过利用容错控制器和SkyPilot中的托管Spot功能,此服务系统可以与来自多个云的更便宜的Spot实例很好地配合使用,以降低服务成本。这目前是一个轻量级实现,我们正在努力将我们最新的研究成果集成进去

How To Evaluate a Chatbot?如何评估聊天机器人?——提出了一种基于GPT-4的评估框架来自动评估聊天机器人的性能

 Figure 3. Response Comparison Assessed by GPT-4

Evaluating AI chatbots is a challenging task, as it requires examining language understanding, reasoning, and context awareness. With AI chatbots becoming more advanced, current open benchmarks may no longer suffice. For instance, the evaluation dataset used in Stanford’s Alpaca, self-instruct, can be effectively answered by SOTA chatbots, making it difficult for humans to discern differences in performance. More limitations include training/test data contamination and the potentially high cost of creating new benchmarks. To tackle these issues, we propose an evaluation framework based on GPT-4 to automate chatbot performance assessment.

评估AI聊天机器人是一项具有挑战性的任务,因为它需要检查语言理解推理上下文意识。随着AI聊天机器人变得更加先进,目前的开放基准可能不再足够。例如,斯坦福大学Alpaca使用的评估数据集self-instruct可以被当前最先进的聊天机器人有效回答,这使人类难以 辨别性能差异。更多限制包括训练/测试数据污染和潜在的创建新基准的高成本。为解决这些问题,我们提出了一种基于GPT-4的评估框架来自动评估聊天机器人的性能。

First, we devised eight question categories, such as Fermi problems, roleplay scenarios, and coding/math tasks, to test various aspects of a chatbot's performance. Through careful prompt engineering, GPT-4 is able to generate diverse, challenging questions that baseline models struggle with. We select ten questions per category and collect answers from five chatbots: LLaMA, Alpaca, ChatGPT, Bard, and Vicuna. We then ask GPT-4 to rate the quality of their answers based on helpfulness, relevance, accuracy, and detail. We discover that GPT-4 can produce not only relatively consistent scores but also detailed explanations on why such scores are given (detailed examples link). However, we also notice that GPT-4 is not very good at judging coding/math tasks.

首先,我们设计了八个问题类别,如费米问题、角色扮演场景和编码/数学任务,以测试聊天机器人性能的各个方面。通过精心设计提示,GPT-4能够生成基线模型难以应对的多样化和具有挑战性的问题。我们从五个聊天机器人中选择每个类别十个问题的答案:LLaMA、Alpaca、ChatGPT、Bard和Vicuna。然后我们要求GPT-4根据有用性、相关性、准确性和细节评价它们的答案质量。我们发现GPT-4不仅可以产生相对一致的分数,而且能够详细解释为什么给出这样的分数(详细示例链接)。然而,我们也注意到GPT-4在判断编码/数学任务方面不是很好

Figure 3 displays the comparison results between all baselines and Vicuna. GPT-4 prefers Vicuna over state-of-the-art open-source models (LLaMA, Alpaca) in more than 90% of the questions, and it achieves competitive performance against proprietary models (ChatGPT, Bard). In 45% of the questions, GPT-4 rates Vicuna's response as better or equal to ChatGPT's. As GPT-4 assigns a quantitative score to each response on a scale of 10, we calculate the total score for each (baseline, Vicuna) comparison pair by adding up the scores obtained by each model on 80 questions. As shown in Table 2, Vicuna’s total score is 92% of ChatGPT’s. Despite recent advancements, these chatbots still face limitations, such as struggling with basic math problems or having limited coding ability.

图3显示了所有基准线和Vicuna之间的比较结果。在80%的问题中,GPT-4更喜欢Vicuna而不是最先进的开源模型(LLaMA,Alpaca),并达到专有模型(ChatGPT,Bard)的竞争性能。在45%的问题中,GPT-4将Vicuna的回答评为优于或等于ChatGPT的回答。由于GPT-4在10点量表上为每个回答分配一个定量得分,我们通过将每个模型在80个问题上获得的得分相加来计算每个(基准,Vicuna)比较对的总得分。如表2所示,Vicuna的总得分是ChatGPT的92%。尽管最近有所进展,但这些聊天机器人仍面临一些限制,如难以应对基本的数学问题或具有有限的编码能力

While this proposed evaluation framework demonstrates the potential for assessing chatbots, it is not yet a rigorous or mature approach, as large language models are prone to hallucinate. Developing a comprehensive, standardized evaluation system for chatbots remains an open question requiring further research.

虽然此提出的评估框架展示了评估聊天机器人的潜力,但由于大语言模型容易产生幻觉,所以这还不是一个严格或成熟的方法。开发全面标准化的聊天机器人评估系统仍然是一个需要进一步研究的开放问题。

Table 2. Total Scores Assessed by GPT-4.

BaselineBaseline ScoreVicuna Score
LLaMA-13B513.0694.0
Alpaca-13B583.0704.0
Bard664.0655.5
ChatGPT693.0638.0

Limitations局限性—不擅长涉及推理数学

We have noticed that, similar to other large language models, Vicuna has certain limitations. For instance, it is not good at tasks involving reasoning or mathematics, and it may have limitations in accurately identifying itself or ensuring the factual accuracy of its outputs. Additionally, it has not been sufficiently optimized to guarantee safety or mitigate potential toxicity or bias. To address the safety concerns, we use the OpenAI moderation API to filter out inappropriate user inputs in our online demo. Nonetheless, we anticipate that Vicuna can serve as an open starting point for future research to tackle these limitations.

我们注意到,与其他大型语言模型一样,Vicuna也存在某些限制。例如,它不擅长涉及推理数学的任务,并且在准确识别自己或确保其输出的事实准确性方面可能存在限制。此外,它还没有得到足够的优化以确保安全性或减轻潜在的 toxicity或偏见。为了解决安全问题,我们在在线演示中使用OpenAI调解API过滤掉不适当的用户输入。尽管如此,我们预计Vicuna可以作为未来研究解决这些限制的开放起点。

Release发行

In our first release, we will share the training, serving, and evaluation code on a GitHub repo: https://github.com/lm-sys/FastChat. We also released the Vicuna-13B model weights, please find the instructions here. There is no plan to release the dataset. Join our Discord server and follow our Twitter to get the latest updates.

在我们的首次发布中,我们将在GitHub repo:https://github.com/lm-sys/FastChat上共享训练,服务和评估代码。我们还发布了Vicuna-13B模型权重,请在这里找相关说明。暂无计划发布数据集。加入我们的Discord服务器并关注我们的Twitter以获取最新动态。

License许可证

The online demo is a research preview intended for non-commercial use only, subject to the model License of LLaMA, Terms of Use of the data generated by OpenAI, and Privacy Practices of ShareGPT. Please contact us If you find any potential violation.\ The code is released under the Apache License 2.0.

在线演示仅供非商业用途,受LLaMA模型许可证、OpenAI生成的数据使用条款和ShareGPT的隐私实践的约束。如果发现任何潜在违规行为,请联系我们。\代码根据Apache许可证2.0版发布。

The Team团队

This is a joint effort with collaborators from multiple institutions, including UC Berkeley, CMU, Stanford, UC San Diego, and MBZUAI.

这是与来自多个机构的合作者的共同努力,包括加州大学伯克利分校、CMU、斯坦福大学、加州大学圣地亚哥分校和MBZUAI。

Students (alphabetical order):

Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang

Advisors (alphabetical order):

Joseph E. Gonzalez, Ion Stoica, Eric P. Xing

学生(按字母顺序):

Wei-Lin Chiang,Zhuohan Li,Zi Lin,Ying Sheng,Zhanghao Wu,Hao Zhang,Lianmin Zheng,Siyuan Zhuang,Yonghao Zhuang

顾问(按字母顺序):

Joseph E. Gonzalez,Ion Stoica,Eric P. Xing

Acknowledgment致谢

We would like to thank Xinyang Geng, Hao Liu, and Eric Wallace from BAIR; Xuecheng Li, and Tianyi Zhang from Stanford Alpaca team for their insightful discussion and feedback; Qirong Ho from MBZUAI for providing support on the serving cluster. Please check out a blog post from BAIR about a concurrent effort on their chatbot, Koala.

我们要感谢BAIR的Xinyang Geng,Hao Liu和Eric Wallace;斯坦福Alpaca团队的Xuecheng Li和Tianyi Zhang提供的精辟讨论和反馈;MBZUAI的Qirong Ho为服务集群提供的支持。请查看BAIR关于他们的聊天机器人Koala的同期工作的博客文章。

Citation

@misc{vicuna2023,title = {Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90\%* ChatGPT Quality},url = {https://lmsys.org/blog/2023-03-30-vicuna/},author = {Chiang, Wei-Lin and Li, Zhuohan and Lin, Zi and Sheng, Ying and Wu, Zhanghao and Zhang, Hao and Zheng, Lianmin and Zhuang, Siyuan and Zhuang, Yonghao and Gonzalez, Joseph E. and Stoica, Ion and Xing, Eric P.},month = {March},year = {2023}
}

本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若转载,请注明出处:http://www.rhkb.cn/news/9739.html

如若内容造成侵权/违法违规/事实不符,请联系长河编程网进行投诉反馈email:809451989@qq.com,一经查实,立即删除!

相关文章

2022掉队的“蔚小理”,按下了兔年加速键

配图来自Canva可画 进入2023年,各大车企又展开了新一轮的“竞速”。尽管1月份汽车整体销量出现了“阴跌”,但从各路车企发布的销量目标来看,车企对于2023依旧保持着较高的信心和预期。在一众车企中,以“蔚小理”为代表的新势力们…

推荐兔年春节的SVG模版

宝藏模版 往期推荐(点击阅读): 🤡推荐几个莫名好玩的SVG模版🤡http://mp.weixin.qq.com/s?__bizMzk0NDIyMjU0OQ&mid2247499869&idx1&sn04e2bf8de065795489bf9c391a249b5d&chksmc3257867f452f17114…

学习css3,使用代码实现一根心爱的二踢脚

2023春节快到了,虽然还在上班,但心情早已开始激动,感叹2022终将过去,喜迎2023兔年吧。让我以激动的心情,利用所学css3代码知识,实现一根心爱的二踢脚吧。 目录 1、实现思路 2、实现二踢脚的主体圆柱部分…

C语言烟花代码—兔年顶呱呱

废话不多说&#xff0c;直接上码 代码&#xff0c;歌曲&#xff0c;烟花图片都在压缩包里 链接: https://pan.baidu.com/s/1_459s0fFCAX1DcQa_BnHMQ?pwdqw12 提取码: qw12 要看效果的也可以看我抖音: 1210246294 #include<stdio.h> #include<easyx.h> #inclu…

Python跳跳兔小游戏源代码,兔年必玩小游戏,兔年大吉

Python跳跳兔小游戏源代码&#xff0c;兔年必玩小游戏&#xff0c;兔年大吉&#xff0c;小兔子跳跳&#xff0c;按空格键向上跳跃&#xff0c;按键盘方向键进行左右移动&#xff0c;以避开飞弹&#xff0c;以防被炸&#xff0c;还可以捡到火箭道具哦。 完整程序下载地址&#…

小程序源码:2022虎年新UI春节送祝福

这是一款网友用以前小编发过的一款端午送祝福改的一款小程序 里面的背景图包括祝福语都已经修改成与虎年相关的内容了 总体来说找的背景图还是可以的,不过有些地方和细节小编也给完善了一下 然后小编测试的时候发现还没有流量主,所以小编也给加了几个流量主进去 到时候大家直…

2023兔年头像框制作生成小程序源码+多模板/抓住红利

正文: 2023了&#xff0c;到了兔年那么也该换新的头像框拉&#xff0c;今天给大家带来兔年的头像框制作小程序&#xff0c;支持一键获取头像支持用户上传头像&#xff0c;丰富的模板&#xff0c;不只于兔年头像框还有其它的&#xff0c;早点给大家发出来也好让大家早点抓住这波…

头像生成器微信小程序代码

微信小程序需要wxml&#xff0c;wxss&#xff0c;js三种类型代码 1&#xff0c;wxml <view class"container"><view class"title">兔年头像生成器</view><view class"description">选择一张图片&#xff0c;生成你的兔…

Python生成兔年祝福词云图,快拿去发给小姐姐吧!祝大家兔年大吉,桃花运多多

今年是兔年&#xff0c;所以今年我们拜年主题就是以兔为主&#xff0c;为了能够在兔年拜年做好准备祝福语可是少不了&#xff0c;发太多的文字已经让人们感到厌烦&#xff0c;所以不如利用自己所学只是做一张兔年的词云图&#xff0c;来去给小姐姐们拜个年。 话不多说&#xff…

2023最新兔年新春好友送祝福小程序源码+文案很多

正文: 去年的时候也发过虎年的送祝福小程序&#xff0c;这不马上2023兔年了所以给大家发兔年的送祝福小程序&#xff0c;早点发也好让大家早点抓住新春的流量福利。 程序: wwxras.lanzoum.com/iPWHz0l7t5ra 图片:

【OpenVI】AIGC纪元,兔年AI绘画实践

团队模型、论文、博文、直播合集&#xff0c;点击此处浏览 一、背景 AIGC指的是AI Generated Content&#xff0c;即由AI创作的内容&#xff0c;是继UGC(User Generated Content用户创造内容&#xff0c;如抖音b站等平台)、PGC(Professional Generated Content专业生产内容&…

2023 兔年春节倒计时微信小程序源码

马上也到2023年兔年了 增加一下喜庆,其实这种小程序还是有流量的 每天看一看倒计时,数着过年 另外还有一个简单的绘画功能

2023兔年新春好友送祝福微信小程序源码

源码简介&#xff1a; 马上2023兔年了所以站长给大家发兔年的送祝福小程序 早点发也好让大家早点抓住新春的流量福利 好了不多说,感兴趣的抓紧下载吧! 原文链接&#xff1a;2023兔年新春好友送祝福微信小程序源码-主题模板站 源码截图&#xff1a;

2023兔年新年春节倒计时微信小程序源码

简介&#xff1a; 马上也到2023年兔年了 给大家分享一个春节倒计时的小程序 增加一下喜庆,其实这种小程序还是有流量的 每天看一看倒计时,数着过年 另外还有一个简单的绘画功能 下方是小编的测试演示图! https://www.skpan.cn/bBibst11KsJ 图片&#xff1a;

2023兔年春节倒计时小程序源码

源码简介&#xff1a; 马上到2023年兔年了 站长给大家分享一个春节倒计时的小程序 增加一下喜庆,其实这种小程序还是有流量的 每天看一看倒计时,数着过年 另外还有一个简单的绘画功能 源码截图&#xff1a;

这么好看的客服组件,还是觉得接入ChatGPT,把它放在博客中那就完美了

我们在使用ChatGPT时&#xff0c;它的返回方式是采用流式回复。感觉这个效果不错。之前做的全都是的等全部结果请求完成&#xff0c;再一次性返回给用户。今天就通过流式回复的效果重写之前的程序。 前端 前端采用一个网页版的客服组件&#xff0c;整体的效果如下&#xff1a;…

C#接入ChatGPT编程实际应用

一&#xff0c;介绍ChatGPT的API可以在前端&#xff0c;以及一些后端进行API的接入&#xff0c;本篇主要介绍基于ChatGPT的API的C#接入研究。 学习更多的科技知识可以来知新文学 最火的莫过于ChatGPT的相关讨论&#xff0c;这个提供了非常强大的AI处理&#xff0c;并且整个平…

Java快速接入Chatgpt简单案例

Chatgpt介绍 ChatGPT是一种由OpenAI训练的大型语言模型。它的原理是基于Transformer架构&#xff0c;通过预训练大量文本数据来学习如何生成人类可读的文本&#xff0c;然后通过接受输入并生成输出来实现对话。 ChatGPT的用途非常广泛&#xff0c;可以用于自然语言处理&#xf…

SpringBoot 接入chatGPT API

SpringBoot 接入chatGPT API 一、准备工作二、补全接口示例三、申请API-KEY**四、JavaScript调用API**五、SpringBoot整合ChatGPT六、使用curl模拟请求ChatGPT平台已经为技术提供了一个入口了,作为一个Java程序员,我们第一时间想到的就是快速开发一个应用,接入ChatGPT的接口…