mysql8 C++源码中创建表函数,表字段最大数量限制,表行最大存储限制

在 MySQL 8 的 C++ 源码中,表的最大字段数量限制体现在 MAX_FIELDS 宏定义中。这个宏定义了表中可以拥有的最大字段数量。

代码中的体现

mysql_prepare_create_table 函数中,有以下代码段检查表的字段数量是否超过最大限制:

cpp

if (alter_info->create_list.elements > MAX_FIELDS) {my_error(ER_TOO_MANY_FIELDS, MYF(0));return true;
}

这里,alter_info->create_list.elements 表示表中字段的数量,MAX_FIELDS 是最大允许的字段数量。

MAX_FIELDS 的定义

MAX_FIELDS 的定义通常在 MySQL 的头文件中,例如 include/mysql_com.h 或其他相关头文件。在 MySQL 8 中,MAX_FIELDS 的值为 4096

源码位置

在 MySQL 8 的源码中,mysql_prepare_create_table 函数位于 sql/sql_table.cc 文件中。这个文件包含了与表创建和修改相关的许多函数。

总结

  • 最大字段数量:MySQL 8 中表的最大字段数量为 4096

  • 代码中的体现:在 mysql_prepare_create_table 函数中通过 MAX_FIELDS 宏检查表的字段数量。

MySQL 中,每行记录的最大长度限制为 65,535 字节。这个限制是由 MySQL 的内部实现决定的,而不是由某个特定的源码文件或函数直接定义的。这个限制适用于所有存储引擎,包括 InnoDB 和 MyISAM。

限制的体现

在 MySQL 8 的源码中,这个限制通常在多个地方体现,特别是在处理表的创建和数据插入时。以下是一些关键点:

  1. 表创建时的验证: 在 mysql_prepare_create_table 函数中,虽然没有直接的代码检查每行记录的长度,但这个限制会在表创建时通过存储引擎的接口进行验证。

  2. 存储引擎的限制: 每个存储引擎都有自己的限制。例如,InnoDB 存储引擎在处理表创建时会检查每行记录的长度是否超过最大限制。

  3. 数据插入时的验证: 在插入数据时,MySQL 会检查每行记录的长度是否超过 65,535 字节。如果超过,会报错。

示例代码

以下是一个简化的示例,展示如何在插入数据时检查每行记录的长度:

cpp

if (reclength > file->max_record_length()) {my_error(ER_TOO_BIG_ROWSIZE, MYF(0),static_cast<long>(file->max_record_length()));return true;
}

在这个代码片段中,reclength 是计算的每行记录的长度,file->max_record_length() 是存储引擎允许的最大记录长度。如果 reclength 超过 file->max_record_length(),则报错。

具体实现

在 MySQL 源码中,file->max_record_length() 通常由存储引擎的实现提供。例如,在 InnoDB 存储引擎中,这个值被定义为 65,535 字节

C++源码

// Prepares the table and key structures for table creation.
bool mysql_prepare_create_table(THD *thd, const char *error_schema_name, const char *error_table_name,HA_CREATE_INFO *create_info, Alter_info *alter_info, handler *file,bool is_partitioned, KEY **key_info_buffer, uint *key_count,FOREIGN_KEY **fk_key_info_buffer, uint *fk_key_count,FOREIGN_KEY *existing_fks, uint existing_fks_count,const dd::Table *existing_fks_table, uint fk_max_generated_name_number,int select_field_count, bool find_parent_keys) {DBUG_TRACE;/*Validation of table properties.*/LEX_STRING *connect_string = &create_info->connect_string;if (connect_string->length != 0 &&connect_string->length > CONNECT_STRING_MAXLEN &&(system_charset_info->cset->charpos(system_charset_info, connect_string->str,(connect_string->str + connect_string->length),CONNECT_STRING_MAXLEN) < connect_string->length)) {my_error(ER_WRONG_STRING_LENGTH, MYF(0), connect_string->str, "CONNECTION",CONNECT_STRING_MAXLEN);return true;}LEX_STRING *compress = &create_info->compress;if (compress->length != 0 && compress->length > TABLE_COMMENT_MAXLEN &&system_charset_info->cset->charpos(system_charset_info, compress->str, compress->str + compress->length,TABLE_COMMENT_MAXLEN) < compress->length) {my_error(ER_WRONG_STRING_LENGTH, MYF(0), compress->str, "COMPRESSION",TABLE_COMMENT_MAXLEN);return true;}LEX_STRING *encrypt_type = &create_info->encrypt_type;if (encrypt_type->length != 0 &&encrypt_type->length > TABLE_COMMENT_MAXLEN &&system_charset_info->cset->charpos(system_charset_info, encrypt_type->str,encrypt_type->str + encrypt_type->length,TABLE_COMMENT_MAXLEN) < encrypt_type->length) {my_error(ER_WRONG_STRING_LENGTH, MYF(0), encrypt_type->str, "ENCRYPTION",TABLE_COMMENT_MAXLEN);return true;}// Validate table comment stringstd::string invalid_sub_str;if (is_invalid_string({create_info->comment.str, create_info->comment.length},system_charset_info, invalid_sub_str)) {my_error(ER_COMMENT_CONTAINS_INVALID_STRING, MYF(0), "table",(std::string(error_schema_name) + "." + std::string(error_table_name)).c_str(),system_charset_info->csname, invalid_sub_str.c_str());return true;}if (validate_comment_length(thd, create_info->comment.str, &create_info->comment.length,TABLE_COMMENT_MAXLEN, ER_TOO_LONG_TABLE_COMMENT, error_table_name)) {return true;}if (alter_info->create_list.elements > MAX_FIELDS) {my_error(ER_TOO_MANY_FIELDS, MYF(0));return true;}/*Checks which previously were done during .FRM creation.TODO: Check if the old .FRM limitations still make sensewith the new DD.*//* Fix this when we have new .frm files;  Current limit is 4G rows (QQ) */constexpr ulonglong u32max = UINT_MAX32;if (create_info->max_rows > UINT_MAX32) {// Values larger than uint32_max are capped to uint32_max.// Emit a warning about this.push_warning_printf(thd, Sql_condition::SL_WARNING, ER_VALUE_OUT_OF_RANGE,ER_THD(thd, ER_VALUE_OUT_OF_RANGE), "max_rows",create_info->max_rows, 0ULL, u32max, u32max);create_info->max_rows = UINT_MAX32;}if (create_info->min_rows > UINT_MAX32) {// Values larger than uint32_max are capped to uint32_max.// Emit a warning about this.push_warning_printf(thd, Sql_condition::SL_WARNING, ER_VALUE_OUT_OF_RANGE,ER_THD(thd, ER_VALUE_OUT_OF_RANGE), "min_rows",create_info->min_rows, 0ULL, u32max, u32max);create_info->min_rows = UINT_MAX32;}if (create_info->row_type == ROW_TYPE_DYNAMIC)create_info->table_options |= HA_OPTION_PACK_RECORD;/*Prepare fields, which must be done before callingadd_functional_index_to_create_list(). The reason is thatprepare_create_field() sets several properties of all Create_fields, such ascharacter set. We need the character set in order to get the correctdisplay width for each Create_field, which is in turn needed to resolve thecorrect data type/length for each hidden generated column added byadd_functional_index_to_create_list().*/int select_field_pos = alter_info->create_list.elements - select_field_count;create_info->null_bits = 0;int field_no = 0;Create_field *sql_field;List_iterator<Create_field> it(alter_info->create_list);for (; (sql_field = it++); field_no++) {if (prepare_create_field(thd, error_schema_name, error_table_name,create_info, &alter_info->create_list,&select_field_pos, file, sql_field, field_no))return true;}// Go through all functional key parts. For each functional key part, resolve// the expression and add a hidden generated column to the create list.for (Key_spec *key : alter_info->key_list) {if (key->type == KEYTYPE_FOREIGN) continue;for (size_t j = 0; j < key->columns.size(); ++j) {Key_part_spec *key_part_spec = key->columns[j];// In the case of procedures, the Key_part_spec may both have an// expression and a field name assigned to it. But the hidden generated// will not exist in the create list, so we will have to add it.if (!key_part_spec->has_expression() ||(key_part_spec->get_field_name() != nullptr &&column_exists_in_create_list(key_part_spec->get_field_name(),alter_info->create_list))) {continue;}Create_field *new_create_field = add_functional_index_to_create_list(thd, key, alter_info, key_part_spec, j, create_info);if (new_create_field == nullptr) {return true;}// Call prepare_create_field on the Create_field that was added by// add_functional_index_to_create_list().assert(is_field_for_functional_index(new_create_field));if (prepare_create_field(thd, error_schema_name, error_table_name,create_info, &alter_info->create_list,&select_field_pos, file, new_create_field,++field_no)) {return true;}}}// Now that we have all the Create_fields available, calculate the offsets// for each column.calculate_field_offsets(&alter_info->create_list);/*Auto increment and blob checks.*/int auto_increment = 0;int blob_columns = 0;it.rewind();while ((sql_field = it++)) {if (sql_field->auto_flags & Field::NEXT_NUMBER) auto_increment++;switch (sql_field->sql_type) {case MYSQL_TYPE_GEOMETRY:case MYSQL_TYPE_BLOB:case MYSQL_TYPE_MEDIUM_BLOB:case MYSQL_TYPE_TINY_BLOB:case MYSQL_TYPE_LONG_BLOB:case MYSQL_TYPE_JSON:blob_columns++;break;default:if (sql_field->is_array) blob_columns++;break;}}if (auto_increment > 1) {my_error(ER_WRONG_AUTO_KEY, MYF(0));return true;}if (auto_increment && (file->ha_table_flags() & HA_NO_AUTO_INCREMENT)) {my_error(ER_TABLE_CANT_HANDLE_AUTO_INCREMENT, MYF(0));return true;}if (blob_columns && (file->ha_table_flags() & HA_NO_BLOBS)) {my_error(ER_TABLE_CANT_HANDLE_BLOB, MYF(0));return true;}/*CREATE TABLE[with auto_increment column] SELECT is unsafe as the rowsinserted in the created table depends on the order of the rows fetchedfrom the select tables. This order may differ on master and slave. Wetherefore mark it as unsafe.*/if (select_field_count > 0 && auto_increment)thd->lex->set_stmt_unsafe(LEX::BINLOG_STMT_UNSAFE_CREATE_SELECT_AUTOINC);/*Count keys and key segments.Also mark redundant keys to be ignored.*/uint key_parts;Mem_root_array<bool> redundant_keys(thd->mem_root,alter_info->key_list.size(), false);if (count_keys(alter_info->key_list, key_count, &key_parts, fk_key_count,&redundant_keys, file->ha_table_flags()))return true;if (*key_count > file->max_keys()) {my_error(ER_TOO_MANY_KEYS, MYF(0), file->max_keys());return true;}/*Make KEY objects for the keys in the new table.*/KEY *key_info;(*key_info_buffer) = key_info = (KEY *)sql_calloc(sizeof(KEY) * (*key_count));KEY_PART_INFO *key_part_info =(KEY_PART_INFO *)sql_calloc(sizeof(KEY_PART_INFO) * key_parts);if (!*key_info_buffer || !key_part_info) return true;  // Out of memoryMem_root_array<const KEY *> keys_to_check(thd->mem_root);if (keys_to_check.reserve(*key_count)) return true;  // Out of memoryuint key_number = 0;bool primary_key = false;// First prepare non-foreign keys so that they are ready when// we prepare foreign keys.for (size_t i = 0; i < alter_info->key_list.size(); i++) {if (redundant_keys[i]) continue;  // Skip redundant keysconst Key_spec *key = alter_info->key_list[i];if (key->type == KEYTYPE_PRIMARY) {if (primary_key) {my_error(ER_MULTIPLE_PRI_KEY, MYF(0));return true;}primary_key = true;}if (key->type != KEYTYPE_FOREIGN) {if (prepare_key(thd, error_schema_name, error_table_name, create_info,&alter_info->create_list, key, key_info_buffer, key_info,&key_part_info, keys_to_check, key_number, file,&auto_increment))return true;key_info++;key_number++;}}// If the table is created without PK, we must check if this has// been disabled and return error. Limit the effect of sql_require_primary_key// to only those SEs that can participate in replication.if (!primary_key && !thd->is_dd_system_thread() &&!thd->is_initialize_system_thread() &&(file->ha_table_flags() &(HA_BINLOG_ROW_CAPABLE | HA_BINLOG_STMT_CAPABLE)) != 0 &&thd->variables.sql_require_primary_key) {my_error(ER_TABLE_WITHOUT_PK, MYF(0));return true;}/*At this point all KEY objects are for indexes are fully constructed.So we can check for duplicate indexes for keys for which it was requested.*/const KEY **dup_check_key;for (dup_check_key = keys_to_check.begin();dup_check_key != keys_to_check.end(); dup_check_key++) {if (check_duplicate_key(thd, error_schema_name, error_table_name,*dup_check_key, *key_info_buffer, *key_count,alter_info))return true;}if (!primary_key && check_promoted_index(file, *key_info_buffer, *key_count))return true;/*Any auto increment columns not found during prepare_key?*/if (auto_increment > 0) {my_error(ER_WRONG_AUTO_KEY, MYF(0));return true;}/* Sort keys in optimized order */std::sort(*key_info_buffer, *key_info_buffer + *key_count, sort_keys());/*Normal keys are done, now prepare foreign keys.We do this after sorting normal keys to get predictable behaviorwhen searching for parent keys for self-referencing foreign keys.*/bool se_supports_fks =(create_info->db_type->flags & HTON_SUPPORTS_FOREIGN_KEYS);assert(se_supports_fks || existing_fks_count == 0);(*fk_key_count) += existing_fks_count;FOREIGN_KEY *fk_key_info;(*fk_key_info_buffer) = fk_key_info =(FOREIGN_KEY *)sql_calloc(sizeof(FOREIGN_KEY) * (*fk_key_count));if (!fk_key_info) return true;  // Out of memory// Copy pre-existing foreign keys.if (existing_fks_count > 0)memcpy(*fk_key_info_buffer, existing_fks,existing_fks_count * sizeof(FOREIGN_KEY));uint fk_number = existing_fks_count;fk_key_info += existing_fks_count;/*Check if we are trying to add partitioning to the table with existingforeign keys and table's storage engine doesn't support foreign keysover partitioned tables.*/if (is_partitioned && existing_fks_count > 0 &&(!create_info->db_type->partition_flags ||create_info->db_type->partition_flags() & HA_CANNOT_PARTITION_FK)) {my_error(ER_FOREIGN_KEY_ON_PARTITIONED, MYF(0));return true;}/*Check that definitions of existing foreign keys are not broken by thisALTER TABLE. Update FOREIGN_KEY::unique_constraint_name if necessary.*/for (FOREIGN_KEY *fk = *fk_key_info_buffer;fk < (*fk_key_info_buffer) + existing_fks_count; fk++) {if (prepare_preexisting_foreign_key(thd, create_info, alter_info, error_schema_name, error_table_name,*key_info_buffer, *key_count, existing_fks_table, fk))return true;}// Prepare new foreign keys.for (size_t i = 0; i < alter_info->key_list.size(); i++) {if (redundant_keys[i]) continue;  // Skip redundant keysKey_spec *key = alter_info->key_list[i];if (key->type == KEYTYPE_FOREIGN) {if (prepare_foreign_key(thd, create_info, alter_info, error_schema_name,error_table_name, is_partitioned,*key_info_buffer, *key_count, *fk_key_info_buffer,fk_number, se_supports_fks, find_parent_keys,down_cast<Foreign_key_spec *>(key),&fk_max_generated_name_number, fk_key_info))return true;if (se_supports_fks) {fk_key_info++;fk_number++;}}}/*Check if  STRICT SQL mode is active and server is not started with--explicit-defaults-for-timestamp. Below check was added to prevent implicitdefault 0 value of timestamp. When explicit-defaults-for-timestamp serveroption is removed, whole set of check can be removed.Note that this check must be after KEYs have been created as thiscan cause the NOT_NULL_FLAG to be set.*/if (thd->variables.sql_mode & MODE_NO_ZERO_DATE &&!thd->variables.explicit_defaults_for_timestamp) {it.rewind();while ((sql_field = it++)) {if (!sql_field->constant_default && !sql_field->gcol_info &&is_timestamp_type(sql_field->sql_type) &&(sql_field->flags & NOT_NULL_FLAG) &&!(sql_field->auto_flags & Field::DEFAULT_NOW)) {/*An error should be reported if:- there is no explicit DEFAULT clause (default column value);- this is a TIMESTAMP column;- the column is not NULL;- this is not the DEFAULT CURRENT_TIMESTAMP column.And from checks before while loop,- STRICT SQL mode is active;- server is not started with --explicit-defaults-for-timestampIn other words, an error should be reported if- STRICT SQL mode is active;- the column definition is equivalent to'column_name TIMESTAMP DEFAULT 0'.*/my_error(ER_INVALID_DEFAULT, MYF(0), sql_field->field_name);return true;}}}/* If fixed row records, we need one bit to check for deleted rows */if (!(create_info->table_options & HA_OPTION_PACK_RECORD))create_info->null_bits++;const ulong data_offset = (create_info->null_bits + 7) / 8;size_t reclength = data_offset;it.rewind();while ((sql_field = it++)) {const size_t length = sql_field->pack_length();if (sql_field->offset + data_offset + length > reclength)reclength = sql_field->offset + data_offset + length;}if (reclength > file->max_record_length()) {my_error(ER_TOO_BIG_ROWSIZE, MYF(0),static_cast<long>(file->max_record_length()));return true;}return false;
}

本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若转载,请注明出处:http://www.rhkb.cn/news/15714.html

如若内容造成侵权/违法违规/事实不符,请联系长河编程网进行投诉反馈email:809451989@qq.com,一经查实,立即删除!

相关文章

将DeepSeek接入Excel实现交互式对话

引言 将DeepSeek接入Excel&#xff0c;为你带来前所未有的交互体验&#xff01;“哪里不懂&#xff0c;选中哪里”&#xff0c;然后直接在侧边栏对话框向DeepSeek发问&#xff0c;非常地方便&#xff01; 案例演示 设置接入方式 既可以通过本地部署的DeepSeek接入Excel&#…

在npm上传属于自己的包

最近在整理代码&#xff0c;上传到npm方便使用&#xff0c;所以学习了如何在npm发布一个包&#xff0c;整理写成一篇文章和大家一起交流。 1、注册npm账号 npm | Home 2、确保是登录状态 &#xff08;在包目录下&#xff0c;终端执行 npm login) 按enter键自动打开页面&…

JS宏进阶:XMLHttpRequest对象

一、概述 XMLHttpRequest简称XHR&#xff0c;它是一个可以在JavaScript中使用的对象&#xff0c;用于在后台与服务器交换数据&#xff0c;实现页面的局部更新&#xff0c;而无需重新加载整个页面&#xff0c;也是Ajax&#xff08;Asynchronous JavaScript and XML&#xff09;…

【快应用】多语言适配案例

【关键词】 多语言,$t 【问题背景】 快应用平台的能力会覆盖多个国家地区,平台支持多语言的能力后,可以让一个快应同时支持多个语言版本的切换,开发者无需开发多个不同语言的源码项目,避免给项目维护带来困难。使用系统默认的语言,开发者配置多语言的方式非常简单,只…

PyQt学习记录

0. 安装配置 0.1 安装相关库 首先打开你的PyCharm程序&#xff0c;然后新建一个目录用于学习&#xff0c;其次在terminal中输入 pip install pyqt5如果你不具有科学上网能力&#xff0c;请改为国内源 pip install pyqt5 -i https://pypi.douban.com/simple然后安装pyqt相关…

【多模态大模型】系列3:语义分割(LSeg、GroupViT)

目录 1 LSeg2 Group ViT 1 LSeg LANGUAGE-DRIVEN SEMANTIC SEGMENTATION LSeg是第一篇将CLIP应用于语义分割的论文。它的分割的效果拔群&#xff0c;容错能力也很高&#xff1a; 模型总览图跟CLIP很像&#xff1a; 对于图像链路&#xff1a;输入一张图片&#xff0c;进入分割…

【深度学习】多目标融合算法(四):多门混合专家网络MMOE(Multi-gate Mixture-of-Experts)

目录 一、引言 二、MMoE&#xff08;Multi-gate Mixture-of-Experts&#xff0c;多门混合专家网络&#xff09; 2.1 技术原理 2.2 技术优缺点 2.3 业务代码实践 2.3.1 业务场景与建模 2.3.2 模型代码实现 2.3.3 模型训练与推理测试 2.3.4 打印模型结构 三、总结 一、…

自动驾驶数据集三剑客:nuScenes、nuImages 与 nuPlan 的技术矩阵与生态协同

目录 1、引言 2、主要内容 2.1、定位对比&#xff1a;感知与规划的全维覆盖 2.2、数据与技术特性对比 2.3、技术协同&#xff1a;构建全栈研发生态 2.4、应用场景与评估体系 2.5、总结与展望 3、参考文献 1、引言 随着自动驾驶技术向全栈化迈进&#xff0c;Motional 团…

使用 AlexNet 实现图片分类 | PyTorch 深度学习实战

前一篇文章&#xff0c;CNN 卷积神经网络处理图片任务 | PyTorch 深度学习实战 本系列文章 GitHub Repo: https://github.com/hailiang-wang/pytorch-get-started 本篇文章内容来自于 强化学习必修课&#xff1a;引领人工智能新时代【梗直哥瞿炜】 使用 AlexNet 实现图片分类…

C# Winform 使用委托实现C++中回调函数的功能

C# Winform 使用委托实现C中回调函数的功能 在项目中遇到了使用C#调用C封装的接口&#xff0c;其中C接口有一个回调函数的参数。参考对比后&#xff0c;在C#中是使用委托(delegate)来实现类似的功能。 下面使用一个示例来介绍具体的使用方式&#xff1a; 第一步&#xff1a;…

攻防世界33 catcat-new【文件包含/flask_session伪造】

题目&#xff1a; 点击一只猫猫&#xff1a; 看这个url像是文件包含漏洞&#xff0c;试试 dirsearch扫出来/admin&#xff0c;访问也没成功&#xff08;--delay 0.1 -t 5&#xff09; 会的那几招全用不了了哈哈&#xff0c;那就继续看答案 先总结几个知识点 1./etc/passwd&am…

ArgoCD实战指南:GitOps驱动下的Kubernetes自动化部署与Helm/Kustomize集成

摘要 ArgoCD 是一种 GitOps 持续交付工具,专为 Kubernetes 设计。它能够自动同步 Git 仓库中的声明性配置,并将其应用到 Kubernetes 集群中。本文将介绍 ArgoCD 的架构、安装步骤,以及如何结合 Helm 和 Kustomize 进行 Kubernetes 自动化部署。 引言 为什么选择 ArgoCD?…

尝试一下,交互式的三维计算python库,py3d

py3d是一个我开发的三维计算python库&#xff0c;目前不定期在PYPI上发版&#xff0c;可以通过pip直接安装 pip install py3d 开发这个库主要可视化是想把自己在工作中常用的三维方法汇总积累下来&#xff0c;不必每次重新造轮子。其实现成的python库也有很多&#xff0c;例如…

【愚公系列】《循序渐进Vue.js 3.x前端开发实践》070-商业项目:电商后台管理系统实战(商品管理模块的开发)

标题详情作者简介愚公搬代码头衔华为云特约编辑&#xff0c;华为云云享专家&#xff0c;华为开发者专家&#xff0c;华为产品云测专家&#xff0c;CSDN博客专家&#xff0c;CSDN商业化专家&#xff0c;阿里云专家博主&#xff0c;阿里云签约作者&#xff0c;腾讯云优秀博主&…

5 个释放 安卓潜力的 Shizuku 应用

Shizuku 软件推荐&#xff1a;释放安卓潜力的五款应用 Shizuku (日语&#xff1a;雫&#xff0c;意为“水滴”) 正如其名&#xff0c;是一款轻巧但功能强大的安卓工具。它无需 Root 权限&#xff0c;通过 ADB (Android Debug Bridge) 授权&#xff0c;即可让应用调用系统 API&…

前端权限控制和管理

前端权限控制和管理 1.前言2.权限相关概念2.1权限的分类(1)后端权限(2)前端权限 2.2前端权限的意义 3.前端权限控制思路3.1菜单的权限控制3.2界面的权限控制3.3按钮的权限控制3.4接口的权限控制 4.实现步骤4.1菜单栏控制4.2界面的控制(1)路由导航守卫(2)动态路由 4.3按钮的控制…

分布式kettle调度平台- web版转换,作业编排新功能介绍

介绍 Kettle&#xff08;也称为Pentaho Data Integration&#xff09;是一款开源的ETL&#xff08;Extract, Transform, Load&#xff09;工具&#xff0c;由Pentaho&#xff08;现为Hitachi Vantara&#xff09;开发和维护。它提供了一套强大的数据集成和转换功能&#xff0c…

Docker容器访问外网:启动时的网络参数配置指南

在启动Docker镜像时,可以通过设置网络参数来确保容器能够访问外网。以下是几种常见的方法: 1. 使用默认的bridge网络 Docker的默认网络模式是bridge,它会创建一个虚拟网桥,将容器连接到宿主机的网络上。在大多数情况下,使用默认的bridge网络配置即可使容器访问外网。 启动…

大语言模型RAG,transformer

1、RAG技术流总结 第一张图是比较经典的RAG知识图谱&#xff0c;第二张图是更加详细扎实的介绍图。 1.1 索引 坦白来说这部分的技术并不是大模型领域的&#xff0c;更像是之前技术在大模型领域的应用&#xff1b;早在2019年我就做过faiss部分的尝试&#xff0c;彼时索引技术已…

数据结构与算法(test3)

七、查找 1. 看图填空 查找表是由同一类型的数据元素&#xff08;或记录&#xff09;构成的集合。例如上图就是一个查找表。 期中&#xff08;1&#xff09;是______________. (2)是______________(3)是_____关键字_______。 2. 查找(Searching) 就是根据给定的某个值, 在查…