Merge remote-tracking branch 'origin/master' into codex/ddd
# Conflicts: # docs/technical/README.md # docs/technical/RUST_API_SERVER_ROUTE_INDEX_2026-04-22.md # docs/technical/SPACETIMEDB_TABLE_CATALOG.md # scripts/generate-spacetime-bindings.mjs # server-rs/crates/api-server/src/app.rs # server-rs/crates/api-server/src/assets.rs # server-rs/crates/api-server/src/big_fish.rs # server-rs/crates/api-server/src/custom_world_ai.rs # server-rs/crates/api-server/src/llm.rs # server-rs/crates/api-server/src/main.rs # server-rs/crates/api-server/src/puzzle.rs # server-rs/crates/api-server/src/runtime_profile.rs # server-rs/crates/api-server/src/runtime_story/compat/ai.rs # server-rs/crates/api-server/src/runtime_story/compat/npc_actions.rs # server-rs/crates/api-server/src/runtime_story/compat/presentation.rs # server-rs/crates/api-server/src/runtime_story/compat/tests.rs # server-rs/crates/api-server/src/state.rs # server-rs/crates/module-auth/src/lib.rs # server-rs/crates/module-big-fish/src/lib.rs # server-rs/crates/module-custom-world/src/lib.rs # server-rs/crates/module-puzzle/src/lib.rs # server-rs/crates/module-runtime/src/lib.rs # server-rs/crates/spacetime-client/src/big_fish.rs # server-rs/crates/spacetime-client/src/lib.rs # server-rs/crates/spacetime-client/src/mapper.rs # server-rs/crates/spacetime-client/src/module_bindings/admin_disable_profile_redeem_code_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/admin_upsert_profile_redeem_code_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/advance_puzzle_next_level_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/append_ai_text_chunk_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/apply_chapter_progression_ledger_entry_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/attach_ai_result_reference_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/authorize_database_migration_operator_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/begin_story_session_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/big_fish_runtime_run_type.rs # server-rs/crates/spacetime-client/src/module_bindings/bind_asset_object_to_entity_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/cancel_ai_task_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/clear_platform_browse_history_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/compile_big_fish_draft_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/compile_custom_world_published_profile_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/compile_puzzle_agent_draft_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/complete_ai_stage_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/complete_ai_task_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/confirm_asset_object_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/consume_profile_wallet_points_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/continue_story_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/create_ai_task_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/create_battle_state_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/create_big_fish_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/create_custom_world_agent_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/create_profile_recharge_order_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/create_puzzle_agent_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/delete_big_fish_work_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/delete_custom_world_agent_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/delete_custom_world_profile_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/delete_puzzle_work_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/delete_runtime_snapshot_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/drag_puzzle_piece_or_group_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/execute_custom_world_agent_action_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/export_auth_store_snapshot_from_tables_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/export_database_migration_to_file_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/fail_ai_task_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/finalize_big_fish_agent_message_turn_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/finalize_custom_world_agent_message_turn_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/finalize_puzzle_agent_message_turn_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/generate_big_fish_asset_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_auth_store_snapshot_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_battle_state_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_big_fish_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_chapter_progression_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_custom_world_agent_card_detail_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_custom_world_agent_operation_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_custom_world_agent_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_custom_world_gallery_detail_by_code_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_custom_world_gallery_detail_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_custom_world_library_detail_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_player_progression_or_default_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_profile_dashboard_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_profile_play_stats_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_profile_recharge_center_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_profile_referral_invite_center_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_puzzle_agent_session_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_puzzle_gallery_detail_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_puzzle_run_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_puzzle_work_detail_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_runtime_inventory_state_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_runtime_setting_or_default_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_runtime_snapshot_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/get_story_session_state_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/grant_player_progression_experience_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/import_auth_store_snapshot_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/import_database_migration_from_file_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/import_database_migration_incremental_from_file_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_asset_history_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_big_fish_works_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_custom_world_gallery_entries_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_custom_world_profiles_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_custom_world_works_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_platform_browse_history_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_profile_save_archives_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_profile_wallet_ledger_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_puzzle_gallery_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/list_puzzle_works_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/mod.rs # server-rs/crates/spacetime-client/src/module_bindings/publish_big_fish_game_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/publish_custom_world_profile_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/publish_custom_world_world_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/publish_puzzle_work_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/record_big_fish_play_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/redeem_profile_referral_invite_code_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/redeem_profile_reward_code_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/refund_profile_wallet_points_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/resolve_combat_action_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/resolve_npc_battle_interaction_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/resolve_npc_interaction_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/resolve_npc_social_action_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/resolve_treasure_interaction_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/resume_profile_save_archive_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/revoke_database_migration_operator_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/save_puzzle_generated_images_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/select_puzzle_cover_image_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/start_puzzle_run_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/submit_big_fish_message_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/submit_custom_world_agent_message_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/submit_puzzle_agent_message_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/submit_puzzle_leaderboard_entry_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/swap_puzzle_pieces_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/unpublish_custom_world_profile_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/update_puzzle_work_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_auth_store_snapshot_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_chapter_progression_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_custom_world_agent_operation_progress_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_custom_world_profile_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_npc_state_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_platform_browse_history_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_runtime_setting_and_return_procedure.rs # server-rs/crates/spacetime-client/src/module_bindings/upsert_runtime_snapshot_and_return_procedure.rs # server-rs/crates/spacetime-module/src/auth/procedures.rs # server-rs/crates/spacetime-module/src/custom_world/mod.rs # server-rs/crates/spacetime-module/src/lib.rs # server-rs/crates/spacetime-module/src/migration.rs # server-rs/crates/spacetime-module/src/puzzle.rs # server-rs/crates/spacetime-module/src/runtime/profile.rs # src/components/platform-entry/PlatformEntryFlowShellImpl.tsx # src/components/rpg-entry/RpgEntryFlowShell.agent.interaction.test.tsx # src/services/aiService.ts # src/services/puzzle-runtime/puzzleRuntimeClient.ts
This commit is contained in:
@@ -5,6 +5,9 @@ use spacetimedb_lib::sats::ser::serde::SerializeWrapper;
|
||||
use std::collections::HashSet;
|
||||
|
||||
use crate::big_fish::big_fish_runtime_run;
|
||||
use crate::match3d::tables::{
|
||||
match3d_agent_message, match3d_agent_session, match3d_runtime_run, match3d_work_profile,
|
||||
};
|
||||
use crate::puzzle::{
|
||||
puzzle_agent_message, puzzle_agent_session, puzzle_event, puzzle_leaderboard_entry,
|
||||
puzzle_runtime_run, puzzle_work_profile,
|
||||
@@ -12,6 +15,8 @@ use crate::puzzle::{
|
||||
|
||||
const MIGRATION_SCHEMA_VERSION: u32 = 1;
|
||||
const MIGRATION_MAX_TABLE_NAME_LEN: usize = 96;
|
||||
const MIGRATION_MAX_IMPORT_UPLOAD_ID_LEN: usize = 128;
|
||||
const MIGRATION_MAX_IMPORT_CHUNK_BYTES: usize = 1024 * 1024;
|
||||
const MIGRATION_MAX_OPERATOR_NOTE_CHARS: usize = 160;
|
||||
const MIGRATION_MIN_BOOTSTRAP_SECRET_LEN: usize = 16;
|
||||
const MIGRATION_BOOTSTRAP_SECRET: Option<&str> =
|
||||
@@ -26,6 +31,21 @@ pub struct DatabaseMigrationOperator {
|
||||
pub note: String,
|
||||
}
|
||||
|
||||
#[spacetimedb::table(
|
||||
accessor = database_migration_import_chunk,
|
||||
index(accessor = by_database_migration_import_upload, btree(columns = [upload_id]))
|
||||
)]
|
||||
pub struct DatabaseMigrationImportChunk {
|
||||
#[primary_key]
|
||||
pub chunk_key: String,
|
||||
pub upload_id: String,
|
||||
pub chunk_index: u32,
|
||||
pub chunk_count: u32,
|
||||
pub operator_identity: Identity,
|
||||
pub created_at: Timestamp,
|
||||
pub chunk: String,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, Eq, SpacetimeType)]
|
||||
pub struct DatabaseMigrationExportInput {
|
||||
pub include_tables: Vec<String>,
|
||||
@@ -39,6 +59,27 @@ pub struct DatabaseMigrationImportInput {
|
||||
pub dry_run: bool,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, Eq, SpacetimeType)]
|
||||
pub struct DatabaseMigrationImportChunkInput {
|
||||
pub upload_id: String,
|
||||
pub chunk_index: u32,
|
||||
pub chunk_count: u32,
|
||||
pub chunk: String,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, Eq, SpacetimeType)]
|
||||
pub struct DatabaseMigrationImportChunksInput {
|
||||
pub upload_id: String,
|
||||
pub include_tables: Vec<String>,
|
||||
pub replace_existing: bool,
|
||||
pub dry_run: bool,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, Eq, SpacetimeType)]
|
||||
pub struct DatabaseMigrationImportChunksClearInput {
|
||||
pub upload_id: String,
|
||||
}
|
||||
|
||||
#[derive(Clone, Copy, Debug, PartialEq, Eq)]
|
||||
enum DatabaseMigrationImportMode {
|
||||
Strict,
|
||||
@@ -65,12 +106,20 @@ pub struct DatabaseMigrationTableStat {
|
||||
pub skipped_row_count: u64,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, Eq, SpacetimeType)]
|
||||
pub struct DatabaseMigrationWarning {
|
||||
pub table_name: String,
|
||||
pub warning_kind: String,
|
||||
pub message: String,
|
||||
}
|
||||
|
||||
#[derive(Clone, Debug, PartialEq, Eq, SpacetimeType)]
|
||||
pub struct DatabaseMigrationProcedureResult {
|
||||
pub ok: bool,
|
||||
pub schema_version: u32,
|
||||
pub migration_json: Option<String>,
|
||||
pub table_stats: Vec<DatabaseMigrationTableStat>,
|
||||
pub warnings: Vec<DatabaseMigrationWarning>,
|
||||
pub error_message: Option<String>,
|
||||
}
|
||||
|
||||
@@ -117,6 +166,8 @@ macro_rules! migration_tables {
|
||||
profile_invite_code,
|
||||
profile_referral_relation,
|
||||
profile_played_world,
|
||||
public_work_play_daily_stat,
|
||||
public_work_like,
|
||||
profile_membership,
|
||||
profile_recharge_order,
|
||||
profile_save_archive,
|
||||
@@ -146,6 +197,10 @@ macro_rules! migration_tables {
|
||||
puzzle_event,
|
||||
puzzle_runtime_run,
|
||||
puzzle_leaderboard_entry,
|
||||
match3d_agent_session,
|
||||
match3d_agent_message,
|
||||
match3d_work_profile,
|
||||
match3d_runtime_run,
|
||||
big_fish_creation_session,
|
||||
big_fish_agent_message,
|
||||
big_fish_asset_slot,
|
||||
@@ -249,6 +304,7 @@ pub fn export_database_migration_to_file(
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: Some(migration_json),
|
||||
table_stats: stats,
|
||||
warnings: Vec::new(),
|
||||
error_message: None,
|
||||
},
|
||||
Err(error) => DatabaseMigrationProcedureResult {
|
||||
@@ -256,6 +312,7 @@ pub fn export_database_migration_to_file(
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: Vec::new(),
|
||||
warnings: Vec::new(),
|
||||
error_message: Some(error),
|
||||
},
|
||||
}
|
||||
@@ -269,11 +326,12 @@ pub fn import_database_migration_from_file(
|
||||
) -> DatabaseMigrationProcedureResult {
|
||||
match import_database_migration_from_file_inner(ctx, input, DatabaseMigrationImportMode::Strict)
|
||||
{
|
||||
Ok(stats) => DatabaseMigrationProcedureResult {
|
||||
Ok((stats, warnings)) => DatabaseMigrationProcedureResult {
|
||||
ok: true,
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: stats,
|
||||
warnings,
|
||||
error_message: None,
|
||||
},
|
||||
Err(error) => DatabaseMigrationProcedureResult {
|
||||
@@ -281,6 +339,7 @@ pub fn import_database_migration_from_file(
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: Vec::new(),
|
||||
warnings: Vec::new(),
|
||||
error_message: Some(error),
|
||||
},
|
||||
}
|
||||
@@ -297,11 +356,12 @@ pub fn import_database_migration_incremental_from_file(
|
||||
input,
|
||||
DatabaseMigrationImportMode::Incremental,
|
||||
) {
|
||||
Ok(stats) => DatabaseMigrationProcedureResult {
|
||||
Ok((stats, warnings)) => DatabaseMigrationProcedureResult {
|
||||
ok: true,
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: stats,
|
||||
warnings,
|
||||
error_message: None,
|
||||
},
|
||||
Err(error) => DatabaseMigrationProcedureResult {
|
||||
@@ -309,11 +369,82 @@ pub fn import_database_migration_incremental_from_file(
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: Vec::new(),
|
||||
warnings: Vec::new(),
|
||||
error_message: Some(error),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// 大迁移 JSON 先按分片写入私有临时表,避免单次 HTTP request body 触发 SpacetimeDB 413。
|
||||
#[spacetimedb::procedure]
|
||||
pub fn put_database_migration_import_chunk(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunkInput,
|
||||
) -> DatabaseMigrationProcedureResult {
|
||||
match put_database_migration_import_chunk_inner(ctx, input) {
|
||||
Ok(()) => empty_database_migration_result(true, None),
|
||||
Err(error) => empty_database_migration_result(false, Some(error)),
|
||||
}
|
||||
}
|
||||
|
||||
// 分片提交保持与直接导入相同的严格追加语义;提交成功后清理临时分片。
|
||||
#[spacetimedb::procedure]
|
||||
pub fn import_database_migration_from_chunks(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunksInput,
|
||||
) -> DatabaseMigrationProcedureResult {
|
||||
match import_database_migration_from_chunks_inner(
|
||||
ctx,
|
||||
input,
|
||||
DatabaseMigrationImportMode::Strict,
|
||||
) {
|
||||
Ok((stats, warnings)) => DatabaseMigrationProcedureResult {
|
||||
ok: true,
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: stats,
|
||||
warnings,
|
||||
error_message: None,
|
||||
},
|
||||
Err(error) => empty_database_migration_result(false, Some(error)),
|
||||
}
|
||||
}
|
||||
|
||||
// 分片增量提交只插入目标库缺失的行;主键或唯一约束冲突的行会跳过。
|
||||
#[spacetimedb::procedure]
|
||||
pub fn import_database_migration_incremental_from_chunks(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunksInput,
|
||||
) -> DatabaseMigrationProcedureResult {
|
||||
match import_database_migration_from_chunks_inner(
|
||||
ctx,
|
||||
input,
|
||||
DatabaseMigrationImportMode::Incremental,
|
||||
) {
|
||||
Ok((stats, warnings)) => DatabaseMigrationProcedureResult {
|
||||
ok: true,
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: stats,
|
||||
warnings,
|
||||
error_message: None,
|
||||
},
|
||||
Err(error) => empty_database_migration_result(false, Some(error)),
|
||||
}
|
||||
}
|
||||
|
||||
// 调用方上传失败或提交失败时可显式清理同一 upload_id 的临时分片。
|
||||
#[spacetimedb::procedure]
|
||||
pub fn clear_database_migration_import_chunks(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunksClearInput,
|
||||
) -> DatabaseMigrationProcedureResult {
|
||||
match clear_database_migration_import_chunks_inner(ctx, input) {
|
||||
Ok(()) => empty_database_migration_result(true, None),
|
||||
Err(error) => empty_database_migration_result(false, Some(error)),
|
||||
}
|
||||
}
|
||||
|
||||
fn export_database_migration_to_file_inner(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationExportInput,
|
||||
@@ -337,7 +468,13 @@ fn import_database_migration_from_file_inner(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportInput,
|
||||
import_mode: DatabaseMigrationImportMode,
|
||||
) -> Result<Vec<DatabaseMigrationTableStat>, String> {
|
||||
) -> Result<
|
||||
(
|
||||
Vec<DatabaseMigrationTableStat>,
|
||||
Vec<DatabaseMigrationWarning>,
|
||||
),
|
||||
String,
|
||||
> {
|
||||
let caller = ctx.sender();
|
||||
let included_tables = normalize_include_tables(&input.include_tables)?;
|
||||
if import_mode == DatabaseMigrationImportMode::Incremental && input.replace_existing {
|
||||
@@ -348,16 +485,9 @@ fn import_database_migration_from_file_inner(
|
||||
}
|
||||
ctx.try_with_tx(|tx| require_migration_operator(tx, caller))?;
|
||||
|
||||
let migration_file = serde_json::from_str::<MigrationFile>(&input.migration_json)
|
||||
.map_err(|error| format!("迁移文件 JSON 解析失败: {error}"))?;
|
||||
if migration_file.schema_version != MIGRATION_SCHEMA_VERSION {
|
||||
return Err(format!(
|
||||
"迁移文件 schema_version 不匹配,期望 {},实际 {}",
|
||||
MIGRATION_SCHEMA_VERSION, migration_file.schema_version
|
||||
));
|
||||
}
|
||||
let migration_file = parse_migration_file(&input.migration_json)?;
|
||||
|
||||
let stats = if input.dry_run {
|
||||
let (stats, warnings) = if input.dry_run {
|
||||
build_import_dry_run_stats(&migration_file.tables, included_tables.as_ref())?
|
||||
} else {
|
||||
ctx.try_with_tx(|tx| {
|
||||
@@ -372,7 +502,159 @@ fn import_database_migration_from_file_inner(
|
||||
})?
|
||||
};
|
||||
|
||||
Ok(stats)
|
||||
Ok((stats, warnings))
|
||||
}
|
||||
|
||||
fn put_database_migration_import_chunk_inner(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunkInput,
|
||||
) -> Result<(), String> {
|
||||
let caller = ctx.sender();
|
||||
let upload_id = normalize_import_upload_id(&input.upload_id)?;
|
||||
if input.chunk_count == 0 {
|
||||
return Err("分片总数必须大于 0".to_string());
|
||||
}
|
||||
if input.chunk_index >= input.chunk_count {
|
||||
return Err(format!(
|
||||
"分片序号越界: {} / {}",
|
||||
input.chunk_index, input.chunk_count
|
||||
));
|
||||
}
|
||||
if input.chunk.is_empty() {
|
||||
return Err("迁移 JSON 分片不能为空".to_string());
|
||||
}
|
||||
if input.chunk.len() > MIGRATION_MAX_IMPORT_CHUNK_BYTES {
|
||||
return Err(format!(
|
||||
"迁移 JSON 分片过大,单片最多 {} bytes",
|
||||
MIGRATION_MAX_IMPORT_CHUNK_BYTES
|
||||
));
|
||||
}
|
||||
|
||||
let chunk_key = build_import_chunk_key(&upload_id, input.chunk_index);
|
||||
ctx.try_with_tx(|tx| {
|
||||
require_migration_operator(tx, caller)?;
|
||||
if let Some(existing) = tx
|
||||
.db
|
||||
.database_migration_import_chunk()
|
||||
.chunk_key()
|
||||
.find(&chunk_key)
|
||||
{
|
||||
if existing.operator_identity != caller {
|
||||
return Err("同名迁移分片已由其他 identity 上传,已拒绝覆盖".to_string());
|
||||
}
|
||||
tx.db
|
||||
.database_migration_import_chunk()
|
||||
.chunk_key()
|
||||
.delete(&chunk_key);
|
||||
}
|
||||
tx.db
|
||||
.database_migration_import_chunk()
|
||||
.insert(DatabaseMigrationImportChunk {
|
||||
chunk_key: chunk_key.clone(),
|
||||
upload_id: upload_id.clone(),
|
||||
chunk_index: input.chunk_index,
|
||||
chunk_count: input.chunk_count,
|
||||
operator_identity: caller,
|
||||
created_at: tx.timestamp,
|
||||
chunk: input.chunk.clone(),
|
||||
});
|
||||
Ok(())
|
||||
})?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn import_database_migration_from_chunks_inner(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunksInput,
|
||||
import_mode: DatabaseMigrationImportMode,
|
||||
) -> Result<
|
||||
(
|
||||
Vec<DatabaseMigrationTableStat>,
|
||||
Vec<DatabaseMigrationWarning>,
|
||||
),
|
||||
String,
|
||||
> {
|
||||
let caller = ctx.sender();
|
||||
let upload_id = normalize_import_upload_id(&input.upload_id)?;
|
||||
let included_tables = normalize_include_tables(&input.include_tables)?;
|
||||
if import_mode == DatabaseMigrationImportMode::Incremental && input.replace_existing {
|
||||
return Err("增量导入不能同时启用 replace_existing".to_string());
|
||||
}
|
||||
|
||||
let migration_json = ctx.try_with_tx(|tx| {
|
||||
require_migration_operator(tx, caller)?;
|
||||
read_database_migration_import_chunks(tx, &upload_id, caller)
|
||||
})?;
|
||||
let migration_file = parse_migration_file(&migration_json)?;
|
||||
|
||||
let (stats, warnings) = if input.dry_run {
|
||||
build_import_dry_run_stats(&migration_file.tables, included_tables.as_ref())?
|
||||
} else {
|
||||
ctx.try_with_tx(|tx| {
|
||||
require_migration_operator(tx, caller)?;
|
||||
apply_migration_file(
|
||||
tx,
|
||||
&migration_file,
|
||||
included_tables.as_ref(),
|
||||
input.replace_existing,
|
||||
import_mode,
|
||||
)
|
||||
})?
|
||||
};
|
||||
|
||||
ctx.try_with_tx(|tx| {
|
||||
require_migration_operator(tx, caller)?;
|
||||
clear_database_migration_import_chunks_tx(tx, &upload_id);
|
||||
Ok::<(), String>(())
|
||||
})?;
|
||||
|
||||
Ok((stats, warnings))
|
||||
}
|
||||
|
||||
fn clear_database_migration_import_chunks_inner(
|
||||
ctx: &mut ProcedureContext,
|
||||
input: DatabaseMigrationImportChunksClearInput,
|
||||
) -> Result<(), String> {
|
||||
let caller = ctx.sender();
|
||||
let upload_id = normalize_import_upload_id(&input.upload_id)?;
|
||||
ctx.try_with_tx(|tx| {
|
||||
require_migration_operator(tx, caller)?;
|
||||
clear_database_migration_import_chunks_tx(tx, &upload_id);
|
||||
Ok::<(), String>(())
|
||||
})?;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn empty_database_migration_result(
|
||||
ok: bool,
|
||||
error_message: Option<String>,
|
||||
) -> DatabaseMigrationProcedureResult {
|
||||
DatabaseMigrationProcedureResult {
|
||||
ok,
|
||||
schema_version: MIGRATION_SCHEMA_VERSION,
|
||||
migration_json: None,
|
||||
table_stats: Vec::new(),
|
||||
warnings: Vec::new(),
|
||||
error_message,
|
||||
}
|
||||
}
|
||||
|
||||
fn parse_migration_file(migration_json: &str) -> Result<MigrationFile, String> {
|
||||
if migration_json.trim().is_empty() {
|
||||
return Err("migration_json 不能为空".to_string());
|
||||
}
|
||||
|
||||
let migration_file = serde_json::from_str::<MigrationFile>(migration_json)
|
||||
.map_err(|error| format!("迁移文件 JSON 解析失败: {error}"))?;
|
||||
if migration_file.schema_version != MIGRATION_SCHEMA_VERSION {
|
||||
return Err(format!(
|
||||
"迁移文件 schema_version 不匹配,期望 {},实际 {}",
|
||||
MIGRATION_SCHEMA_VERSION, migration_file.schema_version
|
||||
));
|
||||
}
|
||||
|
||||
Ok(migration_file)
|
||||
}
|
||||
|
||||
fn authorize_database_migration_operator_inner(
|
||||
@@ -516,6 +798,96 @@ fn normalize_migration_operator_note(input: &str) -> Result<String, String> {
|
||||
Ok(note.to_string())
|
||||
}
|
||||
|
||||
fn normalize_import_upload_id(input: &str) -> Result<String, String> {
|
||||
let upload_id = input.trim();
|
||||
if upload_id.is_empty() {
|
||||
return Err("upload_id 不能为空".to_string());
|
||||
}
|
||||
if upload_id.len() > MIGRATION_MAX_IMPORT_UPLOAD_ID_LEN {
|
||||
return Err(format!(
|
||||
"upload_id 过长,最多 {} bytes",
|
||||
MIGRATION_MAX_IMPORT_UPLOAD_ID_LEN
|
||||
));
|
||||
}
|
||||
if !upload_id
|
||||
.chars()
|
||||
.all(|character| character.is_ascii_alphanumeric() || matches!(character, '-' | '_'))
|
||||
{
|
||||
return Err("upload_id 只能使用 ASCII 字母、数字、短横线或下划线".to_string());
|
||||
}
|
||||
Ok(upload_id.to_string())
|
||||
}
|
||||
|
||||
fn build_import_chunk_key(upload_id: &str, chunk_index: u32) -> String {
|
||||
format!("{upload_id}:{chunk_index:010}")
|
||||
}
|
||||
|
||||
fn read_database_migration_import_chunks(
|
||||
ctx: &ReducerContext,
|
||||
upload_id: &str,
|
||||
caller: Identity,
|
||||
) -> Result<String, String> {
|
||||
let mut chunks = ctx
|
||||
.db
|
||||
.database_migration_import_chunk()
|
||||
.by_database_migration_import_upload()
|
||||
.filter(upload_id)
|
||||
.collect::<Vec<_>>();
|
||||
if chunks.is_empty() {
|
||||
return Err(format!("未找到迁移 JSON 分片: {upload_id}"));
|
||||
}
|
||||
if chunks.iter().any(|chunk| chunk.operator_identity != caller) {
|
||||
return Err("迁移 JSON 分片包含其他 identity 上传的片段,已拒绝提交".to_string());
|
||||
}
|
||||
|
||||
let chunk_count = chunks[0].chunk_count;
|
||||
if chunk_count == 0 {
|
||||
return Err("迁移 JSON 分片总数不合法".to_string());
|
||||
}
|
||||
if chunks
|
||||
.iter()
|
||||
.any(|chunk| chunk.chunk_count != chunk_count || chunk.upload_id != upload_id)
|
||||
{
|
||||
return Err("迁移 JSON 分片总数不一致".to_string());
|
||||
}
|
||||
if chunks.len() != chunk_count as usize {
|
||||
return Err(format!(
|
||||
"迁移 JSON 分片未上传完整,已收到 {} / {}",
|
||||
chunks.len(),
|
||||
chunk_count
|
||||
));
|
||||
}
|
||||
|
||||
chunks.sort_by_key(|chunk| chunk.chunk_index);
|
||||
let mut expected_index = 0u32;
|
||||
let mut migration_json = String::new();
|
||||
for chunk in chunks {
|
||||
if chunk.chunk_index != expected_index {
|
||||
return Err(format!("迁移 JSON 分片缺失序号: {expected_index}"));
|
||||
}
|
||||
migration_json.push_str(&chunk.chunk);
|
||||
expected_index = expected_index.saturating_add(1);
|
||||
}
|
||||
|
||||
Ok(migration_json)
|
||||
}
|
||||
|
||||
fn clear_database_migration_import_chunks_tx(ctx: &ReducerContext, upload_id: &str) {
|
||||
let chunk_keys = ctx
|
||||
.db
|
||||
.database_migration_import_chunk()
|
||||
.by_database_migration_import_upload()
|
||||
.filter(upload_id)
|
||||
.map(|chunk| chunk.chunk_key)
|
||||
.collect::<Vec<_>>();
|
||||
for chunk_key in chunk_keys {
|
||||
ctx.db
|
||||
.database_migration_import_chunk()
|
||||
.chunk_key()
|
||||
.delete(&chunk_key);
|
||||
}
|
||||
}
|
||||
|
||||
fn normalize_include_tables(input: &[String]) -> Result<Option<HashSet<String>>, String> {
|
||||
if input.is_empty() {
|
||||
return Ok(None);
|
||||
@@ -574,11 +946,25 @@ fn build_export_stats(tables: &[MigrationTable]) -> Vec<DatabaseMigrationTableSt
|
||||
fn build_import_dry_run_stats(
|
||||
tables: &[MigrationTable],
|
||||
include_tables: Option<&HashSet<String>>,
|
||||
) -> Result<Vec<DatabaseMigrationTableStat>, String> {
|
||||
) -> Result<
|
||||
(
|
||||
Vec<DatabaseMigrationTableStat>,
|
||||
Vec<DatabaseMigrationWarning>,
|
||||
),
|
||||
String,
|
||||
> {
|
||||
let mut stats = Vec::new();
|
||||
let mut warnings = Vec::new();
|
||||
for table in tables {
|
||||
if !is_supported_migration_table(&table.name) {
|
||||
return Err(format!("迁移文件包含不支持的表: {}", table.name));
|
||||
warnings.push(build_dropped_table_warning(table));
|
||||
stats.push(DatabaseMigrationTableStat {
|
||||
table_name: table.name.clone(),
|
||||
exported_row_count: 0,
|
||||
imported_row_count: 0,
|
||||
skipped_row_count: table.rows.len() as u64,
|
||||
});
|
||||
continue;
|
||||
}
|
||||
if should_include_table(include_tables, &table.name) {
|
||||
stats.push(DatabaseMigrationTableStat {
|
||||
@@ -596,7 +982,7 @@ fn build_import_dry_run_stats(
|
||||
});
|
||||
}
|
||||
}
|
||||
Ok(stats)
|
||||
Ok((stats, warnings))
|
||||
}
|
||||
|
||||
fn apply_migration_file(
|
||||
@@ -605,13 +991,15 @@ fn apply_migration_file(
|
||||
include_tables: Option<&HashSet<String>>,
|
||||
replace_existing: bool,
|
||||
import_mode: DatabaseMigrationImportMode,
|
||||
) -> Result<Vec<DatabaseMigrationTableStat>, String> {
|
||||
) -> Result<
|
||||
(
|
||||
Vec<DatabaseMigrationTableStat>,
|
||||
Vec<DatabaseMigrationWarning>,
|
||||
),
|
||||
String,
|
||||
> {
|
||||
let mut stats = Vec::new();
|
||||
for table in &migration_file.tables {
|
||||
if !is_supported_migration_table(&table.name) {
|
||||
return Err(format!("迁移文件包含不支持的表: {}", table.name));
|
||||
}
|
||||
}
|
||||
let mut warnings = Vec::new();
|
||||
|
||||
let import_table_names = build_import_table_name_set(migration_file, include_tables);
|
||||
if replace_existing {
|
||||
@@ -620,6 +1008,17 @@ fn apply_migration_file(
|
||||
}
|
||||
|
||||
for table in &migration_file.tables {
|
||||
if !is_supported_migration_table(&table.name) {
|
||||
warnings.push(build_dropped_table_warning(table));
|
||||
stats.push(DatabaseMigrationTableStat {
|
||||
table_name: table.name.clone(),
|
||||
exported_row_count: 0,
|
||||
imported_row_count: 0,
|
||||
skipped_row_count: table.rows.len() as u64,
|
||||
});
|
||||
continue;
|
||||
}
|
||||
|
||||
if !should_include_table(include_tables, &table.name) {
|
||||
stats.push(DatabaseMigrationTableStat {
|
||||
table_name: table.name.clone(),
|
||||
@@ -631,7 +1030,7 @@ fn apply_migration_file(
|
||||
}
|
||||
|
||||
let (imported_row_count, skipped_row_count) =
|
||||
insert_migration_table_rows(ctx, table, import_mode)?;
|
||||
insert_migration_table_rows(ctx, table, import_mode, &mut warnings)?;
|
||||
stats.push(DatabaseMigrationTableStat {
|
||||
table_name: table.name.clone(),
|
||||
exported_row_count: 0,
|
||||
@@ -640,7 +1039,7 @@ fn apply_migration_file(
|
||||
});
|
||||
}
|
||||
|
||||
Ok(stats)
|
||||
Ok((stats, warnings))
|
||||
}
|
||||
|
||||
fn build_import_table_name_set(
|
||||
@@ -655,37 +1054,192 @@ fn build_import_table_name_set(
|
||||
.collect()
|
||||
}
|
||||
|
||||
fn build_dropped_table_warning(table: &MigrationTable) -> DatabaseMigrationWarning {
|
||||
DatabaseMigrationWarning {
|
||||
table_name: table.name.clone(),
|
||||
warning_kind: "dropped_table".to_string(),
|
||||
message: format!(
|
||||
"迁移文件包含当前模块已删除或未加入白名单的表 {},已跳过 {} 行",
|
||||
table.name,
|
||||
table.rows.len()
|
||||
),
|
||||
}
|
||||
}
|
||||
|
||||
fn build_dropped_field_warning(table_name: &str, field_name: &str) -> DatabaseMigrationWarning {
|
||||
DatabaseMigrationWarning {
|
||||
table_name: table_name.to_string(),
|
||||
warning_kind: "dropped_field".to_string(),
|
||||
message: format!("表 {table_name} 的旧字段 {field_name} 当前已不存在,已在导入时丢弃"),
|
||||
}
|
||||
}
|
||||
|
||||
fn row_to_json<T: spacetimedb::Serialize>(row: &T) -> Result<serde_json::Value, String> {
|
||||
serde_json::to_value(SerializeWrapper::from_ref(row))
|
||||
.map_err(|error| format!("迁移行序列化失败: {error}"))
|
||||
}
|
||||
|
||||
fn row_from_json<T>(value: &serde_json::Value) -> Result<T, String>
|
||||
fn row_from_json<T>(
|
||||
table_name: &str,
|
||||
value: &serde_json::Value,
|
||||
warnings: &mut Vec<DatabaseMigrationWarning>,
|
||||
) -> Result<T, String>
|
||||
where
|
||||
T: for<'de> spacetimedb::Deserialize<'de>,
|
||||
{
|
||||
let wrapped: DeserializeWrapper<T> = serde_json::from_value(value.clone())
|
||||
.map_err(|error| format!("迁移行反序列化失败: {error}"))?;
|
||||
let wrapped = match serde_json::from_value::<DeserializeWrapper<T>>(value.clone()) {
|
||||
Ok(row) => row,
|
||||
Err(original_error) => recover_row_with_deleted_fields::<T>(
|
||||
table_name,
|
||||
value,
|
||||
&original_error.to_string(),
|
||||
warnings,
|
||||
)
|
||||
.ok_or_else(|| format!("迁移行反序列化失败,且无法通过丢弃旧字段恢复: {original_error}"))?,
|
||||
};
|
||||
Ok(wrapped.0)
|
||||
}
|
||||
|
||||
fn normalize_migration_row(table_name: &str, value: &serde_json::Value) -> serde_json::Value {
|
||||
let mut next_value = value.clone();
|
||||
if table_name == "user_account" {
|
||||
if let Some(object) = next_value.as_object_mut() {
|
||||
// 中文注释:头像字段晚于认证拆表加入,旧迁移包按未设置头像兼容。
|
||||
object
|
||||
.entry("avatar_url".to_string())
|
||||
.or_insert(serde_json::Value::Null);
|
||||
}
|
||||
}
|
||||
if table_name == "profile_invite_code" {
|
||||
if let Some(object) = next_value.as_object_mut() {
|
||||
// 中文注释:邀请码 metadata 晚于邀请表加入,旧迁移包按空对象兼容。
|
||||
object
|
||||
.entry("metadata_json".to_string())
|
||||
.or_insert_with(|| serde_json::Value::String("{}".to_string()));
|
||||
}
|
||||
}
|
||||
if table_name == "big_fish_creation_session" {
|
||||
if let Some(object) = next_value.as_object_mut() {
|
||||
// 中文注释:旧迁移包没有公开游玩次数字段,导入时按新建作品默认 0 兼容。
|
||||
object
|
||||
.entry("play_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("remix_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("like_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("published_at".to_string())
|
||||
.or_insert(serde_json::Value::Null);
|
||||
}
|
||||
}
|
||||
if table_name == "custom_world_profile" || table_name == "custom_world_gallery_entry" {
|
||||
if let Some(object) = next_value.as_object_mut() {
|
||||
// 中文注释:自定义世界公开互动计数字段晚于基础作品表加入,旧迁移包按 0 兼容。
|
||||
object
|
||||
.entry("play_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("remix_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("like_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
}
|
||||
}
|
||||
if table_name == "puzzle_work_profile" {
|
||||
if let Some(object) = next_value.as_object_mut() {
|
||||
// 中文注释:拼图公开互动计数晚于基础作品表加入,旧迁移包按 0 兼容。
|
||||
object
|
||||
.entry("play_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("remix_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("like_count".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("point_incentive_total_half_points".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
object
|
||||
.entry("point_incentive_claimed_points".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(0));
|
||||
// 中文注释:拼图多关卡字段晚于旧作品表加入,旧迁移包留空并由读取层补出首关。
|
||||
object
|
||||
.entry("levels_json".to_string())
|
||||
.or_insert_with(|| serde_json::Value::from(""));
|
||||
// 中文注释:作品名称/描述从旧关卡名/画面摘要拆出,旧行保留旧值做兼容回填。
|
||||
let fallback_title = object
|
||||
.get("level_name")
|
||||
.cloned()
|
||||
.unwrap_or_else(|| serde_json::Value::from(""));
|
||||
object
|
||||
.entry("work_title".to_string())
|
||||
.or_insert(fallback_title);
|
||||
let fallback_description = object
|
||||
.get("summary")
|
||||
.cloned()
|
||||
.unwrap_or_else(|| serde_json::Value::from(""));
|
||||
object
|
||||
.entry("work_description".to_string())
|
||||
.or_insert(fallback_description);
|
||||
}
|
||||
}
|
||||
next_value
|
||||
}
|
||||
|
||||
fn recover_row_with_deleted_fields<T>(
|
||||
table_name: &str,
|
||||
value: &serde_json::Value,
|
||||
error_message: &str,
|
||||
warnings: &mut Vec<DatabaseMigrationWarning>,
|
||||
) -> Option<DeserializeWrapper<T>>
|
||||
where
|
||||
T: for<'de> spacetimedb::Deserialize<'de>,
|
||||
{
|
||||
let mut candidate = value.as_object()?.clone();
|
||||
let mut next_error = error_message.to_string();
|
||||
|
||||
loop {
|
||||
let field_name = extract_unknown_field_name(&next_error)?;
|
||||
candidate.remove(&field_name)?;
|
||||
warnings.push(build_dropped_field_warning(table_name, &field_name));
|
||||
|
||||
match serde_json::from_value::<DeserializeWrapper<T>>(serde_json::Value::Object(
|
||||
candidate.clone(),
|
||||
)) {
|
||||
Ok(row) => return Some(row),
|
||||
Err(error) => next_error = error.to_string(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn extract_unknown_field_name(error_message: &str) -> Option<String> {
|
||||
let marker = "unknown field";
|
||||
let marker_index = error_message.find(marker)?;
|
||||
let after_marker = error_message[marker_index + marker.len()..].trim_start();
|
||||
|
||||
for quote in ['`', '"', '\''] {
|
||||
if let Some(rest) = after_marker.strip_prefix(quote) {
|
||||
let end_index = rest.find(quote)?;
|
||||
return Some(rest[..end_index].to_string());
|
||||
}
|
||||
}
|
||||
|
||||
after_marker
|
||||
.split(|character: char| !character.is_ascii_alphanumeric() && character != '_')
|
||||
.find(|value| !value.is_empty())
|
||||
.map(str::to_string)
|
||||
}
|
||||
|
||||
fn insert_migration_table_rows(
|
||||
ctx: &ReducerContext,
|
||||
table: &MigrationTable,
|
||||
import_mode: DatabaseMigrationImportMode,
|
||||
warnings: &mut Vec<DatabaseMigrationWarning>,
|
||||
) -> Result<(u64, u64), String> {
|
||||
macro_rules! insert_table_match_arm {
|
||||
($($table:ident),+ $(,)?) => {
|
||||
@@ -696,7 +1250,7 @@ fn insert_migration_table_rows(
|
||||
let mut skipped = 0u64;
|
||||
for value in &table.rows {
|
||||
let normalized_value = normalize_migration_row(stringify!($table), value);
|
||||
let row = row_from_json(&normalized_value)
|
||||
let row = row_from_json(stringify!($table), &normalized_value, warnings)
|
||||
.map_err(|error| format!("{}: {error}", stringify!($table)))?;
|
||||
let insert_result = ctx.db
|
||||
.$table()
|
||||
|
||||
Reference in New Issue
Block a user