mirror of https://github.com/nocodb/nocodb
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
903 lines
31 KiB
903 lines
31 KiB
import type { ColumnType, LinkToAnotherRecordType, PaginatedType, RelationTypes, TableType, ViewType } from 'nocodb-sdk' |
|
import { UITypes, isCreatedOrLastModifiedByCol, isCreatedOrLastModifiedTimeCol } from 'nocodb-sdk' |
|
import type { ComputedRef, Ref } from 'vue' |
|
import { |
|
NOCO, |
|
computed, |
|
extractPk, |
|
extractPkFromRow, |
|
extractSdkResponseErrorMsg, |
|
findIndexByPk, |
|
message, |
|
populateInsertObject, |
|
rowDefaultData, |
|
rowPkData, |
|
storeToRefs, |
|
until, |
|
useBase, |
|
useI18n, |
|
useMetas, |
|
useNuxtApp, |
|
} from '#imports' |
|
import type { CellRange, Row, UndoRedoAction } from '#imports' |
|
|
|
export function useData(args: { |
|
meta: Ref<TableType | undefined> | ComputedRef<TableType | undefined> |
|
viewMeta: Ref<ViewType | undefined> | ComputedRef<(ViewType & { id: string }) | undefined> |
|
formattedData: Ref<Row[]> |
|
paginationData: Ref<PaginatedType> |
|
callbacks?: { |
|
changePage?: (page: number) => Promise<void> |
|
loadData?: () => Promise<void> |
|
globalCallback?: (...args: any[]) => void |
|
syncCount?: () => Promise<void> |
|
syncPagination?: () => Promise<void> |
|
} |
|
}) { |
|
const { meta, viewMeta, formattedData, paginationData, callbacks } = args |
|
|
|
const { t } = useI18n() |
|
|
|
const { getMeta, metas } = useMetas() |
|
|
|
const { addUndo, clone, defineViewScope } = useUndoRedo() |
|
|
|
const { base } = storeToRefs(useBase()) |
|
|
|
const { $api } = useNuxtApp() |
|
|
|
const { isPaginationLoading } = storeToRefs(useViewsStore()) |
|
|
|
const selectedAllRecords = computed({ |
|
get() { |
|
return !!formattedData.value.length && formattedData.value.every((row: Row) => row.rowMeta.selected) |
|
}, |
|
set(selected: boolean) { |
|
formattedData.value.forEach((row: Row) => (row.rowMeta.selected = selected)) |
|
}, |
|
}) |
|
|
|
function addEmptyRow(addAfter = formattedData.value.length, metaValue = meta.value) { |
|
formattedData.value.splice(addAfter, 0, { |
|
row: { ...rowDefaultData(metaValue?.columns) }, |
|
oldRow: {}, |
|
rowMeta: { new: true }, |
|
}) |
|
|
|
return formattedData.value[addAfter] |
|
} |
|
|
|
async function insertRow( |
|
currentRow: Row, |
|
ltarState: Record<string, any> = {}, |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
undo = false, |
|
) { |
|
const row = currentRow.row |
|
if (currentRow.rowMeta) currentRow.rowMeta.saving = true |
|
try { |
|
const { missingRequiredColumns, insertObj } = await populateInsertObject({ |
|
meta: metaValue!, |
|
ltarState, |
|
getMeta, |
|
row, |
|
undo, |
|
}) |
|
|
|
if (missingRequiredColumns.size) return |
|
|
|
const insertedData = await $api.dbViewRow.create( |
|
NOCO, |
|
base?.value.id as string, |
|
metaValue?.id as string, |
|
viewMetaValue?.id as string, |
|
{ ...insertObj, ...(ltarState || {}) }, |
|
) |
|
|
|
if (!undo) { |
|
Object.assign(currentRow, { |
|
row: { ...insertedData, ...row }, |
|
rowMeta: { ...(currentRow.rowMeta || {}), new: undefined }, |
|
oldRow: { ...insertedData }, |
|
}) |
|
|
|
const id = extractPkFromRow(insertedData, metaValue?.columns as ColumnType[]) |
|
const pkData = rowPkData(insertedData, metaValue?.columns as ColumnType[]) |
|
const rowIndex = findIndexByPk(pkData, formattedData.value) |
|
|
|
addUndo({ |
|
redo: { |
|
fn: async function redo( |
|
this: UndoRedoAction, |
|
row: Row, |
|
ltarState: Record<string, any>, |
|
pg: { page: number; pageSize: number }, |
|
) { |
|
row.row = { ...pkData, ...row.row } |
|
const insertedData = await insertRow(row, ltarState, undefined, true) |
|
|
|
if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { |
|
if (pg.page === paginationData.value.page) { |
|
formattedData.value.splice(rowIndex, 0, { |
|
row: { ...row, ...insertedData }, |
|
rowMeta: row.rowMeta, |
|
oldRow: row.oldRow, |
|
}) |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
} else { |
|
await callbacks?.loadData?.() |
|
} |
|
}, |
|
args: [ |
|
clone(currentRow), |
|
clone(ltarState), |
|
{ page: paginationData.value.page, pageSize: paginationData.value.pageSize }, |
|
], |
|
}, |
|
undo: { |
|
fn: async function undo(this: UndoRedoAction, id: string) { |
|
await deleteRowById(id) |
|
if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) |
|
paginationData.value.totalRows = paginationData.value.totalRows! - 1 |
|
}, |
|
args: [id], |
|
}, |
|
scope: defineViewScope({ view: viewMeta.value }), |
|
}) |
|
} |
|
|
|
await callbacks?.syncCount?.() |
|
return insertedData |
|
} catch (error: any) { |
|
message.error(await extractSdkResponseErrorMsg(error)) |
|
} finally { |
|
if (currentRow.rowMeta) currentRow.rowMeta.saving = false |
|
await callbacks?.globalCallback?.() |
|
} |
|
} |
|
|
|
async function bulkInsertRows( |
|
rows: Row[], |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
undo = false, |
|
) { |
|
isPaginationLoading.value = true |
|
|
|
const autoGeneratedKeys = clone(metaValue?.columns || []) |
|
.filter((c) => !c.pk && (isCreatedOrLastModifiedTimeCol(c) || isCreatedOrLastModifiedByCol(c))) |
|
.map((c) => c.title) |
|
|
|
try { |
|
const rowsToInsert = |
|
( |
|
await Promise.all( |
|
rows.map(async (currentRow) => { |
|
const { missingRequiredColumns, insertObj } = await populateInsertObject({ |
|
meta: metaValue!, |
|
ltarState: {}, |
|
getMeta, |
|
row: currentRow.row, |
|
undo, |
|
}) |
|
|
|
if (missingRequiredColumns.size === 0) { |
|
autoGeneratedKeys.forEach((key) => delete insertObj[key!]) |
|
return insertObj |
|
} |
|
}), |
|
) |
|
)?.filter(Boolean) ?? [] // Filter out undefined values (if any) |
|
|
|
const bulkInsertedIds = await $api.dbDataTableRow.create(metaValue?.id as string, rowsToInsert, { |
|
viewId: viewMetaValue?.id as string, |
|
}) |
|
|
|
await callbacks?.syncCount?.() |
|
return bulkInsertedIds |
|
} catch (error: any) { |
|
message.error(await extractSdkResponseErrorMsg(error)) |
|
} finally { |
|
await callbacks?.globalCallback?.() |
|
isPaginationLoading.value = false |
|
} |
|
} |
|
|
|
// inside this method use metaValue and viewMetaValue to refer meta |
|
// since sometimes we need to pass old metas |
|
async function updateRowProperty( |
|
toUpdate: Row, |
|
property: string, |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
undo = false, |
|
) { |
|
if (toUpdate.rowMeta) toUpdate.rowMeta.saving = true |
|
|
|
try { |
|
const id = extractPkFromRow(toUpdate.row, metaValue?.columns as ColumnType[]) |
|
|
|
const updatedRowData: Record<string, any> = await $api.dbViewRow.update( |
|
NOCO, |
|
base?.value.id as string, |
|
metaValue?.id as string, |
|
viewMetaValue?.id as string, |
|
id, |
|
{ |
|
// if value is undefined treat it as null |
|
[property]: toUpdate.row[property] ?? null, |
|
}, |
|
// todo: |
|
// { |
|
// query: { ignoreWebhook: !saved } |
|
// } |
|
) |
|
|
|
if (!undo) { |
|
addUndo({ |
|
redo: { |
|
fn: async function redo(toUpdate: Row, property: string, pg: { page: number; pageSize: number }) { |
|
const updatedData = await updateRowProperty(toUpdate, property, undefined, true) |
|
if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { |
|
const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) |
|
if (rowIndex !== -1) { |
|
const row = formattedData.value[rowIndex] |
|
Object.assign(row.row, updatedData) |
|
Object.assign(row.oldRow, updatedData) |
|
} else { |
|
await callbacks?.loadData?.() |
|
} |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
}, |
|
args: [clone(toUpdate), property, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], |
|
}, |
|
undo: { |
|
fn: async function undo(toUpdate: Row, property: string, pg: { page: number; pageSize: number }) { |
|
const updatedData = await updateRowProperty( |
|
{ row: toUpdate.oldRow, oldRow: toUpdate.row, rowMeta: toUpdate.rowMeta }, |
|
property, |
|
undefined, |
|
true, |
|
) |
|
if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { |
|
const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) |
|
if (rowIndex !== -1) { |
|
const row = formattedData.value[rowIndex] |
|
Object.assign(row.row, updatedData) |
|
Object.assign(row.oldRow, updatedData) |
|
} else { |
|
await callbacks?.loadData?.() |
|
} |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
}, |
|
args: [clone(toUpdate), property, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], |
|
}, |
|
scope: defineViewScope({ view: viewMeta.value }), |
|
}) |
|
|
|
/** update row data(to sync formula and other related columns) |
|
* update only formula, rollup and auto updated datetime columns data to avoid overwriting any changes made by user |
|
*/ |
|
Object.assign( |
|
toUpdate.row, |
|
metaValue!.columns!.reduce<Record<string, any>>((acc: Record<string, any>, col: ColumnType) => { |
|
if ( |
|
col.title in updatedRowData && |
|
(col.uidt === UITypes.Formula || |
|
col.uidt === UITypes.QrCode || |
|
col.uidt === UITypes.Barcode || |
|
col.uidt === UITypes.Rollup || |
|
col.uidt === UITypes.Checkbox || |
|
col.uidt === UITypes.User || |
|
col.uidt === UITypes.LastModifiedTime || |
|
col.uidt === UITypes.LastModifiedBy || |
|
col.uidt === UITypes.Lookup || |
|
col.au || |
|
(col.cdf && / on update /i.test(col.cdf))) |
|
) |
|
acc[col.title!] = updatedRowData[col.title!] |
|
return acc |
|
}, {} as Record<string, any>), |
|
) |
|
Object.assign(toUpdate.oldRow, updatedRowData) |
|
} |
|
|
|
await callbacks?.globalCallback?.() |
|
|
|
return updatedRowData |
|
} catch (e: any) { |
|
message.error(`${t('msg.error.rowUpdateFailed')} ${await extractSdkResponseErrorMsg(e)}`) |
|
} finally { |
|
if (toUpdate.rowMeta) toUpdate.rowMeta.saving = false |
|
} |
|
} |
|
|
|
async function updateOrSaveRow( |
|
row: Row, |
|
property?: string, |
|
ltarState?: Record<string, any>, |
|
args: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
) { |
|
// update changed status |
|
if (row.rowMeta) row.rowMeta.changed = false |
|
|
|
// if new row and save is in progress then wait until the save is complete |
|
await until(() => !(row.rowMeta?.new && row.rowMeta?.saving)).toMatch((v) => v) |
|
|
|
if (row.rowMeta.new) { |
|
return await insertRow(row, ltarState, args) |
|
} else { |
|
// if the field name is missing skip update |
|
if (property) { |
|
await updateRowProperty(row, property, args) |
|
} |
|
} |
|
} |
|
|
|
async function bulkUpdateRows( |
|
rows: Row[], |
|
props: string[], |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
undo = false, |
|
) { |
|
const promises = [] |
|
|
|
for (const row of rows) { |
|
// update changed status |
|
if (row.rowMeta) row.rowMeta.changed = false |
|
|
|
// if new row and save is in progress then wait until the save is complete |
|
promises.push(until(() => !(row.rowMeta?.new && row.rowMeta?.saving)).toMatch((v) => v)) |
|
} |
|
|
|
await Promise.all(promises) |
|
|
|
const updateArray = [] |
|
|
|
for (const row of rows) { |
|
if (row.rowMeta) row.rowMeta.saving = true |
|
|
|
const pk = rowPkData(row.row, metaValue?.columns as ColumnType[]) |
|
|
|
const updateData = props.reduce((acc: Record<string, any>, prop) => { |
|
acc[prop] = row.row[prop] |
|
return acc |
|
}, {} as Record<string, any>) |
|
|
|
updateArray.push({ ...updateData, ...pk }) |
|
} |
|
|
|
await $api.dbTableRow.bulkUpdate(NOCO, metaValue?.base_id as string, metaValue?.id as string, updateArray) |
|
|
|
if (!undo) { |
|
addUndo({ |
|
redo: { |
|
fn: async function redo(redoRows: Row[], props: string[], pg: { page: number; pageSize: number }) { |
|
await bulkUpdateRows(redoRows, props, { metaValue, viewMetaValue }, true) |
|
if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { |
|
for (const toUpdate of redoRows) { |
|
const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) |
|
if (rowIndex !== -1) { |
|
const row = formattedData.value[rowIndex] |
|
Object.assign(row.row, toUpdate.row) |
|
Object.assign(row.oldRow, toUpdate.row) |
|
} else { |
|
await callbacks?.loadData?.() |
|
break |
|
} |
|
} |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
}, |
|
args: [clone(rows), clone(props), { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], |
|
}, |
|
undo: { |
|
fn: async function undo(undoRows: Row[], props: string[], pg: { page: number; pageSize: number }) { |
|
await bulkUpdateRows(undoRows, props, { metaValue, viewMetaValue }, true) |
|
if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { |
|
for (const toUpdate of undoRows) { |
|
const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) |
|
if (rowIndex !== -1) { |
|
const row = formattedData.value[rowIndex] |
|
Object.assign(row.row, toUpdate.row) |
|
Object.assign(row.oldRow, toUpdate.row) |
|
} else { |
|
await callbacks?.loadData?.() |
|
break |
|
} |
|
} |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
}, |
|
args: [ |
|
clone( |
|
rows.map((row) => { |
|
return { row: row.oldRow, oldRow: row.row, rowMeta: row.rowMeta } |
|
}), |
|
), |
|
props, |
|
{ page: paginationData.value.page, pageSize: paginationData.value.pageSize }, |
|
], |
|
}, |
|
scope: defineViewScope({ view: viewMetaValue }), |
|
}) |
|
} |
|
|
|
for (const row of rows) { |
|
if (row.rowMeta) row.rowMeta.saving = false |
|
} |
|
|
|
// reload data since row update may change the other columns data( Formula, QrCode, Barcode, Rollup, Checkbox, User, Auto Updated Datetime, etc...) |
|
if ( |
|
metaValue!.columns!.some((col) => |
|
[ |
|
UITypes.QrCode, |
|
UITypes.LastModifiedTime, |
|
UITypes.Barcode, |
|
UITypes.Formula, |
|
UITypes.Lookup, |
|
UITypes.Rollup, |
|
UITypes.LinkToAnotherRecord, |
|
UITypes.LastModifiedBy, |
|
].includes(col.uidt), |
|
) |
|
) { |
|
await callbacks?.loadData?.() |
|
} |
|
await callbacks?.globalCallback?.() |
|
} |
|
|
|
async function bulkUpdateView( |
|
data: Record<string, any>[], |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
) { |
|
if (!viewMetaValue) return |
|
|
|
await $api.dbTableRow.bulkUpdateAll(NOCO, metaValue?.base_id as string, metaValue?.id as string, data, { |
|
viewId: viewMetaValue.id, |
|
}) |
|
|
|
await callbacks?.loadData?.() |
|
await callbacks?.globalCallback?.() |
|
} |
|
|
|
const linkRecord = async ( |
|
rowId: string, |
|
relatedRowId: string, |
|
column: ColumnType, |
|
type: RelationTypes, |
|
{ metaValue = meta.value }: { metaValue?: TableType } = {}, |
|
) => { |
|
try { |
|
await $api.dbTableRow.nestedAdd( |
|
NOCO, |
|
base.value.title as string, |
|
metaValue?.title as string, |
|
encodeURIComponent(rowId), |
|
type as 'mm' | 'hm', |
|
column.title as string, |
|
encodeURIComponent(relatedRowId), |
|
) |
|
} catch (e: any) { |
|
message.error(await extractSdkResponseErrorMsg(e)) |
|
} |
|
} |
|
|
|
// Recover LTAR relations for a row using the row data |
|
const recoverLTARRefs = async (row: Record<string, any>, { metaValue = meta.value }: { metaValue?: TableType } = {}) => { |
|
const id = extractPkFromRow(row, metaValue?.columns as ColumnType[]) |
|
for (const column of metaValue?.columns ?? []) { |
|
if (column.uidt !== UITypes.LinkToAnotherRecord) continue |
|
|
|
const colOptions = column.colOptions as LinkToAnotherRecordType |
|
|
|
const relatedTableMeta = metas.value?.[colOptions?.fk_related_model_id as string] |
|
|
|
if (isHm(column) || isMm(column)) { |
|
const relatedRows = (row[column.title!] ?? []) as Record<string, any>[] |
|
|
|
for (const relatedRow of relatedRows) { |
|
await linkRecord( |
|
id, |
|
extractPkFromRow(relatedRow, relatedTableMeta.columns as ColumnType[]), |
|
column, |
|
colOptions.type as RelationTypes, |
|
{ metaValue }, |
|
) |
|
} |
|
} else if (isBt(column) && row[column.title!]) { |
|
await linkRecord( |
|
id, |
|
extractPkFromRow(row[column.title!] as Record<string, any>, relatedTableMeta.columns as ColumnType[]), |
|
column, |
|
colOptions.type as RelationTypes, |
|
{ metaValue }, |
|
) |
|
} |
|
} |
|
} |
|
|
|
async function deleteRowById( |
|
id: string, |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
) { |
|
if (!id) { |
|
throw new Error("Delete not allowed for table which doesn't have primary Key") |
|
} |
|
|
|
const res: any = await $api.dbViewRow.delete( |
|
'noco', |
|
base.value.id as string, |
|
metaValue?.id as string, |
|
viewMetaValue?.id as string, |
|
encodeURIComponent(id), |
|
) |
|
|
|
if (res.message) { |
|
message.info( |
|
`Record delete failed: ${`Unable to delete record with ID ${id} because of the following: |
|
\n${res.message.join('\n')}.\n |
|
Clear the data first & try again`})}`, |
|
) |
|
return false |
|
} |
|
return true |
|
} |
|
|
|
async function deleteRow(rowIndex: number, undo?: boolean) { |
|
try { |
|
const row = formattedData.value[rowIndex] |
|
if (!row.rowMeta.new) { |
|
const id = meta?.value?.columns |
|
?.filter((c) => c.pk) |
|
.map((c) => row.row[c.title!]) |
|
.join('___') |
|
|
|
const deleted = await deleteRowById(id as string) |
|
if (!deleted) { |
|
return |
|
} |
|
|
|
if (!undo) { |
|
addUndo({ |
|
redo: { |
|
fn: async function redo(this: UndoRedoAction, id: string) { |
|
await deleteRowById(id) |
|
const pk: Record<string, string> = rowPkData(row.row, meta?.value?.columns as ColumnType[]) |
|
const rowIndex = findIndexByPk(pk, formattedData.value) |
|
if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) |
|
paginationData.value.totalRows = paginationData.value.totalRows! - 1 |
|
}, |
|
args: [id], |
|
}, |
|
undo: { |
|
fn: async function undo( |
|
this: UndoRedoAction, |
|
row: Row, |
|
ltarState: Record<string, any>, |
|
pg: { page: number; pageSize: number }, |
|
) { |
|
const pkData = rowPkData(row.row, meta.value?.columns as ColumnType[]) |
|
row.row = { ...pkData, ...row.row } |
|
await insertRow(row, ltarState, {}, true) |
|
recoverLTARRefs(row.row) |
|
if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { |
|
if (pg.page === paginationData.value.page) { |
|
formattedData.value.splice(rowIndex, 0, row) |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
} else { |
|
await callbacks?.loadData?.() |
|
} |
|
}, |
|
args: [clone(row), {}, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], |
|
}, |
|
scope: defineViewScope({ view: viewMeta.value }), |
|
}) |
|
} |
|
} |
|
|
|
formattedData.value.splice(rowIndex, 1) |
|
|
|
await callbacks?.syncCount?.() |
|
} catch (e: any) { |
|
message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) |
|
} |
|
|
|
await callbacks?.globalCallback?.() |
|
} |
|
|
|
async function deleteSelectedRows() { |
|
let row = formattedData.value.length |
|
let removedRowsData: Record<string, any>[] = [] |
|
let compositePrimaryKey = '' |
|
|
|
while (row--) { |
|
const { row: rowObj, rowMeta } = formattedData.value[row] as Record<string, any> |
|
if (!rowMeta.selected) { |
|
continue |
|
} |
|
if (!rowMeta.new) { |
|
const extractedPk = extractPk(meta?.value?.columns as ColumnType[]) |
|
const compositePkValue = extractPkFromRow(rowObj, meta?.value?.columns as ColumnType[]) |
|
|
|
if (extractedPk && compositePkValue) { |
|
if (!compositePrimaryKey) compositePrimaryKey = extractedPk |
|
|
|
removedRowsData.push({ |
|
[compositePrimaryKey]: compositePkValue as string, |
|
row: clone(formattedData.value[row]) as Row, |
|
rowIndex: row as number, |
|
}) |
|
} |
|
} |
|
} |
|
|
|
if (!removedRowsData.length) return |
|
|
|
try { |
|
const removedRowIds: Record<string, any>[] = await bulkDeleteRows( |
|
removedRowsData.map((row) => ({ [compositePrimaryKey]: row[compositePrimaryKey] as string })), |
|
) |
|
|
|
if (Array.isArray(removedRowIds)) { |
|
const removedRowsDataSet = new Set(removedRowIds.map((row) => row[compositePrimaryKey])) |
|
|
|
removedRowsData = removedRowsData.filter((row) => removedRowsDataSet.has(row[compositePrimaryKey] as string)) |
|
|
|
const rowIndexesSet = new Set(removedRowsData.map((row) => row.rowIndex)) |
|
formattedData.value = formattedData.value.filter((_, index) => rowIndexesSet.has(index)) |
|
} else { |
|
removedRowsData = [] |
|
} |
|
} catch (e: any) { |
|
return message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) |
|
} |
|
|
|
if (!removedRowsData.length) return |
|
|
|
addUndo({ |
|
redo: { |
|
fn: async function redo(this: UndoRedoAction, removedRowsData: Record<string, any>[], compositePrimaryKey: string) { |
|
const removedRowIds = await bulkDeleteRows( |
|
removedRowsData.map((row) => ({ [compositePrimaryKey]: row[compositePrimaryKey] as string })), |
|
) |
|
|
|
if (Array.isArray(removedRowIds)) { |
|
for (const { row } of removedRowsData) { |
|
const primaryKey: Record<string, string> = rowPkData(row.row, meta?.value?.columns as ColumnType[]) |
|
const rowIndex = findIndexByPk(primaryKey, formattedData.value) |
|
if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) |
|
paginationData.value.totalRows = paginationData.value.totalRows! - 1 |
|
} |
|
} |
|
|
|
await callbacks?.syncPagination?.() |
|
}, |
|
args: [removedRowsData, compositePrimaryKey], |
|
}, |
|
undo: { |
|
fn: async function undo( |
|
this: UndoRedoAction, |
|
removedRowsData: Record<string, any>[], |
|
pg: { page: number; pageSize: number }, |
|
) { |
|
const rowsToInsert = removedRowsData |
|
.map((row) => { |
|
const pkData = rowPkData(row.row, meta.value?.columns as ColumnType[]) |
|
row.row = { ...pkData, ...row.row } |
|
return row |
|
}) |
|
.reverse() |
|
|
|
const insertedRowIds = await bulkInsertRows( |
|
rowsToInsert.map((row) => row.row), |
|
undefined, |
|
true, |
|
) |
|
|
|
if (Array.isArray(insertedRowIds)) { |
|
for (const { row, rowIndex } of rowsToInsert) { |
|
recoverLTARRefs(row.row) |
|
|
|
if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { |
|
if (pg.page === paginationData.value.page) { |
|
formattedData.value.splice(rowIndex, 0, row) |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
} else { |
|
await callbacks?.loadData?.() |
|
} |
|
} |
|
} |
|
}, |
|
args: [removedRowsData, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], |
|
}, |
|
scope: defineViewScope({ view: viewMeta.value }), |
|
}) |
|
|
|
await callbacks?.syncCount?.() |
|
await callbacks?.syncPagination?.() |
|
await callbacks?.globalCallback?.() |
|
} |
|
|
|
async function deleteRangeOfRows(cellRange: CellRange) { |
|
if (!cellRange._start || !cellRange._end) return |
|
const start = Math.max(cellRange._start.row, cellRange._end.row) |
|
const end = Math.min(cellRange._start.row, cellRange._end.row) |
|
|
|
// plus one because we want to include the end row |
|
let row = start + 1 |
|
|
|
let removedRowsData: Record<string, any>[] = [] |
|
let compositePrimaryKey = '' |
|
|
|
while (row--) { |
|
try { |
|
const { row: rowObj, rowMeta } = formattedData.value[row] as Record<string, any> |
|
|
|
if (!rowMeta.new) { |
|
const extractedPk = extractPk(meta?.value?.columns as ColumnType[]) |
|
const compositePkValue = extractPkFromRow(rowObj, meta?.value?.columns as ColumnType[]) |
|
|
|
if (extractedPk && compositePkValue) { |
|
if (!compositePrimaryKey) compositePrimaryKey = extractedPk |
|
|
|
removedRowsData.push({ |
|
[compositePrimaryKey]: compositePkValue as string, |
|
row: clone(formattedData.value[row]) as Row, |
|
rowIndex: row as number, |
|
}) |
|
} |
|
} |
|
} catch (e: any) { |
|
return message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) |
|
} |
|
|
|
if (row === end) break |
|
} |
|
|
|
if (!removedRowsData.length) return |
|
|
|
try { |
|
const removedRowIds: Record<string, any>[] = await bulkDeleteRows( |
|
removedRowsData.map((row) => ({ [compositePrimaryKey]: row[compositePrimaryKey] as string })), |
|
) |
|
|
|
if (Array.isArray(removedRowIds)) { |
|
const removedRowsDataSet = new Set(removedRowIds.map((row) => row[compositePrimaryKey])) |
|
|
|
removedRowsData = removedRowsData.filter((row) => removedRowsDataSet.has(row[compositePrimaryKey] as string)) |
|
|
|
const rowIndexesSet = new Set(removedRowsData.map((row) => row.rowIndex)) |
|
formattedData.value = formattedData.value.filter((_, index) => rowIndexesSet.has(index)) |
|
} else { |
|
removedRowsData = [] |
|
} |
|
} catch (e: any) { |
|
return message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) |
|
} |
|
|
|
if (!removedRowsData.length) return |
|
|
|
addUndo({ |
|
redo: { |
|
fn: async function redo(this: UndoRedoAction, removedRowsData: Record<string, any>[], compositePrimaryKey: string) { |
|
const removedRowIds = await bulkDeleteRows( |
|
removedRowsData.map((row) => ({ [compositePrimaryKey]: row[compositePrimaryKey] as string })), |
|
) |
|
|
|
if (Array.isArray(removedRowIds)) { |
|
for (const { row } of removedRowsData) { |
|
const primaryKey: Record<string, string> = rowPkData(row.row, meta?.value?.columns as ColumnType[]) |
|
const rowIndex = findIndexByPk(primaryKey, formattedData.value) |
|
if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) |
|
paginationData.value.totalRows = paginationData.value.totalRows! - 1 |
|
} |
|
} |
|
|
|
await callbacks?.syncPagination?.() |
|
}, |
|
args: [removedRowsData, compositePrimaryKey], |
|
}, |
|
undo: { |
|
fn: async function undo( |
|
this: UndoRedoAction, |
|
removedRowsData: Record<string, any>[], |
|
pg: { page: number; pageSize: number }, |
|
) { |
|
const rowsToInsert = removedRowsData |
|
.map((row) => { |
|
const pkData = rowPkData(row.row, meta.value?.columns as ColumnType[]) |
|
row.row = { ...pkData, ...row.row } |
|
return row |
|
}) |
|
.reverse() |
|
|
|
const insertedRowIds = await bulkInsertRows( |
|
rowsToInsert.map((row) => row.row), |
|
undefined, |
|
true, |
|
) |
|
|
|
if (Array.isArray(insertedRowIds)) { |
|
for (const { row, rowIndex } of rowsToInsert) { |
|
recoverLTARRefs(row.row) |
|
|
|
if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { |
|
if (pg.page === paginationData.value.page) { |
|
formattedData.value.splice(rowIndex, 0, row) |
|
} else { |
|
await callbacks?.changePage?.(pg.page) |
|
} |
|
} else { |
|
await callbacks?.loadData?.() |
|
} |
|
} |
|
} |
|
}, |
|
args: [removedRowsData, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], |
|
}, |
|
scope: defineViewScope({ view: viewMeta.value }), |
|
}) |
|
|
|
await callbacks?.syncCount?.() |
|
await callbacks?.syncPagination?.() |
|
await callbacks?.globalCallback?.() |
|
} |
|
|
|
async function bulkDeleteRows( |
|
rows: Record<string, string>[], |
|
{ metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, |
|
) { |
|
isPaginationLoading.value = true |
|
try { |
|
const bulkDeletedRowsData = await $api.dbDataTableRow.delete(metaValue?.id as string, rows.length === 1 ? rows[0] : rows, { |
|
viewId: viewMetaValue?.id as string, |
|
}) |
|
|
|
await callbacks?.syncCount?.() |
|
return rows.length === 1 && bulkDeletedRowsData ? [bulkDeletedRowsData] : bulkDeletedRowsData |
|
} catch (error: any) { |
|
message.error(await extractSdkResponseErrorMsg(error)) |
|
} finally { |
|
await callbacks?.globalCallback?.() |
|
isPaginationLoading.value = false |
|
} |
|
} |
|
|
|
const removeRowIfNew = (row: Row) => { |
|
const index = formattedData.value.indexOf(row) |
|
|
|
if (index > -1 && row.rowMeta.new) { |
|
formattedData.value.splice(index, 1) |
|
return true |
|
} |
|
return false |
|
} |
|
|
|
return { |
|
insertRow, |
|
updateRowProperty, |
|
addEmptyRow, |
|
deleteRow, |
|
deleteRowById, |
|
deleteSelectedRows, |
|
deleteRangeOfRows, |
|
updateOrSaveRow, |
|
bulkUpdateRows, |
|
bulkUpdateView, |
|
selectedAllRecords, |
|
removeRowIfNew, |
|
bulkDeleteRows, |
|
bulkInsertRows, |
|
} |
|
}
|
|
|