import type { ColumnType, LinkToAnotherRecordType, PaginatedType, RelationTypes, TableType, ViewType } from 'nocodb-sdk' import { UITypes, isCreatedOrLastModifiedByCol, isCreatedOrLastModifiedTimeCol } from 'nocodb-sdk' import type { ComputedRef, Ref } from 'vue' import { NOCO, computed, extractPk, extractPkFromRow, extractSdkResponseErrorMsg, findIndexByPk, message, populateInsertObject, rowDefaultData, rowPkData, storeToRefs, until, useBase, useI18n, useMetas, useNuxtApp, } from '#imports' import type { CellRange, Row, UndoRedoAction } from '#imports' export function useData(args: { meta: Ref | ComputedRef viewMeta: Ref | ComputedRef<(ViewType & { id: string }) | undefined> formattedData: Ref paginationData: Ref callbacks?: { changePage?: (page: number) => Promise loadData?: () => Promise globalCallback?: (...args: any[]) => void syncCount?: () => Promise syncPagination?: () => Promise } }) { const { meta, viewMeta, formattedData, paginationData, callbacks } = args const { t } = useI18n() const { getMeta, metas } = useMetas() const { addUndo, clone, defineViewScope } = useUndoRedo() const { base } = storeToRefs(useBase()) const { $api } = useNuxtApp() const { isPaginationLoading } = storeToRefs(useViewsStore()) const selectedAllRecords = computed({ get() { return !!formattedData.value.length && formattedData.value.every((row: Row) => row.rowMeta.selected) }, set(selected: boolean) { formattedData.value.forEach((row: Row) => (row.rowMeta.selected = selected)) }, }) function addEmptyRow(addAfter = formattedData.value.length, metaValue = meta.value) { formattedData.value.splice(addAfter, 0, { row: { ...rowDefaultData(metaValue?.columns) }, oldRow: {}, rowMeta: { new: true }, }) return formattedData.value[addAfter] } async function insertRow( currentRow: Row, ltarState: Record = {}, { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, undo = false, ) { const row = currentRow.row if (currentRow.rowMeta) currentRow.rowMeta.saving = true try { const { missingRequiredColumns, insertObj } = await populateInsertObject({ meta: metaValue!, ltarState, getMeta, row, undo, }) if (missingRequiredColumns.size) return const insertedData = await $api.dbViewRow.create( NOCO, base?.value.id as string, metaValue?.id as string, viewMetaValue?.id as string, { ...insertObj, ...(ltarState || {}) }, ) if (!undo) { Object.assign(currentRow, { row: { ...insertedData, ...row }, rowMeta: { ...(currentRow.rowMeta || {}), new: undefined }, oldRow: { ...insertedData }, }) const id = extractPkFromRow(insertedData, metaValue?.columns as ColumnType[]) const pkData = rowPkData(insertedData, metaValue?.columns as ColumnType[]) const rowIndex = findIndexByPk(pkData, formattedData.value) addUndo({ redo: { fn: async function redo( this: UndoRedoAction, row: Row, ltarState: Record, pg: { page: number; pageSize: number }, ) { row.row = { ...pkData, ...row.row } const insertedData = await insertRow(row, ltarState, undefined, true) if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { if (pg.page === paginationData.value.page) { formattedData.value.splice(rowIndex, 0, { row: { ...row, ...insertedData }, rowMeta: row.rowMeta, oldRow: row.oldRow, }) } else { await callbacks?.changePage?.(pg.page) } } else { await callbacks?.loadData?.() } }, args: [ clone(currentRow), clone(ltarState), { page: paginationData.value.page, pageSize: paginationData.value.pageSize }, ], }, undo: { fn: async function undo(this: UndoRedoAction, id: string) { await deleteRowById(id) if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) paginationData.value.totalRows = paginationData.value.totalRows! - 1 }, args: [id], }, scope: defineViewScope({ view: viewMeta.value }), }) } await callbacks?.syncCount?.() return insertedData } catch (error: any) { message.error(await extractSdkResponseErrorMsg(error)) } finally { if (currentRow.rowMeta) currentRow.rowMeta.saving = false await callbacks?.globalCallback?.() } } async function bulkInsertRows( rows: Row[], { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, undo = false, ) { isPaginationLoading.value = true const autoGeneratedKeys = clone(metaValue?.columns || []) .filter((c) => !c.pk && (isCreatedOrLastModifiedTimeCol(c) || isCreatedOrLastModifiedByCol(c))) .map((c) => c.title) try { const rowsToInsert = ( await Promise.all( rows.map(async (currentRow) => { const { missingRequiredColumns, insertObj } = await populateInsertObject({ meta: metaValue!, ltarState: {}, getMeta, row: currentRow.row, undo, }) if (missingRequiredColumns.size === 0) { autoGeneratedKeys.forEach((key) => delete insertObj[key!]) return insertObj } }), ) )?.filter(Boolean) ?? [] // Filter out undefined values (if any) const bulkInsertedIds = await $api.dbDataTableRow.create(metaValue?.id as string, rowsToInsert, { viewId: viewMetaValue?.id as string, }) await callbacks?.syncCount?.() return bulkInsertedIds } catch (error: any) { message.error(await extractSdkResponseErrorMsg(error)) } finally { await callbacks?.globalCallback?.() isPaginationLoading.value = false } } // inside this method use metaValue and viewMetaValue to refer meta // since sometimes we need to pass old metas async function updateRowProperty( toUpdate: Row, property: string, { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, undo = false, ) { if (toUpdate.rowMeta) toUpdate.rowMeta.saving = true try { const id = extractPkFromRow(toUpdate.row, metaValue?.columns as ColumnType[]) const updatedRowData: Record = await $api.dbViewRow.update( NOCO, base?.value.id as string, metaValue?.id as string, viewMetaValue?.id as string, id, { // if value is undefined treat it as null [property]: toUpdate.row[property] ?? null, }, // todo: // { // query: { ignoreWebhook: !saved } // } ) if (!undo) { addUndo({ redo: { fn: async function redo(toUpdate: Row, property: string, pg: { page: number; pageSize: number }) { const updatedData = await updateRowProperty(toUpdate, property, undefined, true) if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) if (rowIndex !== -1) { const row = formattedData.value[rowIndex] Object.assign(row.row, updatedData) Object.assign(row.oldRow, updatedData) } else { await callbacks?.loadData?.() } } else { await callbacks?.changePage?.(pg.page) } }, args: [clone(toUpdate), property, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], }, undo: { fn: async function undo(toUpdate: Row, property: string, pg: { page: number; pageSize: number }) { const updatedData = await updateRowProperty( { row: toUpdate.oldRow, oldRow: toUpdate.row, rowMeta: toUpdate.rowMeta }, property, undefined, true, ) if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) if (rowIndex !== -1) { const row = formattedData.value[rowIndex] Object.assign(row.row, updatedData) Object.assign(row.oldRow, updatedData) } else { await callbacks?.loadData?.() } } else { await callbacks?.changePage?.(pg.page) } }, args: [clone(toUpdate), property, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], }, scope: defineViewScope({ view: viewMeta.value }), }) /** update row data(to sync formula and other related columns) * update only formula, rollup and auto updated datetime columns data to avoid overwriting any changes made by user */ Object.assign( toUpdate.row, metaValue!.columns!.reduce>((acc: Record, col: ColumnType) => { if ( col.title in updatedRowData && (col.uidt === UITypes.Formula || col.uidt === UITypes.QrCode || col.uidt === UITypes.Barcode || col.uidt === UITypes.Rollup || col.uidt === UITypes.Checkbox || col.uidt === UITypes.User || col.uidt === UITypes.LastModifiedTime || col.uidt === UITypes.LastModifiedBy || col.uidt === UITypes.Lookup || col.au || (col.cdf && / on update /i.test(col.cdf))) ) acc[col.title!] = updatedRowData[col.title!] return acc }, {} as Record), ) Object.assign(toUpdate.oldRow, updatedRowData) } await callbacks?.globalCallback?.() return updatedRowData } catch (e: any) { message.error(`${t('msg.error.rowUpdateFailed')} ${await extractSdkResponseErrorMsg(e)}`) } finally { if (toUpdate.rowMeta) toUpdate.rowMeta.saving = false } } async function updateOrSaveRow( row: Row, property?: string, ltarState?: Record, args: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, ) { // update changed status if (row.rowMeta) row.rowMeta.changed = false // if new row and save is in progress then wait until the save is complete await until(() => !(row.rowMeta?.new && row.rowMeta?.saving)).toMatch((v) => v) if (row.rowMeta.new) { return await insertRow(row, ltarState, args) } else { // if the field name is missing skip update if (property) { await updateRowProperty(row, property, args) } } } async function bulkUpdateRows( rows: Row[], props: string[], { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, undo = false, ) { const promises = [] for (const row of rows) { // update changed status if (row.rowMeta) row.rowMeta.changed = false // if new row and save is in progress then wait until the save is complete promises.push(until(() => !(row.rowMeta?.new && row.rowMeta?.saving)).toMatch((v) => v)) } await Promise.all(promises) const updateArray = [] for (const row of rows) { if (row.rowMeta) row.rowMeta.saving = true const pk = rowPkData(row.row, metaValue?.columns as ColumnType[]) const updateData = props.reduce((acc: Record, prop) => { acc[prop] = row.row[prop] return acc }, {} as Record) updateArray.push({ ...updateData, ...pk }) } await $api.dbTableRow.bulkUpdate(NOCO, metaValue?.base_id as string, metaValue?.id as string, updateArray) if (!undo) { addUndo({ redo: { fn: async function redo(redoRows: Row[], props: string[], pg: { page: number; pageSize: number }) { await bulkUpdateRows(redoRows, props, { metaValue, viewMetaValue }, true) if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { for (const toUpdate of redoRows) { const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) if (rowIndex !== -1) { const row = formattedData.value[rowIndex] Object.assign(row.row, toUpdate.row) Object.assign(row.oldRow, toUpdate.row) } else { await callbacks?.loadData?.() break } } } else { await callbacks?.changePage?.(pg.page) } }, args: [clone(rows), clone(props), { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], }, undo: { fn: async function undo(undoRows: Row[], props: string[], pg: { page: number; pageSize: number }) { await bulkUpdateRows(undoRows, props, { metaValue, viewMetaValue }, true) if (pg.page === paginationData.value.page && pg.pageSize === paginationData.value.pageSize) { for (const toUpdate of undoRows) { const rowIndex = findIndexByPk(rowPkData(toUpdate.row, meta?.value?.columns as ColumnType[]), formattedData.value) if (rowIndex !== -1) { const row = formattedData.value[rowIndex] Object.assign(row.row, toUpdate.row) Object.assign(row.oldRow, toUpdate.row) } else { await callbacks?.loadData?.() break } } } else { await callbacks?.changePage?.(pg.page) } }, args: [ clone( rows.map((row) => { return { row: row.oldRow, oldRow: row.row, rowMeta: row.rowMeta } }), ), props, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }, ], }, scope: defineViewScope({ view: viewMetaValue }), }) } for (const row of rows) { if (row.rowMeta) row.rowMeta.saving = false } // reload data since row update may change the other columns data( Formula, QrCode, Barcode, Rollup, Checkbox, User, Auto Updated Datetime, etc...) if ( metaValue!.columns!.some((col) => [ UITypes.QrCode, UITypes.LastModifiedTime, UITypes.Barcode, UITypes.Formula, UITypes.Lookup, UITypes.Rollup, UITypes.LinkToAnotherRecord, UITypes.LastModifiedBy, ].includes(col.uidt), ) ) { await callbacks?.loadData?.() } await callbacks?.globalCallback?.() } async function bulkUpdateView( data: Record[], { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, ) { if (!viewMetaValue) return await $api.dbTableRow.bulkUpdateAll(NOCO, metaValue?.base_id as string, metaValue?.id as string, data, { viewId: viewMetaValue.id, }) await callbacks?.loadData?.() await callbacks?.globalCallback?.() } const linkRecord = async ( rowId: string, relatedRowId: string, column: ColumnType, type: RelationTypes, { metaValue = meta.value }: { metaValue?: TableType } = {}, ) => { try { await $api.dbTableRow.nestedAdd( NOCO, base.value.id as string, metaValue?.id as string, encodeURIComponent(rowId), type as RelationTypes, column.title as string, encodeURIComponent(relatedRowId), ) } catch (e: any) { message.error(await extractSdkResponseErrorMsg(e)) } } // Recover LTAR relations for a row using the row data const recoverLTARRefs = async (row: Record, { metaValue = meta.value }: { metaValue?: TableType } = {}) => { const id = extractPkFromRow(row, metaValue?.columns as ColumnType[]) for (const column of metaValue?.columns ?? []) { if (column.uidt !== UITypes.LinkToAnotherRecord) continue const colOptions = column.colOptions as LinkToAnotherRecordType const relatedTableMeta = metas.value?.[colOptions?.fk_related_model_id as string] if (isHm(column) || isMm(column)) { const relatedRows = (row[column.title!] ?? []) as Record[] for (const relatedRow of relatedRows) { await linkRecord( id, extractPkFromRow(relatedRow, relatedTableMeta.columns as ColumnType[]), column, colOptions.type as RelationTypes, { metaValue }, ) } } else if (isBt(column) && row[column.title!]) { await linkRecord( id, extractPkFromRow(row[column.title!] as Record, relatedTableMeta.columns as ColumnType[]), column, colOptions.type as RelationTypes, { metaValue }, ) } } } async function deleteRowById( id: string, { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, ) { if (!id) { throw new Error("Delete not allowed for table which doesn't have primary Key") } const res: any = await $api.dbViewRow.delete( 'noco', base.value.id as string, metaValue?.id as string, viewMetaValue?.id as string, encodeURIComponent(id), ) if (res.message) { message.info( `Record delete failed: ${`Unable to delete record with ID ${id} because of the following: \n${res.message.join('\n')}.\n Clear the data first & try again`})}`, ) return false } return true } async function deleteRow(rowIndex: number, undo?: boolean) { try { const row = formattedData.value[rowIndex] if (!row.rowMeta.new) { const id = meta?.value?.columns ?.filter((c) => c.pk) .map((c) => row.row[c.title!]) .join('___') const fullRecord = await $api.dbTableRow.read( NOCO, // todo: base_id missing on view type base?.value.id as string, meta.value?.id as string, encodeURIComponent(id as string), { getHiddenColumn: true, }, ) row.row = fullRecord const deleted = await deleteRowById(id as string) if (!deleted) { return } if (!undo) { addUndo({ redo: { fn: async function redo(this: UndoRedoAction, id: string) { await deleteRowById(id) const pk: Record = rowPkData(row.row, meta?.value?.columns as ColumnType[]) const rowIndex = findIndexByPk(pk, formattedData.value) if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) paginationData.value.totalRows = paginationData.value.totalRows! - 1 }, args: [id], }, undo: { fn: async function undo( this: UndoRedoAction, row: Row, ltarState: Record, pg: { page: number; pageSize: number }, ) { const pkData = rowPkData(row.row, meta.value?.columns as ColumnType[]) row.row = { ...pkData, ...row.row } await insertRow(row, ltarState, {}, true) recoverLTARRefs(row.row) if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { if (pg.page === paginationData.value.page) { formattedData.value.splice(rowIndex, 0, row) } else { await callbacks?.changePage?.(pg.page) } } else { await callbacks?.loadData?.() } }, args: [clone(row), {}, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], }, scope: defineViewScope({ view: viewMeta.value }), }) } } formattedData.value.splice(rowIndex, 1) await callbacks?.syncCount?.() } catch (e: any) { message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) } await callbacks?.globalCallback?.() } async function deleteSelectedRows() { let row = formattedData.value.length const removedRowsData: Record[] = [] let compositePrimaryKey = '' while (row--) { const { row: rowData, rowMeta } = formattedData.value[row] as Record if (!rowMeta.selected) { continue } if (!rowMeta.new) { const extractedPk = extractPk(meta?.value?.columns as ColumnType[]) const compositePkValue = extractPkFromRow(rowData, meta?.value?.columns as ColumnType[]) const pkData = rowPkData(rowData, meta?.value?.columns as ColumnType[]) if (extractedPk && compositePkValue) { if (!compositePrimaryKey) compositePrimaryKey = extractedPk removedRowsData.push({ [compositePrimaryKey]: compositePkValue as string, pkData, row: clone(formattedData.value[row]) as Row, rowIndex: row as number, }) } } } if (!removedRowsData.length) return const { list } = await $api.dbTableRow.list(NOCO, base?.value.id as string, meta.value?.id as string, { pks: removedRowsData.map((row) => row[compositePrimaryKey]).join(','), }) try { for (const removedRow of removedRowsData) { const rowObj = removedRow.row const rowPk = rowPkData(rowObj.row, meta?.value?.columns as ColumnType[]) const fullRecord = list.find((r: Record) => { return Object.keys(rowPk).every((key) => rowPk[key] === r[key]) }) rowObj.row = clone(fullRecord) } await bulkDeleteRows(removedRowsData.map((row) => row.pkData)) } catch (e: any) { return message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) } if (!removedRowsData.length) return addUndo({ redo: { fn: async function redo(this: UndoRedoAction, removedRowsData: Record[]) { const removedRowIds = await bulkDeleteRows(removedRowsData.map((row) => row.pkData)) if (Array.isArray(removedRowIds)) { for (const { row } of removedRowsData) { const primaryKey: Record = rowPkData(row.row, meta?.value?.columns as ColumnType[]) const rowIndex = findIndexByPk(primaryKey, formattedData.value) if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) paginationData.value.totalRows = paginationData.value.totalRows! - 1 } } await callbacks?.syncPagination?.() }, args: [removedRowsData], }, undo: { fn: async function undo( this: UndoRedoAction, removedRowsData: Record[], pg: { page: number; pageSize: number }, ) { const rowsToInsert = removedRowsData .map((row) => { const pkData = rowPkData(row.row, meta.value?.columns as ColumnType[]) row.row = { ...pkData, ...row.row } return row }) .reverse() const insertedRowIds = await bulkInsertRows( rowsToInsert.map((row) => row.row), undefined, true, ) if (Array.isArray(insertedRowIds)) { for (const { row, rowIndex } of rowsToInsert) { recoverLTARRefs(row.row) if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { if (pg.page === paginationData.value.page) { formattedData.value.splice(rowIndex, 0, row) } else { await callbacks?.changePage?.(pg.page) } } else { await callbacks?.loadData?.() } } } }, args: [removedRowsData, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], }, scope: defineViewScope({ view: viewMeta.value }), }) await callbacks?.syncCount?.() await callbacks?.syncPagination?.() await callbacks?.globalCallback?.() } async function deleteRangeOfRows(cellRange: CellRange) { if (!cellRange._start || !cellRange._end) return const start = Math.max(cellRange._start.row, cellRange._end.row) const end = Math.min(cellRange._start.row, cellRange._end.row) // plus one because we want to include the end row let row = start + 1 const removedRowsData: Record[] = [] let compositePrimaryKey = '' while (row--) { try { const { row: rowData, rowMeta } = formattedData.value[row] as Record if (!rowMeta.new) { const extractedPk = extractPk(meta?.value?.columns as ColumnType[]) const compositePkValue = extractPkFromRow(rowData, meta?.value?.columns as ColumnType[]) const pkData = rowPkData(rowData, meta?.value?.columns as ColumnType[]) if (extractedPk && compositePkValue) { if (!compositePrimaryKey) compositePrimaryKey = extractedPk removedRowsData.push({ [compositePrimaryKey]: compositePkValue as string, pkData, row: clone(formattedData.value[row]) as Row, rowIndex: row as number, }) } } } catch (e: any) { return message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) } if (row === end) break } if (!removedRowsData.length) return const { list } = await $api.dbTableRow.list(NOCO, base?.value.id as string, meta.value?.id as string, { pks: removedRowsData.map((row) => row[compositePrimaryKey]).join(','), }) try { for (const removedRow of removedRowsData) { const rowObj = removedRow.row const rowPk = rowPkData(rowObj.row, meta?.value?.columns as ColumnType[]) const fullRecord = list.find((r: Record) => { return Object.keys(rowPk).every((key) => rowPk[key] === r[key]) }) rowObj.row = clone(fullRecord) } await bulkDeleteRows(removedRowsData.map((row) => row.pkData)) } catch (e: any) { return message.error(`${t('msg.error.deleteRowFailed')}: ${await extractSdkResponseErrorMsg(e)}`) } if (!removedRowsData.length) return addUndo({ redo: { fn: async function redo(this: UndoRedoAction, removedRowsData: Record[]) { const removedRowIds = await bulkDeleteRows(removedRowsData.map((row) => row.pkData)) if (Array.isArray(removedRowIds)) { for (const { row } of removedRowsData) { const primaryKey: Record = rowPkData(row.row, meta?.value?.columns as ColumnType[]) const rowIndex = findIndexByPk(primaryKey, formattedData.value) if (rowIndex !== -1) formattedData.value.splice(rowIndex, 1) paginationData.value.totalRows = paginationData.value.totalRows! - 1 } } await callbacks?.syncPagination?.() }, args: [removedRowsData], }, undo: { fn: async function undo( this: UndoRedoAction, removedRowsData: Record[], pg: { page: number; pageSize: number }, ) { const rowsToInsert = removedRowsData .map((row) => { const pkData = rowPkData(row.row, meta.value?.columns as ColumnType[]) row.row = { ...pkData, ...row.row } return row }) .reverse() const insertedRowIds = await bulkInsertRows( rowsToInsert.map((row) => row.row), undefined, true, ) if (Array.isArray(insertedRowIds)) { for (const { row, rowIndex } of rowsToInsert) { recoverLTARRefs(row.row) if (rowIndex !== -1 && pg.pageSize === paginationData.value.pageSize) { if (pg.page === paginationData.value.page) { formattedData.value.splice(rowIndex, 0, row) } else { await callbacks?.changePage?.(pg.page) } } else { await callbacks?.loadData?.() } } } }, args: [removedRowsData, { page: paginationData.value.page, pageSize: paginationData.value.pageSize }], }, scope: defineViewScope({ view: viewMeta.value }), }) await callbacks?.syncCount?.() await callbacks?.syncPagination?.() await callbacks?.globalCallback?.() } async function bulkDeleteRows( rows: Record[], { metaValue = meta.value, viewMetaValue = viewMeta.value }: { metaValue?: TableType; viewMetaValue?: ViewType } = {}, ) { isPaginationLoading.value = true try { const bulkDeletedRowsData = await $api.dbDataTableRow.delete(metaValue?.id as string, rows.length === 1 ? rows[0] : rows, { viewId: viewMetaValue?.id as string, }) await callbacks?.syncCount?.() return rows.length === 1 && bulkDeletedRowsData ? [bulkDeletedRowsData] : bulkDeletedRowsData } catch (error: any) { message.error(await extractSdkResponseErrorMsg(error)) } finally { await callbacks?.globalCallback?.() isPaginationLoading.value = false } } const removeRowIfNew = (row: Row) => { const index = formattedData.value.indexOf(row) if (index > -1 && row.rowMeta.new) { formattedData.value.splice(index, 1) return true } return false } return { insertRow, updateRowProperty, addEmptyRow, deleteRow, deleteRowById, deleteSelectedRows, deleteRangeOfRows, updateOrSaveRow, bulkUpdateRows, bulkUpdateView, selectedAllRecords, removeRowIfNew, bulkDeleteRows, bulkInsertRows, } }