mirror of
https://github.com/chartdb/chartdb.git
synced 2025-10-29 02:53:56 +00:00
Compare commits
19 Commits
v1.7.0
...
jf/wrong_i
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
595e3db0b3 | ||
|
|
ab89bad6d5 | ||
|
|
deb218423f | ||
|
|
48342471ac | ||
|
|
47bb87a88f | ||
|
|
a96c2e1078 | ||
|
|
26d95eed25 | ||
|
|
be65328f24 | ||
|
|
85fd14fa02 | ||
|
|
9c485b3b01 | ||
|
|
e993f1549c | ||
|
|
0db67ea42a | ||
|
|
b9e621bd68 | ||
|
|
93d59f8887 | ||
|
|
190e4f4ffa | ||
|
|
dc404c9d7e | ||
|
|
dd4324d64f | ||
|
|
1878083056 | ||
|
|
7b6271962a |
24
CHANGELOG.md
24
CHANGELOG.md
@@ -1,5 +1,29 @@
|
||||
# Changelog
|
||||
|
||||
## [1.8.0](https://github.com/chartdb/chartdb/compare/v1.7.0...v1.8.0) (2025-02-13)
|
||||
|
||||
|
||||
### Features
|
||||
|
||||
* **dbml-import:** add error highlighting for dbml imports ([#556](https://github.com/chartdb/chartdb/issues/556)) ([190e4f4](https://github.com/chartdb/chartdb/commit/190e4f4ffa834fa621f264dc608ca3f3b393a331))
|
||||
* **docker image:** add support for custom inference servers ([#543](https://github.com/chartdb/chartdb/issues/543)) ([1878083](https://github.com/chartdb/chartdb/commit/1878083056ea4db7a05cdeeb38a4f7b9f5f95bd1))
|
||||
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
* **canvas:** add right-click option to create relationships ([#568](https://github.com/chartdb/chartdb/issues/568)) ([e993f15](https://github.com/chartdb/chartdb/commit/e993f1549c4c86bb9e7e36062db803ba6613b3b3))
|
||||
* **canvas:** locate table from canvas ([#560](https://github.com/chartdb/chartdb/issues/560)) ([dc404c9](https://github.com/chartdb/chartdb/commit/dc404c9d7ee272c93aac69646bac859829a5234e))
|
||||
* **docker:** add option to hide popups ([#580](https://github.com/chartdb/chartdb/issues/580)) ([a96c2e1](https://github.com/chartdb/chartdb/commit/a96c2e107838d2dc13b586923fd9dbe06598cdd8))
|
||||
* **export-sql:** show create script for only filtered schemas ([#570](https://github.com/chartdb/chartdb/issues/570)) ([85fd14f](https://github.com/chartdb/chartdb/commit/85fd14fa02bb2879c36bba53369dbf2e7fa578d4))
|
||||
* **i18n:** fix Ukrainian ([#554](https://github.com/chartdb/chartdb/issues/554)) ([7b62719](https://github.com/chartdb/chartdb/commit/7b6271962a99bfe5ffbd0176e714c76368ef5c41))
|
||||
* **import dbml:** add import for indexes ([#566](https://github.com/chartdb/chartdb/issues/566)) ([0db67ea](https://github.com/chartdb/chartdb/commit/0db67ea42a5f9585ca1d246db7a7ff0239bec0ba))
|
||||
* **import-query:** improve the cleanup for messy json input ([#562](https://github.com/chartdb/chartdb/issues/562)) ([93d59f8](https://github.com/chartdb/chartdb/commit/93d59f8887765098d040a3184aaee32112f67267))
|
||||
* **index unique:** extract unique toggle for faster editing ([#559](https://github.com/chartdb/chartdb/issues/559)) ([dd4324d](https://github.com/chartdb/chartdb/commit/dd4324d64f7638ada5c022a2ab38bd8e6986af25))
|
||||
* **mssql-import:** improve script readability by adding edition comment ([#572](https://github.com/chartdb/chartdb/issues/572)) ([be65328](https://github.com/chartdb/chartdb/commit/be65328f24b0361638b9e2edb39eaa9906e77f67))
|
||||
* **realtionships section:** add the schema to source/target tables ([#561](https://github.com/chartdb/chartdb/issues/561)) ([b9e621b](https://github.com/chartdb/chartdb/commit/b9e621bd680730a0ffbf1054d735bfa418711cae))
|
||||
* **sqlserver-import:** open ssms guide when max chars ([#565](https://github.com/chartdb/chartdb/issues/565)) ([9c485b3](https://github.com/chartdb/chartdb/commit/9c485b3b01a131bf551c7e95916b0c416f6aa0b5))
|
||||
* **table actions:** fix size of table actions ([#578](https://github.com/chartdb/chartdb/issues/578)) ([26d95ee](https://github.com/chartdb/chartdb/commit/26d95eed25d86452d9168a9d93a301ba50d934e3))
|
||||
|
||||
## [1.7.0](https://github.com/chartdb/chartdb/compare/v1.6.1...v1.7.0) (2025-02-03)
|
||||
|
||||
|
||||
|
||||
12
Dockerfile
12
Dockerfile
@@ -1,6 +1,9 @@
|
||||
FROM node:22-alpine AS builder
|
||||
|
||||
ARG VITE_OPENAI_API_KEY
|
||||
ARG VITE_OPENAI_API_ENDPOINT
|
||||
ARG VITE_LLM_MODEL_NAME
|
||||
ARG VITE_HIDE_BUCKLE_DOT_DEV
|
||||
|
||||
WORKDIR /usr/src/app
|
||||
|
||||
@@ -10,9 +13,13 @@ RUN npm ci
|
||||
|
||||
COPY . .
|
||||
|
||||
RUN echo "VITE_OPENAI_API_KEY=${VITE_OPENAI_API_KEY}" > .env && \
|
||||
echo "VITE_OPENAI_API_ENDPOINT=${VITE_OPENAI_API_ENDPOINT}" >> .env && \
|
||||
echo "VITE_LLM_MODEL_NAME=${VITE_LLM_MODEL_NAME}" >> .env && \
|
||||
echo "VITE_HIDE_BUCKLE_DOT_DEV=${VITE_HIDE_BUCKLE_DOT_DEV}" >> .env
|
||||
|
||||
RUN npm run build
|
||||
|
||||
# Use a lightweight web server to serve the production build
|
||||
FROM nginx:stable-alpine AS production
|
||||
|
||||
COPY --from=builder /usr/src/app/dist /usr/share/nginx/html
|
||||
@@ -20,7 +27,6 @@ COPY ./default.conf.template /etc/nginx/conf.d/default.conf.template
|
||||
COPY entrypoint.sh /entrypoint.sh
|
||||
RUN chmod +x /entrypoint.sh
|
||||
|
||||
# Expose the default port for the Nginx web server
|
||||
EXPOSE 80
|
||||
|
||||
ENTRYPOINT ["/entrypoint.sh"]
|
||||
ENTRYPOINT ["/entrypoint.sh"]
|
||||
25
README.md
25
README.md
@@ -107,8 +107,33 @@ docker build -t chartdb .
|
||||
docker run -e OPENAI_API_KEY=<YOUR_OPEN_AI_KEY> -p 8080:80 chartdb
|
||||
```
|
||||
|
||||
#### Using Custom Inference Server
|
||||
|
||||
```bash
|
||||
# Build
|
||||
docker build \
|
||||
--build-arg VITE_OPENAI_API_ENDPOINT=<YOUR_ENDPOINT> \
|
||||
--build-arg VITE_LLM_MODEL_NAME=<YOUR_MODEL_NAME> \
|
||||
-t chartdb .
|
||||
|
||||
# Run
|
||||
docker run \
|
||||
-e OPENAI_API_ENDPOINT=<YOUR_ENDPOINT> \
|
||||
-e LLM_MODEL_NAME=<YOUR_MODEL_NAME> \
|
||||
-p 8080:80 chartdb
|
||||
```
|
||||
|
||||
> **Note:** You must configure either Option 1 (OpenAI API key) OR Option 2 (Custom endpoint and model name) for AI capabilities to work. Do not mix the two options.
|
||||
|
||||
Open your browser and navigate to `http://localhost:8080`.
|
||||
|
||||
Example configuration for a local vLLM server:
|
||||
|
||||
```bash
|
||||
VITE_OPENAI_API_ENDPOINT=http://localhost:8000/v1
|
||||
VITE_LLM_MODEL_NAME=Qwen/Qwen2.5-32B-Instruct-AWQ
|
||||
```
|
||||
|
||||
## Try it on our website
|
||||
|
||||
1. Go to [ChartDB.io](https://chartdb.io?ref=github_readme_2)
|
||||
|
||||
@@ -10,7 +10,12 @@ server {
|
||||
|
||||
location /config.js {
|
||||
default_type application/javascript;
|
||||
return 200 "window.env = { OPENAI_API_KEY: \"$OPENAI_API_KEY\" };";
|
||||
return 200 "window.env = {
|
||||
OPENAI_API_KEY: \"$OPENAI_API_KEY\",
|
||||
OPENAI_API_ENDPOINT: \"$OPENAI_API_ENDPOINT\",
|
||||
LLM_MODEL_NAME: \"$LLM_MODEL_NAME\",
|
||||
HIDE_BUCKLE_DOT_DEV: \"$HIDE_BUCKLE_DOT_DEV\"
|
||||
};";
|
||||
}
|
||||
|
||||
error_page 500 502 503 504 /50x.html;
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
#!/bin/sh
|
||||
|
||||
# Replace placeholders in nginx.conf
|
||||
envsubst '${OPENAI_API_KEY}' < /etc/nginx/conf.d/default.conf.template > /etc/nginx/conf.d/default.conf
|
||||
envsubst '${OPENAI_API_KEY} ${OPENAI_API_ENDPOINT} ${LLM_MODEL_NAME} ${HIDE_BUCKLE_DOT_DEV}' < /etc/nginx/conf.d/default.conf.template > /etc/nginx/conf.d/default.conf
|
||||
|
||||
# Start Nginx
|
||||
nginx -g "daemon off;"
|
||||
|
||||
4
package-lock.json
generated
4
package-lock.json
generated
@@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "chartdb",
|
||||
"version": "1.7.0",
|
||||
"version": "1.8.0",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "chartdb",
|
||||
"version": "1.7.0",
|
||||
"version": "1.8.0",
|
||||
"dependencies": {
|
||||
"@ai-sdk/openai": "^0.0.51",
|
||||
"@dbml/core": "^3.9.5",
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "chartdb",
|
||||
"private": true,
|
||||
"version": "1.7.0",
|
||||
"version": "1.8.0",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "vite",
|
||||
|
||||
@@ -6,6 +6,8 @@ import type { ExportSQLDialogProps } from '@/dialogs/export-sql-dialog/export-sq
|
||||
import type { ExportImageDialogProps } from '@/dialogs/export-image-dialog/export-image-dialog';
|
||||
import type { ExportDiagramDialogProps } from '@/dialogs/export-diagram-dialog/export-diagram-dialog';
|
||||
import type { ImportDiagramDialogProps } from '@/dialogs/import-diagram-dialog/import-diagram-dialog';
|
||||
import type { CreateRelationshipDialogProps } from '@/dialogs/create-relationship-dialog/create-relationship-dialog';
|
||||
import type { ImportDBMLDialogProps } from '@/dialogs/import-dbml-dialog/import-dbml-dialog';
|
||||
|
||||
export interface DialogContext {
|
||||
// Create diagram dialog
|
||||
@@ -21,7 +23,9 @@ export interface DialogContext {
|
||||
closeExportSQLDialog: () => void;
|
||||
|
||||
// Create relationship dialog
|
||||
openCreateRelationshipDialog: () => void;
|
||||
openCreateRelationshipDialog: (
|
||||
params?: Omit<CreateRelationshipDialogProps, 'dialog'>
|
||||
) => void;
|
||||
closeCreateRelationshipDialog: () => void;
|
||||
|
||||
// Import database dialog
|
||||
@@ -63,7 +67,9 @@ export interface DialogContext {
|
||||
closeImportDiagramDialog: () => void;
|
||||
|
||||
// Import DBML dialog
|
||||
openImportDBMLDialog: () => void;
|
||||
openImportDBMLDialog: (
|
||||
params?: Omit<ImportDBMLDialogProps, 'dialog'>
|
||||
) => void;
|
||||
closeImportDBMLDialog: () => void;
|
||||
}
|
||||
|
||||
|
||||
@@ -6,6 +6,7 @@ import { OpenDiagramDialog } from '@/dialogs/open-diagram-dialog/open-diagram-di
|
||||
import type { ExportSQLDialogProps } from '@/dialogs/export-sql-dialog/export-sql-dialog';
|
||||
import { ExportSQLDialog } from '@/dialogs/export-sql-dialog/export-sql-dialog';
|
||||
import { DatabaseType } from '@/lib/domain/database-type';
|
||||
import type { CreateRelationshipDialogProps } from '@/dialogs/create-relationship-dialog/create-relationship-dialog';
|
||||
import { CreateRelationshipDialog } from '@/dialogs/create-relationship-dialog/create-relationship-dialog';
|
||||
import type { ImportDatabaseDialogProps } from '@/dialogs/import-database-dialog/import-database-dialog';
|
||||
import { ImportDatabaseDialog } from '@/dialogs/import-database-dialog/import-database-dialog';
|
||||
@@ -18,6 +19,7 @@ import { ExportImageDialog } from '@/dialogs/export-image-dialog/export-image-di
|
||||
import { ExportDiagramDialog } from '@/dialogs/export-diagram-dialog/export-diagram-dialog';
|
||||
import { ImportDiagramDialog } from '@/dialogs/import-diagram-dialog/import-diagram-dialog';
|
||||
import { BuckleDialog } from '@/dialogs/buckle-dialog/buckle-dialog';
|
||||
import type { ImportDBMLDialogProps } from '@/dialogs/import-dbml-dialog/import-dbml-dialog';
|
||||
import { ImportDBMLDialog } from '@/dialogs/import-dbml-dialog/import-dbml-dialog';
|
||||
|
||||
export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
@@ -28,6 +30,17 @@ export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
|
||||
const [openCreateRelationshipDialog, setOpenCreateRelationshipDialog] =
|
||||
useState(false);
|
||||
const [createRelationshipDialogParams, setCreateRelationshipDialogParams] =
|
||||
useState<Omit<CreateRelationshipDialogProps, 'dialog'>>();
|
||||
const openCreateRelationshipDialogHandler: DialogContext['openCreateRelationshipDialog'] =
|
||||
useCallback(
|
||||
(params) => {
|
||||
setCreateRelationshipDialogParams(params);
|
||||
setOpenCreateRelationshipDialog(true);
|
||||
},
|
||||
[setOpenCreateRelationshipDialog]
|
||||
);
|
||||
|
||||
const [openStarUsDialog, setOpenStarUsDialog] = useState(false);
|
||||
const [openBuckleDialog, setOpenBuckleDialog] = useState(false);
|
||||
|
||||
@@ -99,6 +112,8 @@ export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
|
||||
// Import DBML dialog
|
||||
const [openImportDBMLDialog, setOpenImportDBMLDialog] = useState(false);
|
||||
const [importDBMLDialogParams, setImportDBMLDialogParams] =
|
||||
useState<Omit<ImportDBMLDialogProps, 'dialog'>>();
|
||||
|
||||
return (
|
||||
<dialogContext.Provider
|
||||
@@ -109,8 +124,8 @@ export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
closeOpenDiagramDialog: () => setOpenOpenDiagramDialog(false),
|
||||
openExportSQLDialog: openExportSQLDialogHandler,
|
||||
closeExportSQLDialog: () => setOpenExportSQLDialog(false),
|
||||
openCreateRelationshipDialog: () =>
|
||||
setOpenCreateRelationshipDialog(true),
|
||||
openCreateRelationshipDialog:
|
||||
openCreateRelationshipDialogHandler,
|
||||
closeCreateRelationshipDialog: () =>
|
||||
setOpenCreateRelationshipDialog(false),
|
||||
openImportDatabaseDialog: openImportDatabaseDialogHandler,
|
||||
@@ -130,7 +145,10 @@ export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
openImportDiagramDialog: () => setOpenImportDiagramDialog(true),
|
||||
closeImportDiagramDialog: () =>
|
||||
setOpenImportDiagramDialog(false),
|
||||
openImportDBMLDialog: () => setOpenImportDBMLDialog(true),
|
||||
openImportDBMLDialog: (params) => {
|
||||
setImportDBMLDialogParams(params);
|
||||
setOpenImportDBMLDialog(true);
|
||||
},
|
||||
closeImportDBMLDialog: () => setOpenImportDBMLDialog(false),
|
||||
}}
|
||||
>
|
||||
@@ -143,6 +161,7 @@ export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
/>
|
||||
<CreateRelationshipDialog
|
||||
dialog={{ open: openCreateRelationshipDialog }}
|
||||
{...createRelationshipDialogParams}
|
||||
/>
|
||||
<ImportDatabaseDialog
|
||||
dialog={{ open: openImportDatabaseDialog }}
|
||||
@@ -160,7 +179,10 @@ export const DialogProvider: React.FC<React.PropsWithChildren> = ({
|
||||
<ExportDiagramDialog dialog={{ open: openExportDiagramDialog }} />
|
||||
<ImportDiagramDialog dialog={{ open: openImportDiagramDialog }} />
|
||||
<BuckleDialog dialog={{ open: openBuckleDialog }} />
|
||||
<ImportDBMLDialog dialog={{ open: openImportDBMLDialog }} />
|
||||
<ImportDBMLDialog
|
||||
dialog={{ open: openImportDBMLDialog }}
|
||||
{...importDBMLDialogParams}
|
||||
/>
|
||||
</dialogContext.Provider>
|
||||
);
|
||||
};
|
||||
|
||||
@@ -85,6 +85,10 @@ export const ImportDatabase: React.FC<ImportDatabaseProps> = ({
|
||||
const [showCheckJsonButton, setShowCheckJsonButton] = useState(false);
|
||||
const [isCheckingJson, setIsCheckingJson] = useState(false);
|
||||
|
||||
const [showSSMSInfoDialog, setShowSSMSInfoDialog] = useState(false);
|
||||
|
||||
const helpButtonRef = React.useRef<HTMLButtonElement>(null);
|
||||
|
||||
useEffect(() => {
|
||||
const loadScripts = async () => {
|
||||
const { importMetadataScripts } = await import(
|
||||
@@ -127,6 +131,16 @@ export const ImportDatabase: React.FC<ImportDatabaseProps> = ({
|
||||
(e: React.ChangeEvent<HTMLTextAreaElement>) => {
|
||||
const inputValue = e.target.value;
|
||||
setScriptResult(inputValue);
|
||||
|
||||
// Automatically open SSMS info when input length is exactly 65535
|
||||
if (inputValue.length === 65535) {
|
||||
setShowSSMSInfoDialog(true);
|
||||
}
|
||||
|
||||
// Show instructions when input contains "WITH fk_info as"
|
||||
if (inputValue.toLowerCase().includes('with fk_info as')) {
|
||||
helpButtonRef.current?.click();
|
||||
}
|
||||
},
|
||||
[setScriptResult]
|
||||
);
|
||||
@@ -245,7 +259,10 @@ export const ImportDatabase: React.FC<ImportDatabaseProps> = ({
|
||||
{t('new_diagram_dialog.import_database.step_1')}
|
||||
</div>
|
||||
{databaseType === DatabaseType.SQL_SERVER && (
|
||||
<SSMSInfo />
|
||||
<SSMSInfo
|
||||
open={showSSMSInfoDialog}
|
||||
setOpen={setShowSSMSInfoDialog}
|
||||
/>
|
||||
)}
|
||||
</div>
|
||||
{databaseTypeToClientsMap[databaseType].length > 0 ? (
|
||||
@@ -369,6 +386,8 @@ export const ImportDatabase: React.FC<ImportDatabaseProps> = ({
|
||||
showCheckJsonButton,
|
||||
isCheckingJson,
|
||||
handleCheckJson,
|
||||
showSSMSInfoDialog,
|
||||
setShowSSMSInfoDialog,
|
||||
]);
|
||||
|
||||
const renderFooter = useCallback(() => {
|
||||
@@ -386,7 +405,11 @@ export const ImportDatabase: React.FC<ImportDatabaseProps> = ({
|
||||
)}
|
||||
{isDesktop ? (
|
||||
<ZoomableImage src="/load-new-db-instructions.gif">
|
||||
<Button type="button" variant="link">
|
||||
<Button
|
||||
type="button"
|
||||
variant="link"
|
||||
ref={helpButtonRef}
|
||||
>
|
||||
{t(
|
||||
'new_diagram_dialog.import_database.instructions_link'
|
||||
)}
|
||||
@@ -438,7 +461,11 @@ export const ImportDatabase: React.FC<ImportDatabaseProps> = ({
|
||||
|
||||
{!isDesktop ? (
|
||||
<ZoomableImage src="/load-new-db-instructions.gif">
|
||||
<Button type="button" variant="link">
|
||||
<Button
|
||||
type="button"
|
||||
variant="link"
|
||||
ref={helpButtonRef}
|
||||
>
|
||||
{t(
|
||||
'new_diagram_dialog.import_database.instructions_link'
|
||||
)}
|
||||
|
||||
@@ -4,32 +4,55 @@ import {
|
||||
HoverCardTrigger,
|
||||
} from '@/components/hover-card/hover-card';
|
||||
import { Label } from '@/components/label/label';
|
||||
import { Info } from 'lucide-react';
|
||||
import React from 'react';
|
||||
import { Info, X } from 'lucide-react';
|
||||
import React, { useCallback, useEffect, useMemo } from 'react';
|
||||
import SSMSInstructions from '@/assets/ssms-instructions.png';
|
||||
import { ZoomableImage } from '@/components/zoomable-image/zoomable-image';
|
||||
import { useTranslation } from 'react-i18next';
|
||||
|
||||
export interface SSMSInfoProps {}
|
||||
export interface SSMSInfoProps {
|
||||
open?: boolean;
|
||||
setOpen?: (open: boolean) => void;
|
||||
}
|
||||
|
||||
export const SSMSInfo = React.forwardRef<
|
||||
React.ElementRef<typeof HoverCardTrigger>,
|
||||
SSMSInfoProps
|
||||
>((props, ref) => {
|
||||
>(({ open: controlledOpen, setOpen: setControlledOpen }, ref) => {
|
||||
const [open, setOpen] = React.useState(false);
|
||||
const { t } = useTranslation();
|
||||
|
||||
useEffect(() => {
|
||||
if (controlledOpen) {
|
||||
setOpen(true);
|
||||
}
|
||||
}, [controlledOpen]);
|
||||
|
||||
const closeHandler = useCallback(() => {
|
||||
setOpen(false);
|
||||
setControlledOpen?.(false);
|
||||
}, [setControlledOpen]);
|
||||
|
||||
const isOpen = useMemo(
|
||||
() => open || controlledOpen,
|
||||
[open, controlledOpen]
|
||||
);
|
||||
|
||||
return (
|
||||
<HoverCard
|
||||
open={open}
|
||||
open={isOpen}
|
||||
onOpenChange={(isOpen) => {
|
||||
if (controlledOpen) {
|
||||
return;
|
||||
}
|
||||
setOpen(isOpen);
|
||||
}}
|
||||
>
|
||||
<HoverCardTrigger ref={ref} {...props} asChild>
|
||||
<HoverCardTrigger ref={ref} asChild>
|
||||
<div
|
||||
className="flex flex-row items-center gap-1 text-pink-600"
|
||||
onClick={() => {
|
||||
setOpen(!open);
|
||||
setOpen?.(!open);
|
||||
}}
|
||||
>
|
||||
<Info size={14} />
|
||||
@@ -41,13 +64,21 @@ export const SSMSInfo = React.forwardRef<
|
||||
</div>
|
||||
</HoverCardTrigger>
|
||||
<HoverCardContent className="w-80">
|
||||
<div className="flex">
|
||||
<div className="space-y-1">
|
||||
<div className="flex flex-col">
|
||||
<div className="flex items-start justify-between">
|
||||
<h4 className="text-sm font-semibold">
|
||||
{t(
|
||||
'new_diagram_dialog.import_database.ssms_instructions.title'
|
||||
)}
|
||||
</h4>
|
||||
<button
|
||||
onClick={closeHandler}
|
||||
className="text-muted-foreground hover:text-foreground"
|
||||
>
|
||||
<X size={16} />
|
||||
</button>
|
||||
</div>
|
||||
<div className="space-y-1">
|
||||
<p className="text-xs text-muted-foreground">
|
||||
<span className="font-semibold">1. </span>
|
||||
{t(
|
||||
|
||||
@@ -28,7 +28,7 @@ export const CreateDiagramDialog: React.FC<CreateDiagramDialogProps> = ({
|
||||
const [databaseType, setDatabaseType] = useState<DatabaseType>(
|
||||
DatabaseType.GENERIC
|
||||
);
|
||||
const { closeCreateDiagramDialog } = useDialog();
|
||||
const { closeCreateDiagramDialog, openImportDBMLDialog } = useDialog();
|
||||
const { updateConfig } = useConfig();
|
||||
const [scriptResult, setScriptResult] = useState('');
|
||||
const [databaseEdition, setDatabaseEdition] = useState<
|
||||
@@ -104,6 +104,10 @@ export const CreateDiagramDialog: React.FC<CreateDiagramDialogProps> = ({
|
||||
await updateConfig({ defaultDiagramId: diagram.id });
|
||||
closeCreateDiagramDialog();
|
||||
navigate(`/diagrams/${diagram.id}`);
|
||||
setTimeout(
|
||||
() => openImportDBMLDialog({ withCreateEmptyDiagram: true }),
|
||||
700
|
||||
);
|
||||
}, [
|
||||
databaseType,
|
||||
addDiagram,
|
||||
@@ -112,6 +116,7 @@ export const CreateDiagramDialog: React.FC<CreateDiagramDialogProps> = ({
|
||||
navigate,
|
||||
updateConfig,
|
||||
diagramNumber,
|
||||
openImportDBMLDialog,
|
||||
]);
|
||||
|
||||
return (
|
||||
|
||||
@@ -22,13 +22,17 @@ import { areFieldTypesCompatible } from '@/lib/data/data-types/data-types';
|
||||
const ErrorMessageRelationshipFieldsNotSameType =
|
||||
'Relationships can only be created between fields of the same type';
|
||||
|
||||
export interface CreateRelationshipDialogProps extends BaseDialogProps {}
|
||||
export interface CreateRelationshipDialogProps extends BaseDialogProps {
|
||||
sourceTableId?: string;
|
||||
}
|
||||
|
||||
export const CreateRelationshipDialog: React.FC<
|
||||
CreateRelationshipDialogProps
|
||||
> = ({ dialog }) => {
|
||||
> = ({ dialog, sourceTableId: preSelectedSourceTableId }) => {
|
||||
const { closeCreateRelationshipDialog } = useDialog();
|
||||
const [primaryTableId, setPrimaryTableId] = useState<string | undefined>();
|
||||
const [primaryTableId, setPrimaryTableId] = useState<string | undefined>(
|
||||
preSelectedSourceTableId
|
||||
);
|
||||
const [primaryFieldId, setPrimaryFieldId] = useState<string | undefined>();
|
||||
const [referencedTableId, setReferencedTableId] = useState<
|
||||
string | undefined
|
||||
@@ -43,6 +47,9 @@ export const CreateRelationshipDialog: React.FC<
|
||||
const [canCreateRelationship, setCanCreateRelationship] = useState(false);
|
||||
const { fitView, setEdges } = useReactFlow();
|
||||
const { databaseType } = useChartDB();
|
||||
const [primaryFieldSelectOpen, setPrimaryFieldSelectOpen] = useState(false);
|
||||
const [referencedTableSelectOpen, setReferencedTableSelectOpen] =
|
||||
useState(false);
|
||||
|
||||
const tableOptions = useMemo(() => {
|
||||
return tables.map(
|
||||
@@ -89,8 +96,23 @@ export const CreateRelationshipDialog: React.FC<
|
||||
setReferencedTableId(undefined);
|
||||
setReferencedFieldId(undefined);
|
||||
setErrorMessage('');
|
||||
setPrimaryFieldSelectOpen(false);
|
||||
setReferencedTableSelectOpen(false);
|
||||
}, [dialog.open]);
|
||||
|
||||
useEffect(() => {
|
||||
if (preSelectedSourceTableId) {
|
||||
const table = getTable(preSelectedSourceTableId);
|
||||
if (table) {
|
||||
setPrimaryTableId(preSelectedSourceTableId);
|
||||
}
|
||||
|
||||
setTimeout(() => {
|
||||
setPrimaryFieldSelectOpen(true);
|
||||
}, 100);
|
||||
}
|
||||
}, [preSelectedSourceTableId, getTable]);
|
||||
|
||||
useEffect(() => {
|
||||
setCanCreateRelationship(false);
|
||||
setErrorMessage('');
|
||||
@@ -223,8 +245,14 @@ export const CreateRelationshipDialog: React.FC<
|
||||
)}
|
||||
value={primaryTableId}
|
||||
onChange={(value) => {
|
||||
setPrimaryTableId(value as string);
|
||||
setPrimaryFieldId(undefined);
|
||||
const newTableId = value as string;
|
||||
setPrimaryTableId(newTableId);
|
||||
if (
|
||||
newTableId !==
|
||||
preSelectedSourceTableId
|
||||
) {
|
||||
setPrimaryFieldId(undefined);
|
||||
}
|
||||
}}
|
||||
emptyPlaceholder={t(
|
||||
'create_relationship_dialog.no_tables_found'
|
||||
@@ -253,6 +281,8 @@ export const CreateRelationshipDialog: React.FC<
|
||||
'create_relationship_dialog.primary_field_placeholder'
|
||||
)}
|
||||
value={primaryFieldId}
|
||||
open={primaryFieldSelectOpen}
|
||||
onOpenChange={setPrimaryFieldSelectOpen}
|
||||
onChange={(value) =>
|
||||
setPrimaryFieldId(value as string)
|
||||
}
|
||||
@@ -283,6 +313,8 @@ export const CreateRelationshipDialog: React.FC<
|
||||
'create_relationship_dialog.referenced_table_placeholder'
|
||||
)}
|
||||
value={referencedTableId}
|
||||
open={referencedTableSelectOpen}
|
||||
onOpenChange={setReferencedTableSelectOpen}
|
||||
onChange={(value) => {
|
||||
setReferencedTableId(value as string);
|
||||
setReferencedFieldId(undefined);
|
||||
|
||||
@@ -20,10 +20,12 @@ import {
|
||||
} from '@/lib/data/export-metadata/export-sql-script';
|
||||
import { databaseTypeToLabelMap } from '@/lib/databases';
|
||||
import { DatabaseType } from '@/lib/domain/database-type';
|
||||
import { shouldShowTablesBySchemaFilter } from '@/lib/domain/db-table';
|
||||
import { Annoyed, Sparkles } from 'lucide-react';
|
||||
import React, { useCallback, useEffect, useRef } from 'react';
|
||||
import { Trans, useTranslation } from 'react-i18next';
|
||||
import type { BaseDialogProps } from '../common/base-dialog-props';
|
||||
import type { Diagram } from '@/lib/domain/diagram';
|
||||
|
||||
export interface ExportSQLDialogProps extends BaseDialogProps {
|
||||
targetDatabaseType: DatabaseType;
|
||||
@@ -34,7 +36,7 @@ export const ExportSQLDialog: React.FC<ExportSQLDialogProps> = ({
|
||||
targetDatabaseType,
|
||||
}) => {
|
||||
const { closeExportSQLDialog } = useDialog();
|
||||
const { currentDiagram } = useChartDB();
|
||||
const { currentDiagram, filteredSchemas } = useChartDB();
|
||||
const { t } = useTranslation();
|
||||
const [script, setScript] = React.useState<string>();
|
||||
const [error, setError] = React.useState<boolean>(false);
|
||||
@@ -43,17 +45,58 @@ export const ExportSQLDialog: React.FC<ExportSQLDialogProps> = ({
|
||||
const abortControllerRef = useRef<AbortController | null>(null);
|
||||
|
||||
const exportSQLScript = useCallback(async () => {
|
||||
const filteredDiagram: Diagram = {
|
||||
...currentDiagram,
|
||||
tables: currentDiagram.tables?.filter((table) =>
|
||||
shouldShowTablesBySchemaFilter(table, filteredSchemas)
|
||||
),
|
||||
relationships: currentDiagram.relationships?.filter((rel) => {
|
||||
const sourceTable = currentDiagram.tables?.find(
|
||||
(t) => t.id === rel.sourceTableId
|
||||
);
|
||||
const targetTable = currentDiagram.tables?.find(
|
||||
(t) => t.id === rel.targetTableId
|
||||
);
|
||||
return (
|
||||
sourceTable &&
|
||||
targetTable &&
|
||||
shouldShowTablesBySchemaFilter(
|
||||
sourceTable,
|
||||
filteredSchemas
|
||||
) &&
|
||||
shouldShowTablesBySchemaFilter(targetTable, filteredSchemas)
|
||||
);
|
||||
}),
|
||||
dependencies: currentDiagram.dependencies?.filter((dep) => {
|
||||
const table = currentDiagram.tables?.find(
|
||||
(t) => t.id === dep.tableId
|
||||
);
|
||||
const dependentTable = currentDiagram.tables?.find(
|
||||
(t) => t.id === dep.dependentTableId
|
||||
);
|
||||
return (
|
||||
table &&
|
||||
dependentTable &&
|
||||
shouldShowTablesBySchemaFilter(table, filteredSchemas) &&
|
||||
shouldShowTablesBySchemaFilter(
|
||||
dependentTable,
|
||||
filteredSchemas
|
||||
)
|
||||
);
|
||||
}),
|
||||
};
|
||||
|
||||
if (targetDatabaseType === DatabaseType.GENERIC) {
|
||||
return Promise.resolve(exportBaseSQL(currentDiagram));
|
||||
return Promise.resolve(exportBaseSQL(filteredDiagram));
|
||||
} else {
|
||||
return exportSQL(currentDiagram, targetDatabaseType, {
|
||||
return exportSQL(filteredDiagram, targetDatabaseType, {
|
||||
stream: true,
|
||||
onResultStream: (text) =>
|
||||
setScript((prev) => (prev ? prev + text : text)),
|
||||
signal: abortControllerRef.current?.signal,
|
||||
});
|
||||
}
|
||||
}, [targetDatabaseType, currentDiagram]);
|
||||
}, [targetDatabaseType, currentDiagram, filteredSchemas]);
|
||||
|
||||
useEffect(() => {
|
||||
if (!dialog.open) {
|
||||
|
||||
@@ -1,4 +1,11 @@
|
||||
import React, { useCallback, useEffect, useState, Suspense } from 'react';
|
||||
import React, {
|
||||
useCallback,
|
||||
useEffect,
|
||||
useState,
|
||||
Suspense,
|
||||
useRef,
|
||||
} from 'react';
|
||||
import * as monaco from 'monaco-editor';
|
||||
import { useDialog } from '@/hooks/use-dialog';
|
||||
import {
|
||||
Dialog,
|
||||
@@ -23,11 +30,54 @@ import { useCanvas } from '@/hooks/use-canvas';
|
||||
import { setupDBMLLanguage } from '@/components/code-snippet/languages/dbml-language';
|
||||
import { useToast } from '@/components/toast/use-toast';
|
||||
import { Spinner } from '@/components/spinner/spinner';
|
||||
import { debounce } from '@/lib/utils';
|
||||
|
||||
export interface ImportDBMLDialogProps extends BaseDialogProps {}
|
||||
interface DBMLError {
|
||||
message: string;
|
||||
line: number;
|
||||
column: number;
|
||||
}
|
||||
|
||||
function parseDBMLError(error: unknown): DBMLError | null {
|
||||
try {
|
||||
if (typeof error === 'string') {
|
||||
const parsed = JSON.parse(error);
|
||||
if (parsed.diags?.[0]) {
|
||||
const diag = parsed.diags[0];
|
||||
return {
|
||||
message: diag.message,
|
||||
line: diag.location.start.line,
|
||||
column: diag.location.start.column,
|
||||
};
|
||||
}
|
||||
} else if (error && typeof error === 'object' && 'diags' in error) {
|
||||
const parsed = error as {
|
||||
diags: Array<{
|
||||
message: string;
|
||||
location: { start: { line: number; column: number } };
|
||||
}>;
|
||||
};
|
||||
if (parsed.diags?.[0]) {
|
||||
return {
|
||||
message: parsed.diags[0].message,
|
||||
line: parsed.diags[0].location.start.line,
|
||||
column: parsed.diags[0].location.start.column,
|
||||
};
|
||||
}
|
||||
}
|
||||
} catch (e) {
|
||||
console.error('Error parsing DBML error:', e);
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
export interface ImportDBMLDialogProps extends BaseDialogProps {
|
||||
withCreateEmptyDiagram?: boolean;
|
||||
}
|
||||
|
||||
export const ImportDBMLDialog: React.FC<ImportDBMLDialogProps> = ({
|
||||
dialog,
|
||||
withCreateEmptyDiagram,
|
||||
}) => {
|
||||
const { t } = useTranslation();
|
||||
const initialDBML = `// Use DBML to define your database structure
|
||||
@@ -75,6 +125,16 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
} = useChartDB();
|
||||
const { reorderTables } = useCanvas();
|
||||
const [reorder, setReorder] = useState(false);
|
||||
const editorRef = useRef<monaco.editor.IStandaloneCodeEditor>();
|
||||
const decorationsCollection =
|
||||
useRef<monaco.editor.IEditorDecorationsCollection>();
|
||||
|
||||
const handleEditorDidMount = (
|
||||
editor: monaco.editor.IStandaloneCodeEditor
|
||||
) => {
|
||||
editorRef.current = editor;
|
||||
decorationsCollection.current = editor.createDecorationsCollection();
|
||||
};
|
||||
|
||||
useEffect(() => {
|
||||
if (reorder) {
|
||||
@@ -85,34 +145,97 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
}
|
||||
}, [reorder, reorderTables]);
|
||||
|
||||
useEffect(() => {
|
||||
if (!dialog.open) return;
|
||||
setErrorMessage(undefined);
|
||||
setDBMLContent(initialDBML);
|
||||
}, [dialog.open, initialDBML]);
|
||||
const highlightErrorLine = useCallback((error: DBMLError) => {
|
||||
if (!editorRef.current) return;
|
||||
|
||||
useEffect(() => {
|
||||
const validateDBML = async () => {
|
||||
if (!dbmlContent.trim()) {
|
||||
setErrorMessage(undefined);
|
||||
return;
|
||||
}
|
||||
const model = editorRef.current.getModel();
|
||||
if (!model) return;
|
||||
|
||||
const decorations = [
|
||||
{
|
||||
range: new monaco.Range(
|
||||
error.line,
|
||||
1,
|
||||
error.line,
|
||||
model.getLineMaxColumn(error.line)
|
||||
),
|
||||
options: {
|
||||
isWholeLine: true,
|
||||
className: 'dbml-error-line',
|
||||
glyphMarginClassName: 'dbml-error-glyph',
|
||||
hoverMessage: { value: error.message },
|
||||
overviewRuler: {
|
||||
color: '#ff0000',
|
||||
position: monaco.editor.OverviewRulerLane.Right,
|
||||
darkColor: '#ff0000',
|
||||
},
|
||||
},
|
||||
},
|
||||
];
|
||||
|
||||
decorationsCollection.current?.set(decorations);
|
||||
}, []);
|
||||
|
||||
const clearDecorations = useCallback(() => {
|
||||
decorationsCollection.current?.clear();
|
||||
}, []);
|
||||
|
||||
const validateDBML = useCallback(
|
||||
async (content: string) => {
|
||||
// Clear previous errors
|
||||
setErrorMessage(undefined);
|
||||
clearDecorations();
|
||||
|
||||
if (!content.trim()) return;
|
||||
|
||||
try {
|
||||
const parser = new Parser();
|
||||
parser.parse(dbmlContent, 'dbml');
|
||||
setErrorMessage(undefined);
|
||||
parser.parse(content, 'dbml');
|
||||
} catch (e) {
|
||||
setErrorMessage(
|
||||
e instanceof Error
|
||||
? e.message
|
||||
: t('import_dbml_dialog.error.description')
|
||||
);
|
||||
const parsedError = parseDBMLError(e);
|
||||
if (parsedError) {
|
||||
setErrorMessage(
|
||||
t('import_dbml_dialog.error.description') +
|
||||
` (1 error found - in line ${parsedError.line})`
|
||||
);
|
||||
highlightErrorLine(parsedError);
|
||||
} else {
|
||||
setErrorMessage(
|
||||
e instanceof Error ? e.message : JSON.stringify(e)
|
||||
);
|
||||
}
|
||||
}
|
||||
};
|
||||
},
|
||||
[clearDecorations, highlightErrorLine, t]
|
||||
);
|
||||
|
||||
validateDBML();
|
||||
}, [dbmlContent, t]);
|
||||
const debouncedValidateRef = useRef<((value: string) => void) | null>(null);
|
||||
|
||||
// Set up debounced validation
|
||||
useEffect(() => {
|
||||
debouncedValidateRef.current = debounce((value: string) => {
|
||||
validateDBML(value);
|
||||
}, 500);
|
||||
|
||||
return () => {
|
||||
debouncedValidateRef.current = null;
|
||||
};
|
||||
}, [validateDBML]);
|
||||
|
||||
// Trigger validation when content changes
|
||||
useEffect(() => {
|
||||
if (debouncedValidateRef.current) {
|
||||
debouncedValidateRef.current(dbmlContent);
|
||||
}
|
||||
}, [dbmlContent]);
|
||||
|
||||
useEffect(() => {
|
||||
if (!dialog.open) {
|
||||
setErrorMessage(undefined);
|
||||
clearDecorations();
|
||||
setDBMLContent(initialDBML);
|
||||
}
|
||||
}, [dialog.open, initialDBML, clearDecorations]);
|
||||
|
||||
const handleImport = useCallback(async () => {
|
||||
if (!dbmlContent.trim() || errorMessage) return;
|
||||
@@ -177,7 +300,7 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
description: (
|
||||
<>
|
||||
<div>{t('import_dbml_dialog.error.description')}</div>
|
||||
{e instanceof Error ? <div>{e.message}</div> : null}
|
||||
{e instanceof Error ? e.message : JSON.stringify(e)}
|
||||
</>
|
||||
),
|
||||
});
|
||||
@@ -211,7 +334,11 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
showClose
|
||||
>
|
||||
<DialogHeader>
|
||||
<DialogTitle>{t('import_dbml_dialog.title')}</DialogTitle>
|
||||
<DialogTitle>
|
||||
{withCreateEmptyDiagram
|
||||
? t('import_dbml_dialog.example_title')
|
||||
: t('import_dbml_dialog.title')}
|
||||
</DialogTitle>
|
||||
<DialogDescription>
|
||||
{t('import_dbml_dialog.description')}
|
||||
</DialogDescription>
|
||||
@@ -222,6 +349,7 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
value={dbmlContent}
|
||||
onChange={(value) => setDBMLContent(value || '')}
|
||||
language="dbml"
|
||||
onMount={handleEditorDidMount}
|
||||
theme={
|
||||
effectiveTheme === 'dark'
|
||||
? 'dbml-dark'
|
||||
@@ -232,6 +360,8 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
minimap: { enabled: false },
|
||||
scrollBeyondLastLine: false,
|
||||
automaticLayout: true,
|
||||
glyphMargin: true,
|
||||
lineNumbers: 'on',
|
||||
scrollbar: {
|
||||
vertical: 'visible',
|
||||
horizontal: 'visible',
|
||||
@@ -246,7 +376,9 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
<div className="flex items-center gap-4">
|
||||
<DialogClose asChild>
|
||||
<Button variant="secondary">
|
||||
{t('import_dbml_dialog.cancel')}
|
||||
{withCreateEmptyDiagram
|
||||
? t('import_dbml_dialog.skip_and_empty')
|
||||
: t('import_dbml_dialog.cancel')}
|
||||
</Button>
|
||||
</DialogClose>
|
||||
{errorMessage ? (
|
||||
@@ -266,7 +398,9 @@ Ref: comments.user_id > users.id // Each comment is written by one user`;
|
||||
onClick={handleImport}
|
||||
disabled={!dbmlContent.trim() || !!errorMessage}
|
||||
>
|
||||
{t('import_dbml_dialog.import')}
|
||||
{withCreateEmptyDiagram
|
||||
? t('import_dbml_dialog.show_example')
|
||||
: t('import_dbml_dialog.import')}
|
||||
</Button>
|
||||
</div>
|
||||
</DialogFooter>
|
||||
|
||||
@@ -109,6 +109,10 @@
|
||||
animation: rainbow-text-simple-animation 0.5s ease-in forwards;
|
||||
}
|
||||
|
||||
.dbml-error-line {
|
||||
background-color: rgba(255, 0, 0, 0.2) !important;
|
||||
}
|
||||
|
||||
@keyframes rainbow-text-simple-animation-rev {
|
||||
0% {
|
||||
background-size: 650%;
|
||||
|
||||
@@ -378,12 +378,15 @@ export const ar: LanguageTranslation = {
|
||||
import_dbml_dialog: {
|
||||
// TODO: Translate
|
||||
title: 'Import DBML',
|
||||
example_title: 'Import Example DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -402,6 +405,7 @@ export const ar: LanguageTranslation = {
|
||||
edit_table: 'تعديل الجدول',
|
||||
duplicate_table: 'نسخ الجدول',
|
||||
delete_table: 'حذف الجدول',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: '({{key}} مغنظة الشبكة (اضغط مع الاستمرار على',
|
||||
|
||||
@@ -381,13 +381,16 @@ export const bn: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -406,6 +409,7 @@ export const bn: LanguageTranslation = {
|
||||
edit_table: 'টেবিল সম্পাদনা করুন',
|
||||
duplicate_table: 'টেবিল নকল করুন',
|
||||
delete_table: 'টেবিল মুছে ফেলুন',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'গ্রিডে স্ন্যাপ করুন (অবস্থান {{key}})',
|
||||
|
||||
@@ -384,13 +384,16 @@ export const de: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -409,6 +412,7 @@ export const de: LanguageTranslation = {
|
||||
edit_table: 'Tabelle bearbeiten',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'Tabelle löschen',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
|
||||
@@ -376,13 +376,16 @@ export const en = {
|
||||
},
|
||||
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error importing DBML',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -401,6 +404,7 @@ export const en = {
|
||||
edit_table: 'Edit Table',
|
||||
duplicate_table: 'Duplicate Table',
|
||||
delete_table: 'Delete Table',
|
||||
add_relationship: 'Add Relationship',
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'Snap to Grid (Hold {{key}})',
|
||||
|
||||
@@ -383,13 +383,16 @@ export const es: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -408,6 +411,7 @@ export const es: LanguageTranslation = {
|
||||
edit_table: 'Editar Tabla',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'Eliminar Tabla',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
|
||||
@@ -30,9 +30,8 @@ export const fr: LanguageTranslation = {
|
||||
theme: 'Thème',
|
||||
show_dependencies: 'Afficher les Dépendances',
|
||||
hide_dependencies: 'Masquer les Dépendances',
|
||||
// TODO: Translate
|
||||
show_minimap: 'Show Mini Map',
|
||||
hide_minimap: 'Hide Mini Map',
|
||||
show_minimap: 'Afficher la Mini Carte',
|
||||
hide_minimap: 'Masquer la Mini Carte',
|
||||
},
|
||||
share: {
|
||||
share: 'Partage',
|
||||
@@ -101,9 +100,8 @@ export const fr: LanguageTranslation = {
|
||||
clear: 'Effacer',
|
||||
show_more: 'Afficher Plus',
|
||||
show_less: 'Afficher Moins',
|
||||
// TODO: Translate
|
||||
copy_to_clipboard: 'Copy to Clipboard',
|
||||
copied: 'Copied!',
|
||||
copy_to_clipboard: 'Copier dans le presse-papiers',
|
||||
copied: 'Copié !',
|
||||
|
||||
side_panel: {
|
||||
schema: 'Schéma:',
|
||||
@@ -116,12 +114,11 @@ export const fr: LanguageTranslation = {
|
||||
add_table: 'Ajouter une Table',
|
||||
filter: 'Filtrer',
|
||||
collapse: 'Réduire Tout',
|
||||
// TODO: Translate
|
||||
clear: 'Clear Filter',
|
||||
no_results: 'No tables found matching your filter.',
|
||||
// TODO: Translate
|
||||
show_list: 'Show Table List',
|
||||
show_dbml: 'Show DBML Editor',
|
||||
clear: 'Effacer le Filtre',
|
||||
no_results:
|
||||
'Aucune table trouvée correspondant à votre filtre.',
|
||||
show_list: 'Afficher la Liste des Tableaux',
|
||||
show_dbml: "Afficher l'éditeur DBML",
|
||||
|
||||
table: {
|
||||
fields: 'Champs',
|
||||
@@ -153,7 +150,7 @@ export const fr: LanguageTranslation = {
|
||||
title: 'Actions de la Table',
|
||||
add_field: 'Ajouter un Champ',
|
||||
add_index: 'Ajouter un Index',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
duplicate_table: 'Tableau dupliqué',
|
||||
delete_table: 'Supprimer la Table',
|
||||
change_schema: 'Changer le Schéma',
|
||||
},
|
||||
@@ -236,14 +233,12 @@ export const fr: LanguageTranslation = {
|
||||
step_2: 'Si vous utilisez "Résultats en Grille", changez le nombre maximum de caractères récupérés pour les données non-XML (définir à 9999999).',
|
||||
},
|
||||
instructions_link: "Besoin d'aide ? Regardez comment",
|
||||
// TODO: Translate
|
||||
check_script_result: 'Check Script Result',
|
||||
check_script_result: 'Vérifier le résultat du Script',
|
||||
},
|
||||
|
||||
cancel: 'Annuler',
|
||||
back: 'Retour',
|
||||
// TODO: Translate
|
||||
import_from_file: 'Import from File',
|
||||
import_from_file: "Importer à partir d'un fichier",
|
||||
empty_diagram: 'Diagramme vide',
|
||||
continue: 'Continuer',
|
||||
import: 'Importer',
|
||||
@@ -358,40 +353,42 @@ export const fr: LanguageTranslation = {
|
||||
cancel: 'Annuler',
|
||||
},
|
||||
},
|
||||
// TODO: Translate
|
||||
export_diagram_dialog: {
|
||||
title: 'Export Diagram',
|
||||
description: 'Choose the format for export:',
|
||||
title: 'Exporter le Diagramme',
|
||||
description: "Sélectionner le format d'exportation :",
|
||||
format_json: 'JSON',
|
||||
cancel: 'Cancel',
|
||||
export: 'Export',
|
||||
cancel: 'Annuler',
|
||||
export: 'Exporter',
|
||||
error: {
|
||||
title: 'Error exporting diagram',
|
||||
title: "Erreur lors de l'exportation du diagramme",
|
||||
description:
|
||||
'Something went wrong. Need help? chartdb.io@gmail.com',
|
||||
"Une erreur s'est produite. Besoin d'aide ? chartdb.io@gmail.com",
|
||||
},
|
||||
},
|
||||
// TODO: Translate
|
||||
import_diagram_dialog: {
|
||||
title: 'Import Diagram',
|
||||
description: 'Paste the diagram JSON below:',
|
||||
cancel: 'Cancel',
|
||||
import: 'Import',
|
||||
title: 'Importer un diagramme',
|
||||
description: 'Coller le diagramme au format JSON ci-dessous :',
|
||||
cancel: 'Annuler',
|
||||
import: 'Exporter',
|
||||
error: {
|
||||
title: 'Error importing diagram',
|
||||
title: "Erreur lors de l'exportation du diagramme",
|
||||
description:
|
||||
'The diagram JSON is invalid. Please check the JSON and try again. Need help? chartdb.io@gmail.com',
|
||||
"Le diagramme JSON n'est pas valide. Veuillez vérifier le JSON et réessayer. Besoin d'aide ? chartdb.io@gmail.com",
|
||||
},
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: "Exemple d'importation DBML",
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
description:
|
||||
'Importer un schéma de base de données à partir du format DBML.',
|
||||
import: 'Importer',
|
||||
cancel: 'Annuler',
|
||||
skip_and_empty: 'Passer et vider',
|
||||
show_example: 'Afficher un exemple',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
title: 'Erreur',
|
||||
description:
|
||||
"Erreur d'analyse du DBML. Veuillez vérifier la syntaxe.",
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -408,12 +405,13 @@ export const fr: LanguageTranslation = {
|
||||
|
||||
table_node_context_menu: {
|
||||
edit_table: 'Éditer la Table',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
duplicate_table: 'Tableau Dupliqué',
|
||||
delete_table: 'Supprimer la Table',
|
||||
add_relationship: 'Ajouter une Relation',
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
snap_to_grid_tooltip: 'Snap to Grid (Hold {{key}})',
|
||||
snap_to_grid_tooltip:
|
||||
'Aligner sur la grille (maintenir la touche {{key}})',
|
||||
|
||||
tool_tips: {
|
||||
double_click_to_edit: 'Double-cliquez pour modifier',
|
||||
|
||||
@@ -381,13 +381,16 @@ export const gu: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -406,6 +409,7 @@ export const gu: LanguageTranslation = {
|
||||
edit_table: 'ટેબલ સંપાદિત કરો',
|
||||
duplicate_table: 'ટેબલ નકલ કરો',
|
||||
delete_table: 'ટેબલ કાઢી નાખો',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'ગ્રિડ પર સ્નેપ કરો (જમાવટ {{key}})',
|
||||
|
||||
@@ -385,13 +385,16 @@ export const hi: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -410,6 +413,7 @@ export const hi: LanguageTranslation = {
|
||||
edit_table: 'तालिका संपादित करें',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'तालिका हटाएँ',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
|
||||
@@ -379,13 +379,16 @@ export const id_ID: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
|
||||
@@ -405,6 +408,7 @@ export const id_ID: LanguageTranslation = {
|
||||
edit_table: 'Ubah Tabel',
|
||||
delete_table: 'Hapus Tabel',
|
||||
duplicate_table: 'Duplikat Tabel',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'Snap ke Kisi (Tahan {{key}})',
|
||||
|
||||
@@ -388,13 +388,16 @@ export const ja: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -413,6 +416,7 @@ export const ja: LanguageTranslation = {
|
||||
edit_table: 'テーブルを編集',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'テーブルを削除',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
|
||||
@@ -377,13 +377,16 @@ export const ko_KR: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -402,6 +405,7 @@ export const ko_KR: LanguageTranslation = {
|
||||
edit_table: '테이블 수정',
|
||||
duplicate_table: '테이블 복제',
|
||||
delete_table: '테이블 삭제',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: '그리드에 맞추기 ({{key}}를 누른채 유지)',
|
||||
|
||||
@@ -389,13 +389,16 @@ export const mr: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
|
||||
@@ -414,8 +417,8 @@ export const mr: LanguageTranslation = {
|
||||
table_node_context_menu: {
|
||||
edit_table: 'टेबल संपादित करा',
|
||||
delete_table: 'टेबल हटवा',
|
||||
// TODO: Add translations
|
||||
duplicate_table: 'Duplicate Table',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
|
||||
@@ -382,13 +382,16 @@ export const ne: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
|
||||
@@ -408,6 +411,7 @@ export const ne: LanguageTranslation = {
|
||||
edit_table: 'तालिका सम्पादन गर्नुहोस्',
|
||||
duplicate_table: 'तालिका नक्कली गर्नुहोस्',
|
||||
delete_table: 'तालिका हटाउनुहोस्',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'ग्रिडमा स्न्याप गर्नुहोस् ({{key}} थिच्नुहोस)',
|
||||
|
||||
@@ -382,13 +382,16 @@ export const pt_BR: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -407,6 +410,7 @@ export const pt_BR: LanguageTranslation = {
|
||||
edit_table: 'Editar Tabela',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'Excluir Tabela',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Add translations
|
||||
|
||||
@@ -378,13 +378,16 @@ export const ru: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -403,6 +406,7 @@ export const ru: LanguageTranslation = {
|
||||
edit_table: 'Изменить таблицу',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'Удалить таблицу',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
copy_to_clipboard: 'Скопировать в буфер обмена',
|
||||
|
||||
@@ -385,13 +385,16 @@ export const te: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
|
||||
@@ -409,9 +412,9 @@ export const te: LanguageTranslation = {
|
||||
|
||||
table_node_context_menu: {
|
||||
edit_table: 'పట్టికను సవరించు',
|
||||
// TODO: Translate
|
||||
duplicate_table: 'Duplicate Table',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: 'పట్టికను తొలగించు',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Translate
|
||||
|
||||
@@ -372,13 +372,16 @@ export const tr: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -394,8 +397,8 @@ export const tr: LanguageTranslation = {
|
||||
table_node_context_menu: {
|
||||
edit_table: 'Tabloyu Düzenle',
|
||||
delete_table: 'Tabloyu Sil',
|
||||
// TODO: Translate
|
||||
duplicate_table: 'Duplicate Table',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
// TODO: Translate
|
||||
|
||||
@@ -26,7 +26,7 @@ export const uk: LanguageTranslation = {
|
||||
hide_sidebar: 'Приховати бічну панель',
|
||||
hide_cardinality: 'Приховати потужність',
|
||||
show_cardinality: 'Показати кардинальність',
|
||||
zoom_on_scroll: 'Збільшити прокручування',
|
||||
zoom_on_scroll: 'Масштабувати прокручуванням',
|
||||
theme: 'Тема',
|
||||
show_dependencies: 'Показати залежності',
|
||||
hide_dependencies: 'Приховати залежності',
|
||||
@@ -377,13 +377,16 @@ export const uk: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -402,6 +405,7 @@ export const uk: LanguageTranslation = {
|
||||
edit_table: 'Редагувати таблицю',
|
||||
duplicate_table: 'Дублювати таблицю',
|
||||
delete_table: 'Видалити таблицю',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'Вирівнювати за сіткою (Отримуйте {{key}})',
|
||||
|
||||
@@ -378,13 +378,16 @@ export const vi: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -403,6 +406,7 @@ export const vi: LanguageTranslation = {
|
||||
edit_table: 'Sửa bảng',
|
||||
duplicate_table: 'Nhân đôi bảng',
|
||||
delete_table: 'Xóa bảng',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: 'Căn lưới (Giữ phím {{key}})',
|
||||
|
||||
@@ -374,13 +374,16 @@ export const zh_CN: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -399,6 +402,7 @@ export const zh_CN: LanguageTranslation = {
|
||||
edit_table: '编辑表',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: '删除表',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: '对齐到网格(按住 {{key}})',
|
||||
|
||||
@@ -373,13 +373,16 @@ export const zh_TW: LanguageTranslation = {
|
||||
},
|
||||
// TODO: Translate
|
||||
import_dbml_dialog: {
|
||||
example_title: 'Import Example DBML',
|
||||
title: 'Import DBML',
|
||||
description: 'Import a database schema from DBML format.',
|
||||
import: 'Import',
|
||||
cancel: 'Cancel',
|
||||
skip_and_empty: 'Skip & Empty',
|
||||
show_example: 'Show Example',
|
||||
error: {
|
||||
title: 'Error',
|
||||
description: 'Failed to import DBML. Please check the syntax.',
|
||||
description: 'Failed to parse DBML. Please check the syntax.',
|
||||
},
|
||||
},
|
||||
relationship_type: {
|
||||
@@ -398,6 +401,7 @@ export const zh_TW: LanguageTranslation = {
|
||||
edit_table: '編輯表格',
|
||||
duplicate_table: 'Duplicate Table', // TODO: Translate
|
||||
delete_table: '刪除表格',
|
||||
add_relationship: 'Add Relationship', // TODO: Translate
|
||||
},
|
||||
|
||||
snap_to_grid_tooltip: '對齊網格(按住 {{key}})',
|
||||
|
||||
@@ -12,6 +12,9 @@ export const sqliteDataTypes: readonly DataType[] = [
|
||||
// Blob Type
|
||||
{ name: 'blob', id: 'blob' },
|
||||
|
||||
// Blob Type
|
||||
{ name: 'json', id: 'json' },
|
||||
|
||||
// Date/Time Types (SQLite uses TEXT, REAL, or INTEGER types for dates and times)
|
||||
{ name: 'date', id: 'date' },
|
||||
{ name: 'datetime', id: 'datetime' },
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
import type { Diagram } from '../../domain/diagram';
|
||||
import { OPENAI_API_KEY } from '@/lib/env';
|
||||
import { OPENAI_API_KEY, OPENAI_API_ENDPOINT, LLM_MODEL_NAME } from '@/lib/env';
|
||||
import type { DatabaseType } from '@/lib/domain/database-type';
|
||||
import type { DBTable } from '@/lib/domain/db-table';
|
||||
import type { DataType } from '../data-types/data-types';
|
||||
@@ -196,6 +196,26 @@ export const exportBaseSQL = (diagram: Diagram): string => {
|
||||
return sqlScript;
|
||||
};
|
||||
|
||||
const validateConfiguration = () => {
|
||||
const apiKey = window?.env?.OPENAI_API_KEY ?? OPENAI_API_KEY;
|
||||
const baseUrl = window?.env?.OPENAI_API_ENDPOINT ?? OPENAI_API_ENDPOINT;
|
||||
const modelName = window?.env?.LLM_MODEL_NAME ?? LLM_MODEL_NAME;
|
||||
|
||||
// If using custom endpoint and model, don't require OpenAI API key
|
||||
if (baseUrl && modelName) {
|
||||
return { useCustomEndpoint: true };
|
||||
}
|
||||
|
||||
// If using OpenAI's service, require API key
|
||||
if (apiKey) {
|
||||
return { useCustomEndpoint: false };
|
||||
}
|
||||
|
||||
throw new Error(
|
||||
'Configuration Error: Either provide an OpenAI API key or both a custom endpoint and model name'
|
||||
);
|
||||
};
|
||||
|
||||
export const exportSQL = async (
|
||||
diagram: Diagram,
|
||||
databaseType: DatabaseType,
|
||||
@@ -213,43 +233,73 @@ export const exportSQL = async (
|
||||
return cachedResult;
|
||||
}
|
||||
|
||||
// Validate configuration before proceeding
|
||||
const { useCustomEndpoint } = validateConfiguration();
|
||||
|
||||
const [{ streamText, generateText }, { createOpenAI }] = await Promise.all([
|
||||
import('ai'),
|
||||
import('@ai-sdk/openai'),
|
||||
]);
|
||||
|
||||
const openai = createOpenAI({
|
||||
apiKey: window?.env?.OPENAI_API_KEY ?? OPENAI_API_KEY,
|
||||
});
|
||||
const apiKey = window?.env?.OPENAI_API_KEY ?? OPENAI_API_KEY;
|
||||
const baseUrl = window?.env?.OPENAI_API_ENDPOINT ?? OPENAI_API_ENDPOINT;
|
||||
const modelName = window?.env?.LLM_MODEL_NAME || 'gpt-4o-mini-2024-07-18';
|
||||
|
||||
let config: { apiKey: string; baseUrl?: string };
|
||||
|
||||
if (useCustomEndpoint) {
|
||||
config = {
|
||||
apiKey: 'sk-xxx', // minimal valid API key format
|
||||
baseUrl: baseUrl,
|
||||
};
|
||||
} else {
|
||||
config = {
|
||||
apiKey: apiKey,
|
||||
};
|
||||
}
|
||||
|
||||
const openai = createOpenAI(config);
|
||||
|
||||
const prompt = generateSQLPrompt(databaseType, sqlScript);
|
||||
|
||||
if (options?.stream) {
|
||||
const { textStream, text: textPromise } = await streamText({
|
||||
model: openai('gpt-4o-mini-2024-07-18'),
|
||||
try {
|
||||
if (options?.stream) {
|
||||
const { textStream, text: textPromise } = await streamText({
|
||||
model: openai(modelName),
|
||||
prompt: prompt,
|
||||
});
|
||||
|
||||
for await (const textPart of textStream) {
|
||||
if (options.signal?.aborted) {
|
||||
return '';
|
||||
}
|
||||
options.onResultStream(textPart);
|
||||
}
|
||||
|
||||
const text = await textPromise;
|
||||
|
||||
setInCache(cacheKey, text);
|
||||
return text;
|
||||
}
|
||||
|
||||
const { text } = await generateText({
|
||||
model: openai(modelName),
|
||||
prompt: prompt,
|
||||
});
|
||||
|
||||
for await (const textPart of textStream) {
|
||||
if (options.signal?.aborted) {
|
||||
return '';
|
||||
}
|
||||
options.onResultStream(textPart);
|
||||
}
|
||||
|
||||
const text = await textPromise;
|
||||
|
||||
setInCache(cacheKey, text);
|
||||
return text;
|
||||
} catch (error: unknown) {
|
||||
console.error('Error generating SQL:', error);
|
||||
if (error instanceof Error && error.message.includes('API key')) {
|
||||
throw new Error(
|
||||
'Error: Please check your API configuration. If using a custom endpoint, make sure the endpoint URL is correct.'
|
||||
);
|
||||
}
|
||||
throw new Error(
|
||||
'Error generating SQL script. Please check your configuration and try again.'
|
||||
);
|
||||
}
|
||||
|
||||
const { text } = await generateText({
|
||||
model: openai('gpt-4o-mini-2024-07-18'),
|
||||
prompt: prompt,
|
||||
});
|
||||
|
||||
setInCache(cacheKey, text);
|
||||
return text;
|
||||
};
|
||||
|
||||
function getMySQLDataTypeSize(type: DataType) {
|
||||
|
||||
@@ -85,7 +85,7 @@ export const sqliteQuery = `WITH fk_info AS (
|
||||
ELSE LOWER(p.type)
|
||||
END,
|
||||
'ordinal_position', p.cid,
|
||||
'nullable', (CASE WHEN p."notnull" = 0 THEN 'true' ELSE 'false' END),
|
||||
'nullable', (CASE WHEN p."notnull" = 0 THEN true ELSE false END),
|
||||
'collation', '',
|
||||
'character_maximum_length',
|
||||
CASE
|
||||
|
||||
@@ -1,23 +1,26 @@
|
||||
import { DatabaseEdition } from '@/lib/domain/database-edition';
|
||||
|
||||
const sqlServerQuery = `WITH fk_info AS (
|
||||
const sqlServerQuery = `${`/* SQL Server 2017 and above edition (14.0, 15.0, 16.0, 17.0)*/`}
|
||||
WITH fk_info AS (
|
||||
SELECT
|
||||
JSON_QUERY(
|
||||
'[' + STRING_AGG(
|
||||
N'[' + STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{"schema": "' + COALESCE(REPLACE(tp_schema.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(tp.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column": "' + COALESCE(REPLACE(cp.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "foreign_key_name": "' + COALESCE(REPLACE(fk.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "reference_schema": "' + COALESCE(REPLACE(tr_schema.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "reference_table": "' + COALESCE(REPLACE(tr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "reference_column": "' + COALESCE(REPLACE(cr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "fk_def": "FOREIGN KEY (' + COALESCE(REPLACE(cp.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
') REFERENCES ' + COALESCE(REPLACE(tr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'(' + COALESCE(REPLACE(cr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
') ON DELETE ' + fk.delete_referential_action_desc COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
' ON UPDATE ' + fk.update_referential_action_desc COLLATE SQL_Latin1_General_CP1_CI_AS + '"}')
|
||||
), ','
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(tp_schema.name, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(tp.name, '"', ''), ''), 'json') +
|
||||
'", "column": "' + STRING_ESCAPE(COALESCE(REPLACE(cp.name, '"', ''), ''), 'json') +
|
||||
'", "foreign_key_name": "' + STRING_ESCAPE(COALESCE(REPLACE(fk.name, '"', ''), ''), 'json') +
|
||||
'", "reference_schema": "' + STRING_ESCAPE(COALESCE(REPLACE(tr_schema.name, '"', ''), ''), 'json') +
|
||||
'", "reference_table": "' + STRING_ESCAPE(COALESCE(REPLACE(tr.name, '"', ''), ''), 'json') +
|
||||
'", "reference_column": "' + STRING_ESCAPE(COALESCE(REPLACE(cr.name, '"', ''), ''), 'json') +
|
||||
'", "fk_def": "FOREIGN KEY (' + STRING_ESCAPE(COALESCE(REPLACE(cp.name, '"', ''), ''), 'json') +
|
||||
') REFERENCES ' + STRING_ESCAPE(COALESCE(REPLACE(tr.name, '"', ''), ''), 'json') +
|
||||
'(' + STRING_ESCAPE(COALESCE(REPLACE(cr.name, '"', ''), ''), 'json') +
|
||||
') ON DELETE ' + STRING_ESCAPE(fk.delete_referential_action_desc, 'json') +
|
||||
' ON UPDATE ' + STRING_ESCAPE(fk.update_referential_action_desc, 'json') +
|
||||
'"}') COLLATE DATABASE_DEFAULT
|
||||
), N','
|
||||
) + N']'
|
||||
) AS all_fks_json
|
||||
FROM sys.foreign_keys AS fk
|
||||
@@ -31,299 +34,271 @@ const sqlServerQuery = `WITH fk_info AS (
|
||||
), pk_info AS (
|
||||
SELECT
|
||||
JSON_QUERY(
|
||||
'[' + STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{"schema": "' + COALESCE(REPLACE(pk.TABLE_SCHEMA, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(pk.TABLE_NAME, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column": "' + COALESCE(REPLACE(pk.COLUMN_NAME, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "pk_def": "PRIMARY KEY (' + pk.COLUMN_NAME COLLATE SQL_Latin1_General_CP1_CI_AS + ')"}')
|
||||
), ','
|
||||
) + N']'
|
||||
N'[' +
|
||||
STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(pk.TABLE_SCHEMA, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(pk.TABLE_NAME, '"', ''), ''), 'json') +
|
||||
'", "column": "' + STRING_ESCAPE(COALESCE(REPLACE(pk.COLUMN_NAME, '"', ''), ''), 'json') +
|
||||
'", "pk_def": "PRIMARY KEY (' + STRING_ESCAPE(pk.COLUMN_NAME, 'json') + N')"}') COLLATE DATABASE_DEFAULT
|
||||
), N','
|
||||
) + N']'
|
||||
) AS all_pks_json
|
||||
FROM
|
||||
(
|
||||
SELECT
|
||||
kcu.TABLE_SCHEMA,
|
||||
kcu.TABLE_NAME,
|
||||
kcu.COLUMN_NAME
|
||||
FROM
|
||||
INFORMATION_SCHEMA.KEY_COLUMN_USAGE kcu
|
||||
JOIN
|
||||
INFORMATION_SCHEMA.TABLE_CONSTRAINTS tc
|
||||
ON kcu.CONSTRAINT_NAME = tc.CONSTRAINT_NAME
|
||||
AND kcu.CONSTRAINT_SCHEMA = tc.CONSTRAINT_SCHEMA
|
||||
WHERE
|
||||
tc.CONSTRAINT_TYPE = 'PRIMARY KEY'
|
||||
) pk
|
||||
FROM (
|
||||
SELECT
|
||||
kcu.TABLE_SCHEMA,
|
||||
kcu.TABLE_NAME,
|
||||
kcu.COLUMN_NAME
|
||||
FROM INFORMATION_SCHEMA.KEY_COLUMN_USAGE kcu
|
||||
JOIN INFORMATION_SCHEMA.TABLE_CONSTRAINTS tc
|
||||
ON kcu.CONSTRAINT_NAME = tc.CONSTRAINT_NAME
|
||||
AND kcu.CONSTRAINT_SCHEMA = tc.CONSTRAINT_SCHEMA
|
||||
WHERE tc.CONSTRAINT_TYPE = 'PRIMARY KEY'
|
||||
) pk
|
||||
),
|
||||
cols AS (
|
||||
SELECT
|
||||
JSON_QUERY(
|
||||
'[' + STRING_AGG(
|
||||
JSON_QUERY(N'[' +
|
||||
STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY('{"schema": "' + COALESCE(REPLACE(cols.TABLE_SCHEMA, '"', ''), '') +
|
||||
'", "table": "' + COALESCE(REPLACE(cols.TABLE_NAME, '"', ''), '') +
|
||||
'", "name": "' + COALESCE(REPLACE(cols.COLUMN_NAME, '"', ''), '') +
|
||||
'", "ordinal_position": "' + CAST(cols.ORDINAL_POSITION AS NVARCHAR(MAX)) +
|
||||
'", "type": "' + LOWER(cols.DATA_TYPE) +
|
||||
'", "character_maximum_length": "' +
|
||||
COALESCE(CAST(cols.CHARACTER_MAXIMUM_LENGTH AS NVARCHAR(MAX)), 'null') +
|
||||
'", "precision": ' +
|
||||
CASE
|
||||
WHEN cols.DATA_TYPE IN ('numeric', 'decimal') THEN
|
||||
CONCAT('{"precision":', COALESCE(CAST(cols.NUMERIC_PRECISION AS NVARCHAR(MAX)), 'null'),
|
||||
',"scale":', COALESCE(CAST(cols.NUMERIC_SCALE AS NVARCHAR(MAX)), 'null'), '}')
|
||||
ELSE
|
||||
'null'
|
||||
END +
|
||||
', "nullable": ' +
|
||||
CASE WHEN cols.IS_NULLABLE = 'YES' THEN 'true' ELSE 'false' END +
|
||||
', "default": "' +
|
||||
COALESCE(REPLACE(CAST(cols.COLUMN_DEFAULT AS NVARCHAR(MAX)), '"', '\\"'), '') +
|
||||
'", "collation": "' +
|
||||
COALESCE(cols.COLLATION_NAME, '') +
|
||||
'"}')
|
||||
), ','
|
||||
) + ']'
|
||||
) AS all_columns_json
|
||||
FROM
|
||||
INFORMATION_SCHEMA.COLUMNS cols
|
||||
WHERE
|
||||
cols.TABLE_CATALOG = DB_NAME()
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(cols.TABLE_SCHEMA, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(cols.TABLE_NAME, '"', ''), ''), 'json') +
|
||||
'", "name": "' + STRING_ESCAPE(COALESCE(REPLACE(cols.COLUMN_NAME, '"', ''), ''), 'json') +
|
||||
'", "ordinal_position": ' + CAST(cols.ORDINAL_POSITION AS NVARCHAR(MAX)) +
|
||||
', "type": "' + STRING_ESCAPE(LOWER(cols.DATA_TYPE), 'json') +
|
||||
'", "character_maximum_length": ' +
|
||||
CASE
|
||||
WHEN cols.CHARACTER_MAXIMUM_LENGTH IS NULL THEN 'null'
|
||||
ELSE CAST(cols.CHARACTER_MAXIMUM_LENGTH AS NVARCHAR(MAX))
|
||||
END +
|
||||
', "precision": ' +
|
||||
CASE
|
||||
WHEN cols.DATA_TYPE IN ('numeric', 'decimal')
|
||||
THEN '{"precision":' + COALESCE(CAST(cols.NUMERIC_PRECISION AS NVARCHAR(MAX)), 'null') +
|
||||
',"scale":' + COALESCE(CAST(cols.NUMERIC_SCALE AS NVARCHAR(MAX)), 'null') + '}'
|
||||
ELSE 'null'
|
||||
END +
|
||||
', "nullable": ' + CASE WHEN cols.IS_NULLABLE = 'YES' THEN 'true' ELSE 'false' END +
|
||||
', "default": ' +
|
||||
'"' + STRING_ESCAPE(COALESCE(REPLACE(CAST(cols.COLUMN_DEFAULT AS NVARCHAR(MAX)), '"', '\\"'), ''), 'json') + '"' +
|
||||
', "collation": ' + CASE
|
||||
WHEN cols.COLLATION_NAME IS NULL THEN 'null'
|
||||
ELSE '"' + STRING_ESCAPE(cols.COLLATION_NAME, 'json') + '"'
|
||||
END +
|
||||
N'}') COLLATE DATABASE_DEFAULT
|
||||
), N','
|
||||
) +
|
||||
N']') AS all_columns_json
|
||||
FROM INFORMATION_SCHEMA.COLUMNS cols
|
||||
WHERE cols.TABLE_CATALOG = DB_NAME()
|
||||
),
|
||||
indexes AS (
|
||||
SELECT
|
||||
'[' + STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{"schema": "' + COALESCE(REPLACE(s.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(t.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "name": "' + COALESCE(REPLACE(i.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column": "' + COALESCE(REPLACE(c.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "index_type": "' + LOWER(i.type_desc) COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "unique": ' + CASE WHEN i.is_unique = 1 THEN 'true' ELSE 'false' END +
|
||||
', "direction": "' + CASE WHEN ic.is_descending_key = 1 THEN 'desc' ELSE 'asc' END COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column_position": ' + CAST(ic.key_ordinal AS nvarchar(max)) + N'}'
|
||||
)
|
||||
), ','
|
||||
) + N']' AS all_indexes_json
|
||||
FROM
|
||||
sys.indexes i
|
||||
JOIN
|
||||
sys.tables t ON i.object_id = t.object_id
|
||||
JOIN
|
||||
sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN
|
||||
sys.index_columns ic ON i.object_id = ic.object_id AND i.index_id = ic.index_id
|
||||
JOIN
|
||||
sys.columns c ON ic.object_id = c.object_id AND ic.column_id = c.column_id
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
AND i.name IS NOT NULL
|
||||
N'[' +
|
||||
STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(s.name, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(t.name, '"', ''), ''), 'json') +
|
||||
'", "name": "' + STRING_ESCAPE(COALESCE(REPLACE(i.name, '"', ''), ''), 'json') +
|
||||
'", "column": "' + STRING_ESCAPE(COALESCE(REPLACE(c.name, '"', ''), ''), 'json') +
|
||||
'", "index_type": "' + STRING_ESCAPE(LOWER(i.type_desc), 'json') +
|
||||
'", "unique": ' + CASE WHEN i.is_unique = 1 THEN 'true' ELSE 'false' END +
|
||||
', "direction": "' + CASE WHEN ic.is_descending_key = 1 THEN 'desc' ELSE 'asc' END +
|
||||
'", "column_position": ' + CAST(ic.key_ordinal AS nvarchar(max)) + N'}'
|
||||
) COLLATE DATABASE_DEFAULT
|
||||
), N','
|
||||
) +
|
||||
N']' AS all_indexes_json
|
||||
FROM sys.indexes i
|
||||
JOIN sys.tables t ON i.object_id = t.object_id
|
||||
JOIN sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN sys.index_columns ic ON i.object_id = ic.object_id AND i.index_id = ic.index_id
|
||||
JOIN sys.columns c ON ic.object_id = c.object_id AND ic.column_id = c.column_id
|
||||
WHERE s.name LIKE '%' AND i.name IS NOT NULL
|
||||
),
|
||||
tbls AS (
|
||||
SELECT
|
||||
'[' + STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{"schema": "' + COALESCE(REPLACE(aggregated.schema_name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(aggregated.table_name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "row_count": "' + CAST(aggregated.row_count AS NVARCHAR(MAX)) +
|
||||
'", "table_type": "' + aggregated.table_type COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "creation_date": "' + CONVERT(NVARCHAR(MAX), aggregated.creation_date, 120) + '"}'
|
||||
)
|
||||
), ','
|
||||
) + N']' AS all_tables_json
|
||||
FROM
|
||||
(
|
||||
-- Select from tables
|
||||
SELECT
|
||||
COALESCE(REPLACE(s.name, '"', ''), '') AS schema_name,
|
||||
COALESCE(REPLACE(t.name, '"', ''), '') AS table_name,
|
||||
SUM(p.rows) AS row_count,
|
||||
t.type_desc AS table_type,
|
||||
t.create_date AS creation_date
|
||||
FROM
|
||||
sys.tables t
|
||||
JOIN
|
||||
sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN
|
||||
sys.partitions p ON t.object_id = p.object_id AND p.index_id IN (0, 1)
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
GROUP BY
|
||||
s.name, t.name, t.type_desc, t.create_date
|
||||
N'[' + STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(aggregated.schema_name, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(aggregated.table_name, '"', ''), ''), 'json') +
|
||||
'", "row_count": ' + CAST(aggregated.row_count AS NVARCHAR(MAX)) +
|
||||
', "table_type": "' + STRING_ESCAPE(aggregated.table_type, 'json') +
|
||||
'", "creation_date": "' + CONVERT(NVARCHAR(MAX), aggregated.creation_date, 120) + N'"}'
|
||||
) COLLATE DATABASE_DEFAULT
|
||||
), N','
|
||||
) +
|
||||
N']' AS all_tables_json
|
||||
FROM (
|
||||
SELECT
|
||||
COALESCE(REPLACE(s.name, '"', ''), '') AS schema_name,
|
||||
COALESCE(REPLACE(t.name, '"', ''), '') AS table_name,
|
||||
SUM(p.rows) AS row_count,
|
||||
t.type_desc AS table_type,
|
||||
t.create_date AS creation_date
|
||||
FROM sys.tables t
|
||||
JOIN sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN sys.partitions p ON t.object_id = p.object_id AND p.index_id IN (0, 1)
|
||||
WHERE s.name LIKE '%'
|
||||
GROUP BY s.name, t.name, t.type_desc, t.create_date
|
||||
|
||||
UNION ALL
|
||||
UNION ALL
|
||||
|
||||
-- Select from views
|
||||
SELECT
|
||||
COALESCE(REPLACE(s.name, '"', ''), '') AS table_name,
|
||||
COALESCE(REPLACE(v.name, '"', ''), '') AS object_name,
|
||||
0 AS row_count, -- Views don't have row counts
|
||||
'VIEW' AS table_type,
|
||||
v.create_date AS creation_date
|
||||
FROM
|
||||
sys.views v
|
||||
JOIN
|
||||
sys.schemas s ON v.schema_id = s.schema_id
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
) AS aggregated
|
||||
SELECT
|
||||
COALESCE(REPLACE(s.name, '"', ''), '') AS table_name,
|
||||
COALESCE(REPLACE(v.name, '"', ''), '') AS object_name,
|
||||
0 AS row_count,
|
||||
'VIEW' AS table_type,
|
||||
v.create_date AS creation_date
|
||||
FROM sys.views v
|
||||
JOIN sys.schemas s ON v.schema_id = s.schema_id
|
||||
WHERE s.name LIKE '%'
|
||||
) AS aggregated
|
||||
),
|
||||
views AS (
|
||||
SELECT
|
||||
'[' + STRING_AGG(
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{"schema": "' + STRING_ESCAPE(COALESCE(s.name, ''), 'json') +
|
||||
'", "view_name": "' + STRING_ESCAPE(COALESCE(v.name, ''), 'json') +
|
||||
'", "view_definition": "' +
|
||||
STRING_ESCAPE(
|
||||
CAST(
|
||||
'' AS XML
|
||||
).value(
|
||||
'xs:base64Binary(sql:column("DefinitionBinary"))',
|
||||
'VARCHAR(MAX)'
|
||||
), 'json') +
|
||||
'"}'
|
||||
)
|
||||
), ','
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(s.name, '"', ''), ''), 'json') +
|
||||
'", "view_name": "' + STRING_ESCAPE(COALESCE(REPLACE(v.name, '"', ''), ''), 'json') +
|
||||
'", "view_definition": "' +
|
||||
STRING_ESCAPE(
|
||||
CAST(
|
||||
'' AS XML
|
||||
).value(
|
||||
'xs:base64Binary(sql:column("DefinitionBinary"))',
|
||||
'VARCHAR(MAX)'
|
||||
), 'json') +
|
||||
N'"}') COLLATE DATABASE_DEFAULT
|
||||
), N','
|
||||
) + N']' AS all_views_json
|
||||
FROM
|
||||
sys.views v
|
||||
JOIN
|
||||
sys.schemas s ON v.schema_id = s.schema_id
|
||||
JOIN
|
||||
sys.sql_modules m ON v.object_id = m.object_id
|
||||
FROM sys.views v
|
||||
JOIN sys.schemas s ON v.schema_id = s.schema_id
|
||||
JOIN sys.sql_modules m ON v.object_id = m.object_id
|
||||
CROSS APPLY
|
||||
(SELECT CONVERT(VARBINARY(MAX), m.definition) AS DefinitionBinary) AS bin
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
WHERE s.name LIKE '%'
|
||||
)
|
||||
SELECT JSON_QUERY(
|
||||
N'{"fk_info": ' + ISNULL((SELECT cast(all_fks_json as nvarchar(max)) FROM fk_info), N'[]') +
|
||||
N'{
|
||||
"fk_info": ' + ISNULL((SELECT cast(all_fks_json as nvarchar(max)) FROM fk_info), N'[]') +
|
||||
', "pk_info": ' + ISNULL((SELECT cast(all_pks_json as nvarchar(max)) FROM pk_info), N'[]') +
|
||||
', "columns": ' + ISNULL((SELECT cast(all_columns_json as nvarchar(max)) FROM cols), N'[]') +
|
||||
', "indexes": ' + ISNULL((SELECT cast(all_indexes_json as nvarchar(max)) FROM indexes), N'[]') +
|
||||
', "tables": ' + ISNULL((SELECT cast(all_tables_json as nvarchar(max)) FROM tbls), N'[]') +
|
||||
', "views": ' + ISNULL((SELECT cast(all_views_json as nvarchar(max)) FROM views), N'[]') +
|
||||
', "database_name": "' + DB_NAME() + '"' +
|
||||
', "version": ""}'
|
||||
', "database_name": "' + STRING_ESCAPE(DB_NAME(), 'json') +
|
||||
'", "version": ""
|
||||
}'
|
||||
) AS metadata_json_to_import;
|
||||
`;
|
||||
|
||||
const sqlServer2016AndBelowQuery = `WITH fk_info AS (
|
||||
SELECT
|
||||
JSON_QUERY(
|
||||
'[' + ISNULL(
|
||||
STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{"schema": "' + COALESCE(REPLACE(tp_schema.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(tp.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column": "' + COALESCE(REPLACE(cp.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "foreign_key_name": "' + COALESCE(REPLACE(fk.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "reference_schema": "' + COALESCE(REPLACE(tr_schema.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "reference_table": "' + COALESCE(REPLACE(tr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "reference_column": "' + COALESCE(REPLACE(cr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "fk_def": "FOREIGN KEY (' + COALESCE(REPLACE(cp.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
') REFERENCES ' + COALESCE(REPLACE(tr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'(' + COALESCE(REPLACE(cr.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
') ON DELETE ' + fk.delete_referential_action_desc COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
' ON UPDATE ' + fk.update_referential_action_desc COLLATE SQL_Latin1_General_CP1_CI_AS + '"}')
|
||||
)
|
||||
FROM
|
||||
sys.foreign_keys AS fk
|
||||
JOIN
|
||||
sys.foreign_key_columns AS fkc ON fk.object_id = fkc.constraint_object_id
|
||||
JOIN
|
||||
sys.tables AS tp ON fkc.parent_object_id = tp.object_id
|
||||
JOIN
|
||||
sys.schemas AS tp_schema ON tp.schema_id = tp_schema.schema_id
|
||||
JOIN
|
||||
sys.columns AS cp ON fkc.parent_object_id = cp.object_id AND fkc.parent_column_id = cp.column_id
|
||||
JOIN
|
||||
sys.tables AS tr ON fkc.referenced_object_id = tr.object_id
|
||||
JOIN
|
||||
sys.schemas AS tr_schema ON tr.schema_id = tr_schema.schema_id
|
||||
JOIN
|
||||
sys.columns AS cr ON fkc.referenced_object_id = cr.object_id AND fkc.referenced_column_id = cr.column_id
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
+ N']'
|
||||
) AS all_fks_json
|
||||
const sqlServer2016AndBelowQuery = `${`/* SQL Server 2016 and below edition (13.0, 12.0, 11.0..) */`}
|
||||
WITH fk_info AS (
|
||||
SELECT JSON_QUERY('[' +
|
||||
ISNULL(
|
||||
STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(tp_schema.name, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(tp.name, '"', ''), ''), 'json') +
|
||||
'", "column": "' + STRING_ESCAPE(COALESCE(REPLACE(cp.name, '"', ''), ''), 'json') +
|
||||
'", "foreign_key_name": "' + STRING_ESCAPE(COALESCE(REPLACE(fk.name, '"', ''), ''), 'json') +
|
||||
'", "reference_schema": "' + STRING_ESCAPE(COALESCE(REPLACE(tr_schema.name, '"', ''), ''), 'json') +
|
||||
'", "reference_table": "' + STRING_ESCAPE(COALESCE(REPLACE(tr.name, '"', ''), ''), 'json') +
|
||||
'", "reference_column": "' + STRING_ESCAPE(COALESCE(REPLACE(cr.name, '"', ''), ''), 'json') +
|
||||
'", "fk_def": "FOREIGN KEY (' + STRING_ESCAPE(COALESCE(REPLACE(cp.name, '"', ''), ''), 'json') +
|
||||
') REFERENCES ' + STRING_ESCAPE(COALESCE(REPLACE(tr.name, '"', ''), ''), 'json') +
|
||||
'(' + STRING_ESCAPE(COALESCE(REPLACE(cr.name, '"', ''), ''), 'json') +
|
||||
') ON DELETE ' + STRING_ESCAPE(fk.delete_referential_action_desc, 'json') +
|
||||
' ON UPDATE ' + STRING_ESCAPE(fk.update_referential_action_desc, 'json') +
|
||||
'"}') COLLATE DATABASE_DEFAULT
|
||||
)
|
||||
FROM sys.foreign_keys AS fk
|
||||
JOIN sys.foreign_key_columns AS fkc ON fk.object_id = fkc.constraint_object_id
|
||||
JOIN sys.tables AS tp ON fkc.parent_object_id = tp.object_id
|
||||
JOIN sys.schemas AS tp_schema ON tp.schema_id = tp_schema.schema_id
|
||||
JOIN sys.columns AS cp ON fkc.parent_object_id = cp.object_id AND fkc.parent_column_id = cp.column_id
|
||||
JOIN sys.tables AS tr ON fkc.referenced_object_id = tr.object_id
|
||||
JOIN sys.schemas AS tr_schema ON tr.schema_id = tr_schema.schema_id
|
||||
JOIN sys.columns AS cr ON fkc.referenced_object_id = cr.object_id AND fkc.referenced_column_id = cr.column_id
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
+ N']') AS all_fks_json
|
||||
),
|
||||
pk_info AS (
|
||||
SELECT
|
||||
JSON_QUERY(
|
||||
'[' + ISNULL(
|
||||
STUFF((
|
||||
SELECT JSON_QUERY('[' +
|
||||
ISNULL(STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(N'{"schema": "' + COALESCE(REPLACE(pk.TABLE_SCHEMA, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(pk.TABLE_NAME, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column": "' + COALESCE(REPLACE(pk.COLUMN_NAME, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "pk_def": "PRIMARY KEY (' + pk.COLUMN_NAME COLLATE SQL_Latin1_General_CP1_CI_AS + ')"}')
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(pk.TABLE_SCHEMA, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(pk.TABLE_NAME, '"', ''), ''), 'json') +
|
||||
'", "column": "' + STRING_ESCAPE(COALESCE(REPLACE(pk.COLUMN_NAME, '"', ''), ''), 'json') +
|
||||
'", "pk_def": "PRIMARY KEY (' + STRING_ESCAPE(pk.COLUMN_NAME, 'json') + N')"}') COLLATE DATABASE_DEFAULT
|
||||
)
|
||||
FROM
|
||||
(
|
||||
SELECT
|
||||
kcu.TABLE_SCHEMA,
|
||||
kcu.TABLE_NAME,
|
||||
kcu.COLUMN_NAME
|
||||
FROM
|
||||
INFORMATION_SCHEMA.KEY_COLUMN_USAGE kcu
|
||||
JOIN
|
||||
INFORMATION_SCHEMA.TABLE_CONSTRAINTS tc
|
||||
SELECT kcu.TABLE_SCHEMA,
|
||||
kcu.TABLE_NAME,
|
||||
kcu.COLUMN_NAME
|
||||
FROM INFORMATION_SCHEMA.KEY_COLUMN_USAGE kcu
|
||||
JOIN INFORMATION_SCHEMA.TABLE_CONSTRAINTS tc
|
||||
ON kcu.CONSTRAINT_NAME = tc.CONSTRAINT_NAME
|
||||
AND kcu.CONSTRAINT_SCHEMA = tc.CONSTRAINT_SCHEMA
|
||||
WHERE
|
||||
tc.CONSTRAINT_TYPE = 'PRIMARY KEY'
|
||||
WHERE tc.CONSTRAINT_TYPE = 'PRIMARY KEY'
|
||||
) pk
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
+ N']'
|
||||
) AS all_pks_json
|
||||
+ N']') AS all_pks_json
|
||||
),
|
||||
cols AS (
|
||||
SELECT
|
||||
JSON_QUERY(
|
||||
'[' + ISNULL(
|
||||
STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY('{"schema": "' + COALESCE(REPLACE(cols.TABLE_SCHEMA, '"', ''), '') +
|
||||
'", "table": "' + COALESCE(REPLACE(cols.TABLE_NAME, '"', ''), '') +
|
||||
'", "name": "' + COALESCE(REPLACE(cols.COLUMN_NAME, '"', ''), '') +
|
||||
'", "ordinal_position": "' + CAST(cols.ORDINAL_POSITION AS NVARCHAR(MAX)) +
|
||||
'", "type": "' + LOWER(cols.DATA_TYPE) +
|
||||
'", "character_maximum_length": "' +
|
||||
COALESCE(CAST(cols.CHARACTER_MAXIMUM_LENGTH AS NVARCHAR(MAX)), 'null') +
|
||||
'", "precision": ' +
|
||||
CASE
|
||||
WHEN cols.DATA_TYPE IN ('numeric', 'decimal') THEN
|
||||
CONCAT('{"precision":', COALESCE(CAST(cols.NUMERIC_PRECISION AS NVARCHAR(MAX)), 'null'),
|
||||
',"scale":', COALESCE(CAST(cols.NUMERIC_SCALE AS NVARCHAR(MAX)), 'null'), '}')
|
||||
ELSE
|
||||
'null'
|
||||
END +
|
||||
', "nullable": ' +
|
||||
CASE WHEN cols.IS_NULLABLE = 'YES' THEN 'true' ELSE 'false' END +
|
||||
', "default": "' +
|
||||
COALESCE(REPLACE(CAST(cols.COLUMN_DEFAULT AS NVARCHAR(MAX)), '"', '"'), '') +
|
||||
'", "collation": "' +
|
||||
COALESCE(cols.COLLATION_NAME, '') +
|
||||
'"}')
|
||||
)
|
||||
FROM
|
||||
INFORMATION_SCHEMA.COLUMNS cols
|
||||
WHERE
|
||||
cols.TABLE_CATALOG = DB_NAME()
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
+ ']'
|
||||
) AS all_columns_json
|
||||
SELECT JSON_QUERY('[' +
|
||||
ISNULL(
|
||||
STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY('{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(cols.TABLE_SCHEMA, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(cols.TABLE_NAME, '"', ''), ''), 'json') +
|
||||
'", "name": "' + STRING_ESCAPE(COALESCE(REPLACE(cols.COLUMN_NAME, '"', ''), ''), 'json') +
|
||||
'", "ordinal_position": ' + CAST(cols.ORDINAL_POSITION AS NVARCHAR(MAX)) +
|
||||
', "type": "' + STRING_ESCAPE(LOWER(cols.DATA_TYPE), 'json') +
|
||||
'", "character_maximum_length": ' +
|
||||
CASE
|
||||
WHEN cols.CHARACTER_MAXIMUM_LENGTH IS NULL THEN 'null'
|
||||
ELSE CAST(cols.CHARACTER_MAXIMUM_LENGTH AS NVARCHAR(MAX))
|
||||
END +
|
||||
', "precision": ' +
|
||||
CASE
|
||||
WHEN cols.DATA_TYPE IN ('numeric', 'decimal')
|
||||
THEN '{"precision":' + COALESCE(CAST(cols.NUMERIC_PRECISION AS NVARCHAR(MAX)), 'null') +
|
||||
',"scale":' + COALESCE(CAST(cols.NUMERIC_SCALE AS NVARCHAR(MAX)), 'null') + '}'
|
||||
ELSE 'null'
|
||||
END +
|
||||
', "nullable": ' + CASE WHEN cols.IS_NULLABLE = 'YES' THEN 'true' ELSE 'false' END +
|
||||
', "default": ' +
|
||||
'"' + STRING_ESCAPE(COALESCE(REPLACE(CAST(cols.COLUMN_DEFAULT AS NVARCHAR(MAX)), '"', '\\"'), ''), 'json') + '"' +
|
||||
', "collation": ' +
|
||||
CASE
|
||||
WHEN cols.COLLATION_NAME IS NULL THEN 'null'
|
||||
ELSE '"' + STRING_ESCAPE(cols.COLLATION_NAME, 'json') + '"'
|
||||
END +
|
||||
N'}')
|
||||
)
|
||||
FROM
|
||||
INFORMATION_SCHEMA.COLUMNS cols
|
||||
WHERE
|
||||
cols.TABLE_CATALOG = DB_NAME()
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
+ ']') AS all_columns_json
|
||||
),
|
||||
indexes AS (
|
||||
SELECT
|
||||
@@ -331,30 +306,24 @@ indexes AS (
|
||||
STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{"schema": "' + COALESCE(REPLACE(s.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(t.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "name": "' + COALESCE(REPLACE(i.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "column": "' + COALESCE(REPLACE(c.name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "index_type": "' + LOWER(i.type_desc) COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(s.name, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(t.name, '"', ''), ''), 'json') +
|
||||
'", "name": "' + STRING_ESCAPE(COALESCE(REPLACE(i.name, '"', ''), ''), 'json') +
|
||||
'", "column": "' + STRING_ESCAPE(COALESCE(REPLACE(c.name, '"', ''), ''), 'json') +
|
||||
'", "index_type": "' + STRING_ESCAPE(LOWER(i.type_desc), 'json') +
|
||||
'", "unique": ' + CASE WHEN i.is_unique = 1 THEN 'true' ELSE 'false' END +
|
||||
', "direction": "' + CASE WHEN ic.is_descending_key = 1 THEN 'desc' ELSE 'asc' END COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
', "direction": "' + CASE WHEN ic.is_descending_key = 1 THEN 'desc' ELSE 'asc' END +
|
||||
'", "column_position": ' + CAST(ic.key_ordinal AS nvarchar(max)) + N'}'
|
||||
)
|
||||
) COLLATE DATABASE_DEFAULT
|
||||
)
|
||||
FROM
|
||||
sys.indexes i
|
||||
JOIN
|
||||
sys.tables t ON i.object_id = t.object_id
|
||||
JOIN
|
||||
sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN
|
||||
sys.index_columns ic ON i.object_id = ic.object_id AND i.index_id = ic.index_id
|
||||
JOIN
|
||||
sys.columns c ON ic.object_id = c.object_id AND ic.column_id = c.column_id
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
AND i.name IS NOT NULL
|
||||
FROM sys.indexes i
|
||||
JOIN sys.tables t ON i.object_id = t.object_id
|
||||
JOIN sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN sys.index_columns ic ON i.object_id = ic.object_id AND i.index_id = ic.index_id
|
||||
JOIN sys.columns c ON ic.object_id = c.object_id AND ic.column_id = c.column_id
|
||||
WHERE s.name LIKE '%'
|
||||
AND i.name IS NOT NULL
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
+ N']' AS all_indexes_json
|
||||
@@ -365,12 +334,12 @@ tbls AS (
|
||||
STUFF((
|
||||
SELECT ',' +
|
||||
CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{"schema": "' + COALESCE(REPLACE(aggregated.schema_name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "table": "' + COALESCE(REPLACE(aggregated.object_name, '"', ''), '') COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "row_count": "' + CAST(aggregated.row_count AS NVARCHAR(MAX)) +
|
||||
'", "object_type": "' + aggregated.object_type COLLATE SQL_Latin1_General_CP1_CI_AS +
|
||||
'", "creation_date": "' + CONVERT(NVARCHAR(MAX), aggregated.creation_date, 120) + '"}'
|
||||
JSON_QUERY(N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(aggregated.schema_name, '"', ''), ''), 'json') +
|
||||
'", "table": "' + STRING_ESCAPE(COALESCE(REPLACE(aggregated.table_name, '"', ''), ''), 'json') +
|
||||
'", "row_count": ' + CAST(aggregated.row_count AS NVARCHAR(MAX)) +
|
||||
', "table_type": "' + STRING_ESCAPE(aggregated.table_type, 'json') +
|
||||
'", "creation_date": "' + CONVERT(NVARCHAR(MAX), aggregated.creation_date, 120) + N'"}'
|
||||
)
|
||||
)
|
||||
FROM
|
||||
@@ -378,20 +347,15 @@ tbls AS (
|
||||
-- Select from tables
|
||||
SELECT
|
||||
COALESCE(REPLACE(s.name, '"', ''), '') AS schema_name,
|
||||
COALESCE(REPLACE(t.name, '"', ''), '') AS object_name,
|
||||
COALESCE(REPLACE(t.name, '"', ''), '') AS table_name,
|
||||
SUM(p.rows) AS row_count,
|
||||
t.type_desc AS object_type,
|
||||
t.type_desc AS table_type,
|
||||
t.create_date AS creation_date
|
||||
FROM
|
||||
sys.tables t
|
||||
JOIN
|
||||
sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN
|
||||
sys.partitions p ON t.object_id = p.object_id AND p.index_id IN (0, 1)
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
GROUP BY
|
||||
s.name, t.name, t.type_desc, t.create_date
|
||||
FROM sys.tables t
|
||||
JOIN sys.schemas s ON t.schema_id = s.schema_id
|
||||
JOIN sys.partitions p ON t.object_id = p.object_id AND p.index_id IN (0, 1)
|
||||
WHERE s.name LIKE '%'
|
||||
GROUP BY s.name, t.name, t.type_desc, t.create_date
|
||||
|
||||
UNION ALL
|
||||
|
||||
@@ -402,12 +366,9 @@ tbls AS (
|
||||
0 AS row_count, -- Views don't have row counts
|
||||
'VIEW' AS object_type,
|
||||
v.create_date AS creation_date
|
||||
FROM
|
||||
sys.views v
|
||||
JOIN
|
||||
sys.schemas s ON v.schema_id = s.schema_id
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
FROM sys.views v
|
||||
JOIN sys.schemas s ON v.schema_id = s.schema_id
|
||||
WHERE s.name LIKE '%'
|
||||
) AS aggregated
|
||||
FOR XML PATH('')
|
||||
), 1, 1, ''), '')
|
||||
@@ -417,38 +378,40 @@ views AS (
|
||||
SELECT
|
||||
'[' +
|
||||
(
|
||||
SELECT
|
||||
STUFF((
|
||||
SELECT ',' + CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{"schema": "' + COALESCE(REPLACE(s.name, '"', ''), '') +
|
||||
'", "view_name": "' + COALESCE(REPLACE(v.name, '"', ''), '') +
|
||||
'", "view_definition": "' +
|
||||
CAST(
|
||||
(
|
||||
SELECT CAST(OBJECT_DEFINITION(v.object_id) AS VARBINARY(MAX)) FOR XML PATH('')
|
||||
) AS NVARCHAR(MAX)
|
||||
) + '"}'
|
||||
SELECT STUFF((
|
||||
SELECT ',' + CONVERT(nvarchar(max),
|
||||
JSON_QUERY(
|
||||
N'{
|
||||
"schema": "' + STRING_ESCAPE(COALESCE(REPLACE(s.name, '"', ''), ''), 'json') +
|
||||
'", "view_name": "' + STRING_ESCAPE(COALESCE(REPLACE(v.name, '"', ''), ''), 'json') +
|
||||
'", "view_definition": "' +
|
||||
CAST(
|
||||
(
|
||||
SELECT CAST(OBJECT_DEFINITION(v.object_id) AS VARBINARY(MAX)) FOR XML PATH('')
|
||||
) AS NVARCHAR(MAX)
|
||||
) + N'"}'
|
||||
)
|
||||
)
|
||||
)
|
||||
FROM
|
||||
sys.views v
|
||||
JOIN
|
||||
sys.schemas s ON v.schema_id = s.schema_id
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
FOR XML PATH(''), TYPE).value('.', 'NVARCHAR(MAX)'), 1, 1, '')
|
||||
FROM
|
||||
sys.views v
|
||||
JOIN
|
||||
sys.schemas s ON v.schema_id = s.schema_id
|
||||
WHERE
|
||||
s.name LIKE '%'
|
||||
FOR XML PATH(''), TYPE).value('.', 'NVARCHAR(MAX)'), 1, 1, '')
|
||||
) + ']' AS all_views_json
|
||||
)
|
||||
SELECT JSON_QUERY(
|
||||
N'{"fk_info": ' + ISNULL((SELECT cast(all_fks_json as nvarchar(max)) FROM fk_info), N'[]') +
|
||||
N'{
|
||||
"fk_info": ' + ISNULL((SELECT cast(all_fks_json as nvarchar(max)) FROM fk_info), N'[]') +
|
||||
', "pk_info": ' + ISNULL((SELECT cast(all_pks_json as nvarchar(max)) FROM pk_info), N'[]') +
|
||||
', "columns": ' + ISNULL((SELECT cast(all_columns_json as nvarchar(max)) FROM cols), N'[]') +
|
||||
', "indexes": ' + ISNULL((SELECT cast(all_indexes_json as nvarchar(max)) FROM indexes), N'[]') +
|
||||
', "tables": ' + ISNULL((SELECT cast(all_objects_json as nvarchar(max)) FROM tbls), N'[]') +
|
||||
', "views": ' + ISNULL((SELECT cast(all_views_json as nvarchar(max)) FROM views), N'[]') +
|
||||
', "database_name": "' + DB_NAME() + '"' +
|
||||
', "version": ""}'
|
||||
', "version": ""
|
||||
}'
|
||||
) AS metadata_json_to_import;`;
|
||||
|
||||
export const getSqlServerQuery = (
|
||||
|
||||
@@ -10,14 +10,20 @@ export const fixMetadataJson = async (
|
||||
return (
|
||||
metadataJson
|
||||
.trim()
|
||||
// First unescape the JSON string
|
||||
.replace(/\\"/g, '"')
|
||||
.replace(/\\\\/g, '\\')
|
||||
.replace(/^[^{]*/, '') // Remove everything before the first '{'
|
||||
.replace(/}[^}]*$/, '}') // Remove everything after the last '}'
|
||||
.replace(/:""([^"]+)""/g, ':"$1"') // Convert :""value"" to :"value"
|
||||
.replace(/""(\w+)""/g, '"$1"') // Convert ""key"" to "key"
|
||||
.replace(/^\s+|\s+$/g, '')
|
||||
.replace(/^"|"$/g, '')
|
||||
.replace(/^'|'$/g, '')
|
||||
.replace(/""""/g, '""') // Remove Quadruple quotes from keys
|
||||
.replace(/"""([^",}]+)"""/g, '"$1"') // Remove tripple quotes from keys
|
||||
.replace(/""([^",}]+)""/g, '"$1"') // Remove double quotes from keys
|
||||
|
||||
/* eslint-disable-next-line no-useless-escape */
|
||||
.replace(/\"/g, '___ESCAPED_QUOTE___') // Temporarily replace empty strings
|
||||
.replace(/(?<=:\s*)""(?=\s*[,}])/g, '___EMPTY___') // Temporarily replace empty strings
|
||||
|
||||
@@ -28,10 +28,24 @@ interface DBMLField {
|
||||
increment?: boolean;
|
||||
}
|
||||
|
||||
interface DBMLIndexColumn {
|
||||
value: string;
|
||||
type?: string;
|
||||
length?: number;
|
||||
order?: 'asc' | 'desc';
|
||||
}
|
||||
|
||||
interface DBMLIndex {
|
||||
columns: string | (string | DBMLIndexColumn)[];
|
||||
unique?: boolean;
|
||||
name?: string;
|
||||
}
|
||||
|
||||
interface DBMLTable {
|
||||
name: string;
|
||||
schema?: string | { name: string };
|
||||
fields: DBMLField[];
|
||||
indexes?: DBMLIndex[];
|
||||
}
|
||||
|
||||
interface DBMLEndpoint {
|
||||
@@ -99,18 +113,60 @@ export const importDBMLToDiagram = async (
|
||||
|
||||
// Extract only the necessary data from the parsed DBML
|
||||
const extractedData = {
|
||||
tables: dbmlData.tables.map((table: DBMLTable) => ({
|
||||
name: table.name,
|
||||
schema: table.schema,
|
||||
fields: table.fields.map((field: DBMLField) => ({
|
||||
name: field.name,
|
||||
type: field.type,
|
||||
unique: field.unique,
|
||||
pk: field.pk,
|
||||
not_null: field.not_null,
|
||||
increment: field.increment,
|
||||
})),
|
||||
})),
|
||||
tables: (dbmlData.tables as unknown as DBMLTable[]).map(
|
||||
(table) => ({
|
||||
name: table.name,
|
||||
schema: table.schema,
|
||||
fields: table.fields.map((field: DBMLField) => ({
|
||||
name: field.name,
|
||||
type: field.type,
|
||||
unique: field.unique,
|
||||
pk: field.pk,
|
||||
not_null: field.not_null,
|
||||
increment: field.increment,
|
||||
})),
|
||||
indexes:
|
||||
table.indexes?.map((dbmlIndex) => {
|
||||
let indexColumns: string[];
|
||||
|
||||
// Handle composite index case "(col1, col2)"
|
||||
if (typeof dbmlIndex.columns === 'string') {
|
||||
if (dbmlIndex.columns.includes('(')) {
|
||||
// Composite index
|
||||
const columnsStr =
|
||||
dbmlIndex.columns.replace(/[()]/g, '');
|
||||
indexColumns = columnsStr
|
||||
.split(',')
|
||||
.map((c) => c.trim());
|
||||
} else {
|
||||
// Single column
|
||||
indexColumns = [dbmlIndex.columns.trim()];
|
||||
}
|
||||
} else {
|
||||
// Handle array of columns
|
||||
indexColumns = Array.isArray(dbmlIndex.columns)
|
||||
? dbmlIndex.columns.map((col) =>
|
||||
typeof col === 'object' &&
|
||||
'value' in col
|
||||
? (col.value as string).trim()
|
||||
: (col as string).trim()
|
||||
)
|
||||
: [String(dbmlIndex.columns).trim()];
|
||||
}
|
||||
|
||||
// Generate a consistent index name
|
||||
const indexName =
|
||||
dbmlIndex.name ||
|
||||
`idx_${table.name}_${indexColumns.join('_')}`;
|
||||
|
||||
return {
|
||||
columns: indexColumns,
|
||||
unique: dbmlIndex.unique || false,
|
||||
name: indexName,
|
||||
};
|
||||
}) || [],
|
||||
})
|
||||
),
|
||||
refs: (dbmlData.refs as unknown as DBMLRef[]).map((ref) => ({
|
||||
endpoints: (ref.endpoints as [DBMLEndpoint, DBMLEndpoint]).map(
|
||||
(endpoint) => ({
|
||||
@@ -126,7 +182,42 @@ export const importDBMLToDiagram = async (
|
||||
const tables: DBTable[] = extractedData.tables.map((table, index) => {
|
||||
const row = Math.floor(index / 4);
|
||||
const col = index % 4;
|
||||
const tableSpacing = 300; // Increased spacing between tables
|
||||
const tableSpacing = 300;
|
||||
|
||||
// Create fields first so we have their IDs
|
||||
const fields = table.fields.map((field) => ({
|
||||
id: generateId(),
|
||||
name: field.name.replace(/['"]/g, ''),
|
||||
type: mapDBMLTypeToGenericType(field.type.type_name),
|
||||
nullable: !field.not_null,
|
||||
primaryKey: field.pk || false,
|
||||
unique: field.unique || false,
|
||||
createdAt: Date.now(),
|
||||
}));
|
||||
|
||||
// Convert DBML indexes to ChartDB indexes
|
||||
const indexes =
|
||||
table.indexes?.map((dbmlIndex) => {
|
||||
const fieldIds = dbmlIndex.columns.map((columnName) => {
|
||||
const field = fields.find((f) => f.name === columnName);
|
||||
if (!field) {
|
||||
throw new Error(
|
||||
`Index references non-existent column: ${columnName}`
|
||||
);
|
||||
}
|
||||
return field.id;
|
||||
});
|
||||
|
||||
return {
|
||||
id: generateId(),
|
||||
name:
|
||||
dbmlIndex.name ||
|
||||
`idx_${table.name}_${dbmlIndex.columns.join('_')}`,
|
||||
fieldIds,
|
||||
unique: dbmlIndex.unique || false,
|
||||
createdAt: Date.now(),
|
||||
};
|
||||
}) || [];
|
||||
|
||||
return {
|
||||
id: generateId(),
|
||||
@@ -136,18 +227,10 @@ export const importDBMLToDiagram = async (
|
||||
? table.schema
|
||||
: table.schema?.name || '',
|
||||
order: index,
|
||||
fields: table.fields.map((field) => ({
|
||||
id: generateId(),
|
||||
name: field.name.replace(/['"]/g, ''),
|
||||
type: mapDBMLTypeToGenericType(field.type.type_name),
|
||||
nullable: !field.not_null,
|
||||
primaryKey: field.pk || false,
|
||||
unique: field.unique || false,
|
||||
createdAt: Date.now(),
|
||||
})),
|
||||
fields,
|
||||
indexes,
|
||||
x: col * tableSpacing,
|
||||
y: row * tableSpacing,
|
||||
indexes: [],
|
||||
color: randomColor(),
|
||||
isView: false,
|
||||
createdAt: Date.now(),
|
||||
|
||||
@@ -1,5 +1,11 @@
|
||||
export const OPENAI_API_KEY: string = import.meta.env.VITE_OPENAI_API_KEY;
|
||||
export const OPENAI_API_ENDPOINT: string = import.meta.env
|
||||
.VITE_OPENAI_API_ENDPOINT;
|
||||
export const LLM_MODEL_NAME: string = import.meta.env.VITE_LLM_MODEL_NAME;
|
||||
export const IS_CHARTDB_IO: boolean =
|
||||
import.meta.env.VITE_IS_CHARTDB_IO === 'true';
|
||||
export const APP_URL: string = import.meta.env.VITE_APP_URL;
|
||||
export const HOST_URL: string = import.meta.env.VITE_HOST_URL ?? '';
|
||||
export const HIDE_BUCKLE_DOT_DEV: boolean =
|
||||
(window?.env?.HIDE_BUCKLE_DOT_DEV ??
|
||||
import.meta.env.VITE_HIDE_BUCKLE_DOT_DEV) === 'true';
|
||||
|
||||
@@ -9,9 +9,10 @@ import { useChartDB } from '@/hooks/use-chartdb';
|
||||
import { useLayout } from '@/hooks/use-layout';
|
||||
import { cloneTable } from '@/lib/clone';
|
||||
import type { DBTable } from '@/lib/domain/db-table';
|
||||
import { Copy, Pencil, Trash2 } from 'lucide-react';
|
||||
import { Copy, Pencil, Trash2, Workflow } from 'lucide-react';
|
||||
import React, { useCallback } from 'react';
|
||||
import { useTranslation } from 'react-i18next';
|
||||
import { useDialog } from '@/hooks/use-dialog';
|
||||
|
||||
export interface TableNodeContextMenuProps {
|
||||
table: DBTable;
|
||||
@@ -24,6 +25,7 @@ export const TableNodeContextMenu: React.FC<
|
||||
const { openTableFromSidebar } = useLayout();
|
||||
const { t } = useTranslation();
|
||||
const { isMd: isDesktop } = useBreakpoint('md');
|
||||
const { openCreateRelationshipDialog } = useDialog();
|
||||
|
||||
const duplicateTableHandler = useCallback(() => {
|
||||
const clonedTable = cloneTable(table);
|
||||
@@ -43,6 +45,12 @@ export const TableNodeContextMenu: React.FC<
|
||||
removeTable(table.id);
|
||||
}, [removeTable, table.id]);
|
||||
|
||||
const addRelationshipHandler = useCallback(() => {
|
||||
openCreateRelationshipDialog({
|
||||
sourceTableId: table.id,
|
||||
});
|
||||
}, [openCreateRelationshipDialog, table.id]);
|
||||
|
||||
if (!isDesktop || readonly) {
|
||||
return <>{children}</>;
|
||||
}
|
||||
@@ -64,6 +72,13 @@ export const TableNodeContextMenu: React.FC<
|
||||
<span>{t('table_node_context_menu.duplicate_table')}</span>
|
||||
<Copy className="size-3.5" />
|
||||
</ContextMenuItem>
|
||||
<ContextMenuItem
|
||||
onClick={addRelationshipHandler}
|
||||
className="flex justify-between gap-3"
|
||||
>
|
||||
<span>{t('table_node_context_menu.add_relationship')}</span>
|
||||
<Workflow className="size-3.5" />
|
||||
</ContextMenuItem>
|
||||
<ContextMenuItem
|
||||
onClick={removeTableHandler}
|
||||
className="flex justify-between gap-3"
|
||||
|
||||
@@ -5,11 +5,11 @@ import { Button } from '@/components/button/button';
|
||||
import {
|
||||
ChevronsLeftRight,
|
||||
ChevronsRightLeft,
|
||||
Pencil,
|
||||
Table2,
|
||||
ChevronDown,
|
||||
ChevronUp,
|
||||
Check,
|
||||
CircleDotDashed,
|
||||
} from 'lucide-react';
|
||||
import { Label } from '@/components/label/label';
|
||||
import type { DBTable } from '@/lib/domain/db-table';
|
||||
@@ -247,7 +247,7 @@ export const TableNode: React.FC<NodeProps<TableNodeType>> = React.memo(
|
||||
className="size-6 p-0 text-slate-500 hover:bg-primary-foreground hover:text-slate-700 dark:text-slate-400 dark:hover:bg-slate-800 dark:hover:text-slate-200"
|
||||
onClick={openTableInEditor}
|
||||
>
|
||||
<Pencil className="size-4" />
|
||||
<CircleDotDashed className="size-4" />
|
||||
</Button>
|
||||
)}
|
||||
{editMode ? null : (
|
||||
|
||||
@@ -38,6 +38,7 @@ import { Helmet } from 'react-helmet-async';
|
||||
import { useStorage } from '@/hooks/use-storage';
|
||||
import { AlertProvider } from '@/context/alert-context/alert-provider';
|
||||
import { CanvasProvider } from '@/context/canvas-context/canvas-provider';
|
||||
import { HIDE_BUCKLE_DOT_DEV } from '@/lib/env';
|
||||
|
||||
const OPEN_STAR_US_AFTER_SECONDS = 30;
|
||||
const SHOW_STAR_US_AGAIN_AFTER_DAYS = 1;
|
||||
@@ -153,6 +154,10 @@ const EditorPageComponent: React.FC = () => {
|
||||
]);
|
||||
|
||||
useEffect(() => {
|
||||
if (HIDE_BUCKLE_DOT_DEV) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (!currentDiagram?.id || githubRepoOpened) {
|
||||
return;
|
||||
}
|
||||
@@ -174,6 +179,10 @@ const EditorPageComponent: React.FC = () => {
|
||||
]);
|
||||
|
||||
useEffect(() => {
|
||||
if (HIDE_BUCKLE_DOT_DEV) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (!currentDiagram?.id) {
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -98,10 +98,16 @@ export const RelationshipListItemContent: React.FC<
|
||||
<Tooltip>
|
||||
<TooltipTrigger>
|
||||
<div className="truncate text-left text-sm">
|
||||
{sourceTable?.schema
|
||||
? `${sourceTable.schema}.`
|
||||
: ''}
|
||||
{sourceTable?.name}({sourceField?.name})
|
||||
</div>
|
||||
</TooltipTrigger>
|
||||
<TooltipContent>
|
||||
{sourceTable?.schema
|
||||
? `${sourceTable.schema}.`
|
||||
: ''}
|
||||
{sourceTable?.name}({sourceField?.name})
|
||||
</TooltipContent>
|
||||
</Tooltip>
|
||||
@@ -117,11 +123,17 @@ export const RelationshipListItemContent: React.FC<
|
||||
</div>
|
||||
<Tooltip>
|
||||
<TooltipTrigger>
|
||||
<div className="truncate text-left text-sm ">
|
||||
<div className="truncate text-left text-sm">
|
||||
{targetTable?.schema
|
||||
? `${targetTable.schema}.`
|
||||
: ''}
|
||||
{targetTable?.name}({targetField?.name})
|
||||
</div>
|
||||
</TooltipTrigger>
|
||||
<TooltipContent>
|
||||
{targetTable?.schema
|
||||
? `${targetTable.schema}.`
|
||||
: ''}
|
||||
{targetTable?.name}({targetField?.name})
|
||||
</TooltipContent>
|
||||
</Tooltip>
|
||||
|
||||
@@ -0,0 +1,18 @@
|
||||
import React from 'react';
|
||||
import { Toggle } from '@/components/toggle/toggle';
|
||||
|
||||
export const TableIndexToggle = React.forwardRef<
|
||||
React.ElementRef<typeof Toggle>,
|
||||
React.ComponentPropsWithoutRef<typeof Toggle>
|
||||
>((props, ref) => {
|
||||
return (
|
||||
<Toggle
|
||||
{...props}
|
||||
ref={ref}
|
||||
variant="default"
|
||||
className="h-8 w-[32px] p-2 text-xs text-slate-500 hover:bg-primary-foreground hover:text-slate-700 dark:text-slate-400 dark:hover:text-slate-200"
|
||||
/>
|
||||
);
|
||||
});
|
||||
|
||||
TableIndexToggle.displayName = Toggle.displayName;
|
||||
@@ -14,6 +14,12 @@ import { Label } from '@/components/label/label';
|
||||
import { Input } from '@/components/input/input';
|
||||
import { useTranslation } from 'react-i18next';
|
||||
import { SelectBox } from '@/components/select-box/select-box';
|
||||
import { TableIndexToggle } from './table-index-toggle';
|
||||
import {
|
||||
Tooltip,
|
||||
TooltipContent,
|
||||
TooltipTrigger,
|
||||
} from '@/components/tooltip/tooltip';
|
||||
|
||||
export interface TableIndexProps {
|
||||
index: DBIndex;
|
||||
@@ -54,7 +60,28 @@ export const TableIndex: React.FC<TableIndexProps> = ({
|
||||
)}
|
||||
keepOrder
|
||||
/>
|
||||
<div className="flex shrink-0">
|
||||
<div className="flex shrink-0 gap-1">
|
||||
<Tooltip>
|
||||
<TooltipTrigger asChild>
|
||||
<span>
|
||||
<TableIndexToggle
|
||||
pressed={index.unique}
|
||||
onPressedChange={(value) =>
|
||||
updateIndex({
|
||||
unique: !!value,
|
||||
})
|
||||
}
|
||||
>
|
||||
U
|
||||
</TableIndexToggle>
|
||||
</span>
|
||||
</TooltipTrigger>
|
||||
<TooltipContent>
|
||||
{t(
|
||||
'side_panel.tables_section.table.index_actions.unique'
|
||||
)}
|
||||
</TooltipContent>
|
||||
</Tooltip>
|
||||
<Popover>
|
||||
<PopoverTrigger asChild>
|
||||
<Button
|
||||
|
||||
@@ -162,7 +162,7 @@ export const TableListItemHeader: React.FC<TableListItemHeaderProps> = ({
|
||||
<EllipsisVertical />
|
||||
</ListItemHeaderButton>
|
||||
</DropdownMenuTrigger>
|
||||
<DropdownMenuContent className="w-fit">
|
||||
<DropdownMenuContent className="w-fit min-w-40">
|
||||
<DropdownMenuLabel>
|
||||
{t(
|
||||
'side_panel.tables_section.table.table_actions.title'
|
||||
|
||||
@@ -7,6 +7,7 @@ import { DiagramName } from './diagram-name';
|
||||
import { LastSaved } from './last-saved';
|
||||
import { LanguageNav } from './language-nav/language-nav';
|
||||
import { Menu } from './menu/menu';
|
||||
import { HIDE_BUCKLE_DOT_DEV } from '@/lib/env';
|
||||
|
||||
export interface TopNavbarProps {}
|
||||
|
||||
@@ -30,6 +31,10 @@ export const TopNavbar: React.FC<TopNavbarProps> = () => {
|
||||
}, []);
|
||||
|
||||
const renderGetBuckleButton = useCallback(() => {
|
||||
if (HIDE_BUCKLE_DOT_DEV) {
|
||||
return null;
|
||||
}
|
||||
|
||||
return (
|
||||
<button
|
||||
className="gradient-background relative inline-flex items-center justify-center overflow-hidden rounded-lg p-0.5 text-base text-gray-700 focus:outline-none focus:ring-0"
|
||||
|
||||
Reference in New Issue
Block a user