Darkseal
GuinsooLabGitHub
  • Welcome to Darkseal!
  • Overview
    • Tutorial
    • Features
      • Basic Features
      • Discovery and Collaboration
      • Glossaries
      • Data Insights
      • Integrations
      • Alerts and Notifications
    • Releases
    • Roadmap
  • Deployment
    • Docker Deployment
    • Bare Metal Deployment
    • Enable Security
      • Basic Authentication
      • Google SSO
      • Okta SSO
      • One Login SSO
      • Keycloak SSO
      • MorganaToken SSO
      • JWT Token
    • Server Configuration
    • Airflow
  • Connectors
    • Database
      • Clickhouse
      • Databricks
      • DeltaLake
      • Druid
      • Hive
      • MariaDB
      • MySQL
      • Pinot
      • Postgres
      • Presto
      • Trino
    • Dashboard
      • Metabase
      • Redash
      • Spotrix
      • Superset
    • Messaging
      • Kafka
      • Redpanda
    • Pipeline
      • Airflow
      • Airbyte
      • Dagster
    • ML Model
      • ML Flow
    • Metadata
      • Amundsen
      • Datahub
    • Ingestion
      • Workflows
      • Lineage
      • Versioning
  • Main Concepts
    • High Level Design
    • Metadata Standard
      • Schemas
  • Developers
    • Architecture
    • Contribute
    • Web Hooks
  • Appendix
    • FAQs
    • APIs
Powered by GitBook
On this page
  1. Deployment

Airflow

PreviousServer ConfigurationNextDatabase

Last updated 2 years ago

If you already have an Airflow instance up and running, you might want to reuse it to host the metadata workflows as well. Here we will guide you on the different aspects to consider when configuring an existing Airflow.

There are three different angles here:

  1. Installing the ingestion modules directly on the host to enable the .

  2. Installing connector modules on the host to run specific workflows.

  3. Installing the Airflow APIs to enable the workflow deployment through the UI.

Depending on what you wish to use, you might just need some of these installations. Note that the installation commands shown below need to be run in the Airflow instances.

Airflow Lineage Backend