Connector Catalog
The Matih platform provides access to 600+ data source connectors through its integration with Airbyte. Connectors handle the complexities of authentication, pagination, rate limiting, schema discovery, incremental extraction, and error recovery for each source type.
Connector Categories
| Category | Count | Description |
|---|---|---|
| Databases | 30+ | Relational databases, NoSQL databases, data warehouses |
| SaaS | 200+ | CRM, marketing, finance, support, project management, analytics |
| Cloud Storage | 10+ | Object storage, SFTP, file-based sources |
| APIs | 100+ | REST, GraphQL, and webhook-based integrations |
| Event Streams | 10+ | Kafka, Kinesis, Pub/Sub, EventHub |
Connector Architecture
Every connector follows the same lifecycle within the Matih ingestion subsystem.
+-------------------+ +-----------------+ +------------------+
| Connector Config |---->| Airbyte Source |---->| Schema Discovery |
| (credentials, | | (connector | | (streams, |
| connection params)| | container) | | columns, types) |
+-------------------+ +-----------------+ +--------+---------+
|
+--------v---------+
| Stream Selection |
| (user picks |
| tables to sync) |
+--------+---------+
|
+--------v---------+
| Sync Execution |
| (extract, load |
| to Iceberg) |
+------------------+Connector Capabilities
Each connector declares its supported capabilities during schema discovery.
| Capability | Description |
|---|---|
| Full Refresh | Re-extracts all data on every sync. Supported by all connectors. |
| Incremental - Append | Extracts only new rows since the last sync using a cursor column. |
| Incremental - Deduped | Extracts new and updated rows, deduplicating by primary key. |
| CDC | Captures inserts, updates, and deletes from the database transaction log. Only available for databases with log-based replication (PostgreSQL WAL, MySQL binlog, MongoDB oplog). |
Connector Configuration Model
All connectors are configured through the CreateSourceRequest API.
{
"name": "my-postgres-source",
"description": "Production orders database",
"connectorType": "postgres",
"connectionConfig": {
"host": "orders-db.example.com",
"port": 5432,
"database": "orders",
"username": "readonly_user",
"password": "********",
"ssl_mode": "require",
"schemas": ["public"]
}
}The connectorType field identifies which Airbyte connector to use. The connectionConfig map contains connector-specific parameters.
Listing Available Connectors
The Ingestion Service provides an API endpoint to list all available connector types.
GET /api/v1/sources/connector-typesResponse:
[
"postgres",
"mysql",
"mongodb",
"mssql",
"oracle",
"salesforce",
"hubspot",
"stripe",
"s3",
"gcs",
"azure-blob-storage",
"sftp",
"github",
"jira",
"zendesk",
"google-analytics",
"snowflake",
"bigquery",
"redshift",
...
]Connector Support Tiers
| Tier | Definition | Examples |
|---|---|---|
| Generally Available | Fully tested, production-ready, supported by the platform team | PostgreSQL, MySQL, MongoDB, Salesforce, S3 |
| Beta | Functional but may have edge cases. Community-maintained. | Notion, Airtable, Asana |
| Alpha | Experimental. Use with caution. | Custom REST API connectors |
Common Configuration Patterns
Database Connectors
All database connectors share a common configuration structure:
| Field | Type | Description |
|---|---|---|
host | string | Database server hostname or IP address |
port | integer | Database server port |
database | string | Database name |
username | string | Authentication username |
password | string | Authentication password |
ssl_mode | string | SSL configuration (disable, require, verify-ca, verify-full) |
schemas | string[] | Schemas to include in discovery (optional, defaults to all) |
SaaS Connectors
SaaS connectors typically use OAuth2 or API key authentication:
| Field | Type | Description |
|---|---|---|
api_key | string | API key for key-based authentication |
client_id | string | OAuth2 client ID |
client_secret | string | OAuth2 client secret |
refresh_token | string | OAuth2 refresh token |
start_date | string | ISO 8601 date for initial data extraction window |
Cloud Storage Connectors
Cloud storage connectors read files from object stores:
| Field | Type | Description |
|---|---|---|
bucket | string | Bucket or container name |
path_prefix | string | Path prefix to filter objects |
file_format | string | Expected file format (csv, parquet, json, avro) |
access_key_id | string | Cloud provider access key |
secret_access_key | string | Cloud provider secret key |
region | string | Cloud provider region |
Next Steps
- Database Connectors -- detailed configuration for PostgreSQL, MySQL, MongoDB, and more
- SaaS Connectors -- detailed configuration for Salesforce, HubSpot, Stripe, and more
- Cloud Storage Connectors -- detailed configuration for S3, GCS, Azure Blob, and SFTP