Schema-Driven DynamoDB Normalization & Event Publishing for Python
daplug-ddb is a lightweight package that provides schema-aware CRUD helpers, batch utilities, and optional SNS publishing so you can treat DynamoDB as a structured datastore without rewriting boilerplate for every project.
- Schema Mapping β Convert inbound payloads into strongly typed DynamoDB items driven by your OpenAPI (or JSON schema) definitions.
- Idempotent CRUD β Consistent
create,overwrite,update,delete, andreadoperations with optional optimistic locking via anidempotence_key. - Batch Helpers β Simplified batch insert/delete flows that validate data and handle chunking for you.
- SNS Integration β Optional event publishing for every write operation so downstream systems stay in sync.
pip install daplug-ddb
# pipenv install daplug-ddb
# poetry add daplug-ddb
# uv pip install daplug-ddbimport daplug_ddb
adapter = daplug_ddb.adapter(
table="example-table",
endpoint="https://dynamodb.us-east-2.amazonaws.com", # optional, will use AWS conventional env vars if using on lambda
schema_file="openapi.yml",
hash_key="record_id",
idempotence_key="modified",
)
item = adapter.create(
data={
"record_id": "abc123",
"object_key": {"string_key": "value"},
"array_number": [1, 2, 3],
"modified": "2024-01-01",
},
schema="ExampleModel",
)
print(item)Because the adapter is configured with a schema_file, every call can opt into
mapping by supplying schema. Skip the schema argument when you want to write
the data exactly as provided.
# Merge partial updates while preserving existing attributes
adapter.update(
operation="get", # fetch original item via get; use "query" for indexes
query={
"Key": {"record_id": "abc123", "sort_key": "v1"}
},
data={
"record_id": "abc123",
"sort_key": "v1",
"array_number": [1, 2, 3, 4],
},
update_list_operation="replace",
)adapter = daplug_ddb.adapter(
table="tenant-config",
endpoint="https://dynamodb.us-east-2.amazonaws.com",
schema_file="openapi.yml",
hash_key="tenant_id",
)
prefix_args = {
"hash_key": "tenant_id",
"hash_prefix": "tenant#",
"range_key": "sort_key",
"range_prefix": "config#",
}
item = adapter.create(
data={
"tenant_id": "abc",
"sort_key": "default",
"modified": "2024-01-01",
},
schema="TenantModel",
**prefix_args,
)
# DynamoDB stores tenant_id as "tenant#abc", but the adapter returns "abc"When prefixes are provided, the adapter automatically applies them on the way
into DynamoDB (including batch operations and deletes) and removes them before
returning data or publishing SNS events. Pass the same prefix_args to reads
(get, query, scan) so query keys are expanded and responses are cleaned.
Codex workflow references:
adapter.batch_insert(
data=[
{"record_id": str(idx), "sort_key": str(idx)}
for idx in range(100)
],
batch_size=25,
)
adapter.batch_delete(
data=[
{"record_id": str(idx), "sort_key": str(idx)}
for idx in range(100)
]
)adapter = daplug_ddb.adapter(
table="orders",
endpoint="https://dynamodb.us-east-2.amazonaws.com",
schema_file="openapi.yml",
hash_key="order_id",
idempotence_key="modified",
)
updated = adapter.update(
data={"order_id": "abc123", "modified": "2024-02-01"},
operation="get",
query={"Key": {"order_id": "abc123"}},
schema="OrderModel",
)The adapter fetches the current item, merges the update, and executes a
conditional PutItem to ensure the stored modified value still matches
what was read. If another writer changes the record first, the operation
fails with a conditional check error rather than overwriting the data.
Set raise_idempotence_error=True if you prefer the adapter to raise a
ValueError instead of relying on DynamoDB's conditional failure. Leaving it
at the default (False) allows you to detect conflicts without breaking the
update flow.
adapter = daplug_ddb.adapter(
table="orders",
schema_file="openapi.yml",
hash_key="order_id",
idempotence_key="modified",
raise_idempotence_error=True,
)Enable idempotence_use_latest=True when you want the adapter to keep the
most recent copy based on the timestamp stored in the idempotence key. Stale
updates are ignored automatically.
adapter = daplug_ddb.adapter(
table="orders",
schema_file="openapi.yml",
hash_key="order_id",
idempotence_key="modified",
idempotence_use_latest=True,
)Stale updates are short-circuited before DynamoDB writes occur.
Client Update Request
β
βΌ
[Adapter.fetch]
β (reads original item)
βΌ
ββββββββββββββββββββββββββββ
β Original Item β
β modified = "2024-01-01" β
ββββββββββββββββββββββββββββ
β merge + map
βΌ
PutItem rejected β original returnedClient Update Request
β
βΌ
[Adapter.fetch]
β (reads original item)
βΌ
ββββββββββββββββββββββββββββ
β Original Item β
β idempotence_key = "v1" β
ββββββββββββββββββββββββββββ
β merge + map
βΌ
PutItem(Item=β¦, ConditionExpression=Attr(idempotence_key).eq("v1"))
β
ββββββ΄ββββββββ
β β
βΌ βΌ
Success ConditionalCheckFailed
(another writer changed key)You can supply request-scoped SNS message attributes by passing 'sns_attributes' into any adapter operation (e.g. 'create', 'update', 'delete'). These merge with adapter defaults and schema-derived metadata.
adapter = daplug_ddb.adapter(
table="audit-table",
schema_file="openapi.yml",
hash_key="audit_id",
idempotence_key="version",
sns_arn="arn:aws:sns:us-east-2:123456789012:audit-events",
sns_endpoint="https://sns.us-east-2.amazonaws.com",
sns_attributes={"source": "daplug"},
)
adapter.create(
data=item,
schema="AuditModel",
sns_attributes={"source": "billing", "priority": "high"},
)
# => publishes a formatted SNS event with schema metadataAdapter-level 'sns_attributes' supplied when constructing the adapter act as defaults for every publish. Use per-call 'sns_attributes' to extend or override those defaults without touching the adapter configuration. Each publish always adds an 'operation' attribute reflecting the CRUD action so subscribers can route by verb.
adapter = daplug_ddb.adapter(
table="audit-table",
schema_file="openapi.yml",
hash_key="audit_id",
sns_arn="arn:aws:sns:...",
sns_attributes={"source": "daplug", "env": "prod"},
)
# emits {source: "daplug", env: "prod", operation: "create"}
adapter.create(data=item, schema="AuditModel")
# overrides only the env attribute for this publish
adapter.update(
data=item,
schema="AuditModel",
sns_attributes={"env": "staging"},
)Each adapter instance holds shared configuration such as schema_file, SNS
defaults, and optional key prefixes. Pass the schema name (and any
operation-specific overrides) when you invoke a method.
adapter = daplug_ddb.adapter(
table="orders",
schema_file="openapi.yml",
hash_key="order_id",
idempotence_key="modified",
)# default: behaves like insert (requires hash_key)
adapter.create(data=payload, schema="OrderModel")
# explicit overwrite (upsert semantics)
adapter.create(
operation="overwrite",
data=payload,
schema="OrderModel",
)adapter.insert(data=payload, schema="OrderModel")adapter.overwrite(data=payload, schema="OrderModel")adapter.get(
query={"Key": {"order_id": "abc123"}},
schema="OrderModel",
)adapter.query(
query={
"IndexName": "test_query_id",
"KeyConditionExpression": "test_query_id = :id",
"ExpressionAttributeValues": {":id": "def345"},
},
schema="OrderModel",
)adapter.scan(schema="OrderModel")
# raw DynamoDB response
adapter.scan(raw_scan=True)read delegates to get, query, or scan based on the
operation kwarg.
# single item
adapter.read(operation="get", query={"Key": {"order_id": "abc123"}}, schema="OrderModel")
# query
adapter.read(
operation="query",
query={
"KeyConditionExpression": "test_query_id = :id",
"ExpressionAttributeValues": {":id": "def345"},
},
schema="OrderModel",
)adapter.update(
data={"order_id": "abc123", "modified": "2024-03-02"},
operation="get",
query={"Key": {"order_id": "abc123"}},
schema="OrderModel",
)adapter.delete(query={"Key": {"order_id": "abc123"}})adapter.batch_insert(data=[{...} for _ in range(10)], schema="OrderModel", batch_size=25)adapter.batch_delete(data=[{...} for _ in range(10)], batch_size=25)Include per-call prefix overrides whenever you need to scope keys.
adapter.insert(
data=payload,
schema="OrderModel",
hash_key="order_id",
hash_prefix="tenant#",
)- Python 3.10+
- Pipenv
- Docker (for running DynamoDB Local during tests)
git clone https://github.com/paulcruse3/daplug-ddb.git
cd daplug-ddb
pipenv install --devKeep setup.py aligned with the locked Pipenv dependencies before publishing.
pipenv run pipenv-setup sync# unit tests (no DynamoDB required)
pipenv run test
# integration tests (spins up local DynamoDB when available)
pipenv run integrationsSupplying an idempotence_key enables optimistic concurrency for updates and overwrites. The adapter reads the original item, captures the keyβs value, and issues a PutItem with a ConditionExpression asserting the value is unchanged. If another writer updates the record first, DynamoDB returns a conditional check failure instead of silently overwriting data.
Client Update Request
β
βΌ
[Adapter.fetch]
β (reads original item)
βΌ
ββββββββββββββββββββββββββββ
β Original Item β
β idempotence_key = "v1" β
ββββββββββββββββββββββββββββ
β merge + map
βΌ
PutItem(Item=β¦, ConditionExpression=Attr(idempotence_key).eq("v1"))
β
ββββββ΄ββββββββ
β β
βΌ βΌ
Success ConditionalCheckFailed
(another writer changed key)- Optional: Omit
idempotence_keyto mirror DynamoDBβs default βlast write winsβ behavior while still benefiting from schema normalization. - Safety: When the key is configured but missing on the fetched item, the adapter raises
ValueError, surfacing misconfigurations early. - Events: SNS notifications include the idempotence metadata so downstream services can reason about version changes.
# generates HTML, XML, and JUnit reports under ./coverage/
pipenv run coverage
# pylint configuration aligned with the legacy project
pipenv run lintdaplug-ddb/
βββ daplug_ddb/
βΒ Β βββ adapter.py # DynamoDB adapter implementation
βΒ Β βββ prefixer.py # DynamoDB prefixer implementation
βΒ Β βββ common/ # Shared helpers (merging, schema loading, logging)
βΒ Β βββ __init__.py # Public adapter factory & exports
βββ tests/
βΒ Β βββ integration/ # Integration suite against DynamoDB Local
βΒ Β βββ unit/ # Isolated unit tests using mocks
βΒ Β βββ openapi.yml # Sample schema used for mapping tests
βββ Pipfile # Runtime and dev dependencies
βββ setup.py # Packaging metadata
βββ README.mdContributions are welcome! Open an issue or submit a pull request if youβd like to add new features, improve documentation, or expand test coverage.
git checkout -b feature/amazing-improvement
# make your changes
pipenv run lint
pipenv run test
pipenv run integrations
git commit -am "feat: amazing improvement"
git push origin feature/amazing-improvementApache License 2.0 β see LICENSE for full text.
Built to keep DynamoDB integrations DRY, predictable, and schema-driven.