9.7 KiB
DeltaGlider Python SDK Documentation
The DeltaGlider Python SDK provides a boto3-compatible API for core S3 operations (~20% of methods covering 80% of use cases), while achieving 99%+ compression for very similar versioned artifacts through intelligent binary delta compression.
🎯 Key Highlights
- boto3-Compatible Core API: 21 essential S3 methods that work exactly like boto3
- 99%+ Compression: Automatically for versioned files and archives
- Familiar API: If you know boto3, you already know DeltaGlider's core methods
- Full S3 Compatibility: Works with AWS S3, MinIO, Cloudflare R2, and all S3-compatible storage
- See BOTO3_COMPATIBILITY.md: For complete method coverage details
Quick Links
- Getting Started - Installation and first steps
- Examples - Real-world usage patterns
- API Reference - Complete API documentation
- Architecture - How it works under the hood
Overview
DeltaGlider provides three ways to interact with your S3 storage:
1. boto3-Compatible API (Recommended) 🌟
Core boto3 S3 methods with automatic compression (see BOTO3_COMPATIBILITY.md for full list):
from deltaglider import create_client
# Core boto3 S3 methods work exactly the same, with 99% compression!
client = create_client()
# Standard boto3 S3 methods - just work!
client.put_object(Bucket='releases', Key='v1.0.0/app.zip', Body=data)
response = client.get_object(Bucket='releases', Key='v1.0.0/app.zip')
# Optimized list_objects with smart performance defaults (NEW!)
# Fast by default - no unnecessary metadata fetching
response = client.list_objects(Bucket='releases', Prefix='v1.0.0/')
for obj in response['Contents']:
print(f"{obj['Key']}: {obj['Size']} bytes")
# Pagination for large buckets
response = client.list_objects(Bucket='releases', MaxKeys=100)
while response.get('IsTruncated'):
# Process current page
for obj in response['Contents']:
print(obj['Key'])
# Get next page
response = client.list_objects(
Bucket='releases',
MaxKeys=100,
ContinuationToken=response.get('NextContinuationToken')
)
# Get detailed compression stats only when needed
response = client.list_objects(Bucket='releases', FetchMetadata=True) # Slower but detailed
# Bucket statistics with intelligent S3-based caching (NEW!)
stats = client.get_bucket_stats('releases') # Fast (~100ms with cache)
stats = client.get_bucket_stats('releases', mode='detailed') # Accurate compression metrics
stats = client.get_bucket_stats('releases', refresh_cache=True) # Force fresh computation
client.delete_object(Bucket='releases', Key='old-version.zip')
2. Simple API
For straightforward use cases:
from deltaglider import create_client
client = create_client()
summary = client.upload("my-app-v1.0.0.zip", "s3://releases/v1.0.0/")
client.download("s3://releases/v1.0.0/my-app-v1.0.0.zip", "local.zip")
3. CLI (Command Line Interface)
Drop-in replacement for AWS S3 CLI:
deltaglider cp my-app-v1.0.0.zip s3://releases/
deltaglider ls s3://releases/
deltaglider sync ./builds/ s3://releases/
Migration from boto3
For core S3 operations, migrating is as simple as changing your import:
# Before (boto3)
import boto3
client = boto3.client('s3')
client.put_object(Bucket='mybucket', Key='myfile.zip', Body=data)
# After (DeltaGlider) - Core methods work the same, with 99% compression!
from deltaglider import create_client
client = create_client()
client.put_object(Bucket='mybucket', Key='myfile.zip', Body=data)
Note: DeltaGlider implements ~21 core S3 methods. For advanced features (versioning, ACLs, multipart uploads >5GB), use boto3 directly. See BOTO3_COMPATIBILITY.md for details.
Key Features
- Core boto3 Compatibility: 21 essential S3 methods work exactly as expected (~20% coverage, 80% use cases)
- 99%+ Compression: For versioned artifacts and similar files
- Intelligent Detection: Automatically determines when to use delta compression
- Data Integrity: SHA256 verification on every operation
- Transparent: Works with existing tools and workflows
- Production Ready: Battle-tested with 200K+ files
- Thoroughly Tested: 99 integration/unit tests with comprehensive coverage
- Type Safe: Full mypy type checking, zero type errors
When to Use DeltaGlider
Perfect For
- Software releases and versioned artifacts
- Container images and layers
- Database backups and snapshots
- Machine learning model checkpoints
- Game assets and updates
- Any versioned binary data
Not Ideal For
- Already compressed unique files
- Streaming media files
- Frequently changing unstructured data
- Files smaller than 1MB
Installation
pip install deltaglider
For development or testing with MinIO:
docker run -p 9000:9000 minio/minio server /data
export AWS_ENDPOINT_URL=http://localhost:9000
Basic Usage
boto3-Compatible Usage (Recommended)
from deltaglider import create_client
# Create client (uses AWS credentials automatically)
client = create_client()
# Upload using boto3 API
with open('release-v2.0.0.zip', 'rb') as f:
response = client.put_object(
Bucket='releases',
Key='v2.0.0/release.zip',
Body=f,
Metadata={'version': '2.0.0'}
)
# Check compression stats (DeltaGlider extension)
if 'DeltaGliderInfo' in response:
info = response['DeltaGliderInfo']
print(f"Saved {info['SavingsPercent']:.0f}% storage space")
# Download using boto3 API
response = client.get_object(Bucket='releases', Key='v2.0.0/release.zip')
with open('local-copy.zip', 'wb') as f:
f.write(response['Body'].read())
# List objects
response = client.list_objects(Bucket='releases', Prefix='v2.0.0/')
for obj in response.get('Contents', []):
print(f"{obj['Key']}: {obj['Size']} bytes")
# Delete object
client.delete_object(Bucket='releases', Key='old-version.zip')
Simple API Usage
from deltaglider import create_client
# Create client (uses AWS credentials from environment)
client = create_client()
# Upload a file
summary = client.upload("release-v2.0.0.zip", "s3://releases/v2.0.0/")
print(f"Saved {summary.savings_percent:.0f}% storage space")
# Download a file
client.download("s3://releases/v2.0.0/release-v2.0.0.zip", "local-copy.zip")
With Custom Configuration
from deltaglider import create_client
client = create_client(
endpoint_url="http://minio.internal:9000", # Custom S3 endpoint
log_level="DEBUG", # Detailed logging
aws_access_key_id="minio",
aws_secret_access_key="minio",
region_name="eu-west-1",
max_ratio=0.3, # Stricter delta acceptance
)
ℹ️ The SDK now manages an encrypted, process-isolated cache automatically in
/tmp/deltaglider-*. Tune cache behavior via environment variables such asDG_CACHE_BACKEND,DG_CACHE_MEMORY_SIZE_MB, andDG_CACHE_ENCRYPTION_KEYinstead of passing acache_dirargument.
Real-World Example
from deltaglider import create_client
# Core boto3 methods work exactly like boto3!
client = create_client()
# Upload multiple software versions
versions = ["v1.0.0", "v1.0.1", "v1.0.2", "v1.1.0"]
for version in versions:
with open(f"dist/my-app-{version}.zip", 'rb') as f:
response = client.put_object(
Bucket='releases',
Key=f'{version}/my-app.zip',
Body=f
)
# DeltaGlider provides compression stats
if 'DeltaGliderInfo' in response:
info = response['DeltaGliderInfo']
print(f"{version}: {info['StoredSizeMB']:.1f}MB "
f"(saved {info['SavingsPercent']:.0f}%)")
# Result:
# v1.0.0: 100.0MB (saved 0%) <- First file becomes reference
# v1.0.1: 0.2MB (saved 99.8%) <- Only differences stored
# v1.0.2: 0.3MB (saved 99.7%) <- Delta from reference
# v1.1.0: 5.2MB (saved 94.8%) <- Larger changes, still huge savings
How It Works
- First Upload: The first file uploaded to a prefix becomes the reference
- Delta Compression: Subsequent similar files are compared using xdelta3
- Smart Storage: Only the differences (deltas) are stored
- Transparent Reconstruction: Files are automatically reconstructed on download
- Core boto3 Compatibility: Essential operations maintain full boto3 API compatibility
Performance
Based on real-world usage:
- Compression: 99%+ for similar versions
- Upload Speed: 3-4 files/second
- Download Speed: <100ms reconstruction
- Storage Savings: 4TB → 5GB (ReadOnlyREST case study)
Advanced Features
Multipart Upload Support
# Large file uploads work automatically
with open('large-file.zip', 'rb') as f:
client.put_object(
Bucket='backups',
Key='database/backup.zip',
Body=f # Handles multipart automatically for large files
)
Batch Operations
# Upload multiple files efficiently
files = ['app.zip', 'docs.zip', 'assets.zip']
for file in files:
with open(file, 'rb') as f:
client.put_object(Bucket='releases', Key=file, Body=f)
Presigned URLs
# Generate presigned URLs for secure sharing
url = client.generate_presigned_url(
'get_object',
Params={'Bucket': 'releases', 'Key': 'v1.0.0/app.zip'},
ExpiresIn=3600
)
Support
- GitHub Issues: github.com/beshu-tech/deltaglider/issues
- Documentation: github.com/beshu-tech/deltaglider#readme
License
MIT License - See LICENSE for details.