get rid of legacy commands

This commit is contained in:
Simone Scarduzio
2025-10-03 19:12:50 +02:00
parent 03106b76a8
commit b760890a61
7 changed files with 27 additions and 317 deletions

View File

@@ -1,21 +1,23 @@
# AWS S3 CLI Compatibility Plan for DeltaGlider
# AWS S3 CLI Compatibility for DeltaGlider
## Current State
DeltaGlider currently provides a custom CLI with the following commands:
DeltaGlider provides AWS S3 CLI compatible commands with automatic delta compression:
### Existing Commands
- `deltaglider put <file> <s3_url>` - Upload file with delta compression
- `deltaglider get <s3_url> [-o output]` - Download and reconstruct file
### Commands
- `deltaglider cp <source> <destination>` - Copy files with delta compression
- `deltaglider ls [s3_url]` - List buckets and objects
- `deltaglider rm <s3_url>` - Remove objects
- `deltaglider sync <source> <destination>` - Synchronize directories
- `deltaglider verify <s3_url>` - Verify file integrity
### Current Usage Examples
```bash
# Upload a file
deltaglider put myfile.zip s3://bucket/path/to/file.zip
deltaglider cp myfile.zip s3://bucket/path/to/file.zip
# Download a file (auto-detects .delta)
deltaglider get s3://bucket/path/to/file.zip
# Download a file
deltaglider cp s3://bucket/path/to/file.zip .
# Verify integrity
deltaglider verify s3://bucket/path/to/file.zip.delta
@@ -168,18 +170,7 @@ Additional flags specific to DeltaGlider's delta compression:
3. Create migration guide from aws-cli
4. Performance benchmarks comparing to aws-cli
## Migration Path for Existing Users
### Alias Support During Transition
```bash
# Old command -> New command mapping
deltaglider put FILE S3_URL -> deltaglider cp FILE S3_URL
deltaglider get S3_URL -> deltaglider cp S3_URL .
deltaglider verify S3_URL -> deltaglider ls --verify S3_URL
```
### Environment Variables
- `DELTAGLIDER_LEGACY_MODE=1` - Use old command syntax
## Environment Variables
- `DELTAGLIDER_AWS_COMPAT=1` - Strict AWS S3 CLI compatibility mode
## Success Criteria

View File

@@ -57,7 +57,7 @@ aws s3 cp readonlyrest-1.66.1_es8.0.0.zip s3://releases/
# Size on S3: 82.5MB
# With DeltaGlider
deltaglider put readonlyrest-1.66.1_es8.0.0.zip s3://releases/
deltaglider cp readonlyrest-1.66.1_es8.0.0.zip s3://releases/
# Size on S3: 65KB (99.92% smaller!)
```
@@ -186,7 +186,7 @@ This intelligence meant our 127,455 checksum files were uploaded directly, avoid
```bash
# Simple integration into our CI/CD
- aws s3 cp $FILE s3://releases/
+ deltaglider put $FILE s3://releases/
+ deltaglider cp $FILE s3://releases/
```
### Week 4: Full Migration
@@ -253,10 +253,10 @@ Storage costs scale linearly with data growth. Without DeltaGlider:
pip install deltaglider
# Upload a file (automatic compression)
deltaglider put my-release-v1.0.0.zip s3://releases/
deltaglider cp my-release-v1.0.0.zip s3://releases/
# Download (automatic reconstruction)
deltaglider get s3://releases/my-release-v1.0.0.zip
deltaglider cp s3://releases/my-release-v1.0.0.zip .
# It's that simple.
```
@@ -277,12 +277,12 @@ completely_different: 0% # No compression (uploaded as-is)
**GitHub Actions**:
```yaml
- name: Upload Release
run: deltaglider put dist/*.zip s3://releases/${{ github.ref_name }}/
run: deltaglider cp dist/*.zip s3://releases/${{ github.ref_name }}/
```
**Jenkins Pipeline**:
```groovy
sh "deltaglider put ${WORKSPACE}/target/*.jar s3://artifacts/"
sh "deltaglider cp ${WORKSPACE}/target/*.jar s3://artifacts/"
```
**Python Script**:
@@ -327,7 +327,7 @@ python calculate_savings.py --path /your/releases
# Try it yourself
docker run -p 9000:9000 minio/minio # Local S3
pip install deltaglider
deltaglider put your-file.zip s3://test/
deltaglider cp your-file.zip s3://test/
```
---