Satya (ΰ€Έΰ€€ΰ₯ΰ€―) is the Sanskrit word for truth and reality, embodying our commitment to data integrity and validation. Just as truth is fundamental and unwavering, Satya ensures your data validation is reliable, fast, and efficient.
Satya is a blazingly fast data validation library for Python, powered by Rust. It provides comprehensive validation capabilities while maintaining exceptional performance through innovative batch processing techniques.
β οΈ Latest Version: v0.3.8 - Upgrading from v0.2? Read the migration guide: docs/migration.md. v0.3 introduces a Pydantic-like DX with breaking changes.
- Dict[str, CustomModel] Support: Complete validation support for dictionary structures containing custom model instances
- MAP-Elites Algorithm Support: Native support for complex archive structures like
Dict[str, ArchiveEntry]
- Hierarchical Data Structures: Full support for nested model dictionaries in configuration management and ML pipelines
- Recursive Model Resolution: Automatic dependency analysis and topological sorting for proper validation order
- Dependency Tracking: Automatically analyzes and tracks model relationships
- Topological Sorting: Ensures models are validated in the correct dependency order
- Circular Dependency Detection: Prevents infinite loops in complex model graphs
- SDist Builds: Proper source distribution builds enabling
--no-binary
installations - Docker Run CI/CD: Improved GitHub Actions compatibility with direct docker run commands
- Cross-Platform Compatibility: Full support for Linux, macOS, and Windows across all architectures
from satya import Model, Field
from typing import Dict
class ArchiveEntry(Model):
config: SystemConfig
performance: float = Field(ge=-1000.0, le=100000.0)
class MapElitesArchive(Model):
resolution: int = Field(ge=1, le=20)
archive: Dict[str, ArchiveEntry] = Field(description="Archive entries")
# This now works perfectly!
data = {
"resolution": 5,
"archive": {
"cell_1_2": {"config": {"buffer_size": 1024}, "performance": 95.5}
}
}
archive = MapElitesArchive(**data) # Works perfectly!
- Added complete test suite with 4 test methods covering nested Dict[str, Model] patterns
- All 150+ tests pass with comprehensive coverage
- Source distribution builds tested and verified
- High-performance validation with Rust-powered core
- Batch processing with configurable batch sizes for optimal throughput
- Stream processing support for handling large datasets
- Comprehensive validation including email, URL, regex, numeric ranges, and more
- Type coercion with intelligent type conversion
- Decimal support for financial-grade precision
- Compatible with standard Python type hints
- OpenAI-compatible schema generation
- Minimal memory overhead
from satya import Model, Field, ModelValidationError
class User(Model):
id: int = Field(description="User ID")
name: str = Field(description="User name")
email: str = Field(description="Email address")
active: bool = Field(default=True)
# Enable batching for optimal performance
validator = User.validator()
validator.set_batch_size(1000) # Recommended for most workloads
# Process data efficiently
for valid_item in validator.validate_stream(data):
process(valid_item)
from typing import Optional
from decimal import Decimal
from satya import Model, Field, List
# Pretty printing (optional)
Model.PRETTY_REPR = True
class User(Model):
id: int
name: str = Field(default='John Doe')
email: str = Field(email=True) # RFC 5322 compliant email validation
signup_ts: Optional[str] = Field(required=False)
friends: List[int] = Field(default=[])
balance: Decimal = Field(ge=0, description="Account balance") # Decimal support
external_data = {
'id': '123',
'email': '[email protected]',
'signup_ts': '2017-06-01 12:22',
'friends': [1, '2', b'3'],
'balance': '1234.56'
}
validator = User.validator()
validator.set_batch_size(1000) # Enable batching for performance
result = validator.validate(external_data)
user = User(**result.value)
print(user)
#> User(id=123, name='John Doe', email='[email protected]', signup_ts='2017-06-01 12:22', friends=[1, 2, 3], balance=1234.56)
Our comprehensive benchmarks demonstrate Satya's exceptional performance when using batch processing:
- Satya (batch=1000): 2,072,070 items/second
- msgspec: 1,930,466 items/second
- Satya (single-item): 637,362 items/second
Key findings:
- Batch processing provides up to 3.3x performance improvement
- Optimal batch size of 1,000 items for complex validation workloads
- Competitive performance with msgspec while providing comprehensive validation
Memory usage remains comparable across all approaches, demonstrating that performance gains don't come at the cost of increased memory consumption.
Our earlier benchmarks also show significant performance improvements:
- Satya: 207,321 items/second
- Pydantic: 72,302 items/second
- Speed improvement: 2.9x
- Memory usage: Nearly identical (Satya: 158.2MB, Pydantic: 162.5MB)
- Satya: 177,790 requests/second
- Pydantic: 1,323 requests/second
- Average latency improvement: 134.4x
- P99 latency improvement: 134.4x
Validation Mode | Throughput | Memory Usage | Use Case |
---|---|---|---|
Satya dict-path | 5.7M items/s | 7.2MB | Pre-parsed Python dicts |
Satya JSON streaming | 3.2M items/s | 0.4MB | Large JSON datasets |
Satya JSON non-stream | 1.2M items/s | 0.4MB | Small JSON datasets |
orjson + Satya dict | 2.6M items/s | 21.5MB | End-to-end JSON processing |
msgspec + JSON | 7.5M items/s | 0.4MB | Comparison baseline |
Pydantic + orjson | 0.8M items/s | 0.4MB | Traditional validation |
- 7.9x faster than Pydantic for dict validation
- 4x faster than Pydantic for JSON processing
- Memory bounded: <8MB even for 5M records
- Competitive with msgspec: 76% of msgspec's speed with more flexibility
- Streaming support: Process unlimited datasets with constant memory
- Small Scale (100k): 7.9M items/s - matches msgspec performance
- Large Scale (5M): 5.7M items/s - maintains high throughput
- Memory Efficiency: Bounded growth, predictable resource usage
Note: Benchmarks run on Apple Silicon M-series. Results include comprehensive comparison with msgspec and Pydantic using fair JSON parsing (orjson). See
/benchmarks/
for detailed methodology.
- High Performance: Rust-powered core with efficient batch processing
- Comprehensive Validation:
- Email validation (RFC 5322 compliant)
- URL format validation
- Regex pattern matching
- Numeric constraints (min/max, ge/le/gt/lt)
- Decimal precision handling
- UUID format validation
- Enum and literal type support
- Array constraints (min/max items, unique items)
- Deep nested object validation
- Stream Processing: Efficient handling of large datasets
- Type Safety: Full compatibility with Python type hints
- Error Reporting: Detailed validation error messages
- Memory Efficient: Minimal overhead design
Satya brings together high performance and comprehensive validation capabilities. While inspired by projects like Pydantic (for its elegant API) and msgspec (for performance benchmarks), Satya offers:
- Rust-powered performance with zero-cost abstractions
- Batch processing for optimal throughput
- Comprehensive validation beyond basic type checking
- Production-ready error handling and reporting
- Memory-efficient design for large-scale applications
- High-throughput API services
- Real-time data processing pipelines
- Large dataset validation
- Stream processing applications
- Financial and healthcare systems requiring strict validation
- Performance-critical microservices
pip install satya
- Python 3.8 or higher
Note for developers: If you're contributing to Satya or building from source, you'll need Rust toolchain 1.70.0 or higher:
# Install Rust if you don't have it curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh # Update existing Rust installation rustup updateYou can check your Rust version with:
rustc --version
For optimal performance, always use batch processing:
# Configure batch size based on your workload
validator = MyModel.validator()
validator.set_batch_size(1000) # Start with 1000, adjust as needed
# Use stream processing for large datasets
for valid_item in validator.validate_stream(data):
process(valid_item)
- Default recommendation: 1,000 items
- Large objects: Consider smaller batches (500-1000)
- Small objects: Can use larger batches (5000-10000)
- Memory constrained: Use smaller batches
- Always benchmark with your specific data
Satya provides comprehensive validation that goes beyond basic type checking:
Feature | Satya | msgspec | Pydantic |
---|---|---|---|
Basic type validation | β | β | β |
Email validation (RFC 5322) | β | β | β |
URL validation | β | β | β |
Regex patterns | β | β | β |
Numeric constraints | β | β | β |
Decimal precision | β | β | β |
UUID validation | β | β | β |
Enum/Literal types | β | β | β |
Array constraints | β | β | β |
Deep nesting (4+ levels) | β | β | β |
Custom error messages | β | Limited | β |
Batch processing | β | β | β |
Satya provides comprehensive JSON Schema generation with OpenAI compatibility:
from satya import Model, Field
class User(Model):
name: str = Field(description="User name")
age: int = Field(description="User age")
# Standard JSON Schema
schema = User.json_schema()
print(schema)
# {
# "type": "object",
# "title": "User",
# "properties": {
# "name": {"type": "string", "description": "User name"},
# "age": {"type": "integer", "description": "User age"}
# },
# "required": ["name", "age"]
# }
# OpenAI-compatible schema (flattened types, strict validation)
openai_schema = User.model_json_schema()
# Fixes nested type objects and ensures OpenAI API compatibility
If you previously used the low-level core (_satya.StreamValidatorCore
) or manually registered schemas with StreamValidator
, migrate to the new model-first API. See the full guide: docs/migration.md
.
Quick before/after:
# Before (legacy manual schema)
from satya._satya import StreamValidatorCore
core = StreamValidatorCore()
core.add_field('id', 'int', True)
core.add_field('email', 'str', True)
core.set_field_constraints('email', email=True)
oks = core.validate_batch([{"id": 1, "email": "[email protected]"}])
# After (model-first)
from satya import Model, Field
class User(Model):
id: int
email: str = Field(email=True)
oks = User.validator().validate_batch([{"id": 1, "email": "[email protected]"}])
JSON bytes helpers (streaming):
ok = User.model_validate_json_bytes(b'{"id":1, "email":"[email protected]"}', streaming=True)
oks = User.model_validate_json_array_bytes(b'[{"id":1},{"id":2}]', streaming=True)
Satya v0.3.8 is stable and production-ready. The core functionality includes comprehensive validation, schema generation, enhanced nested model support, and source distribution builds. Key capabilities include:
- Complete Dict[str, CustomModel] Support: Full validation for complex nested structures
- MAP-Elites Algorithm Compatibility: Native support for evolutionary optimization archives
- Hierarchical Data Validation: Recursive model resolution with dependency tracking
- Source Distribution Support: Enable
uv pip install --no-binary satya satya==0.3.8
- Provider-Agnostic Architecture: Clean separation of core validation from provider-specific features
We're actively working on:
- Expanding type support
- Adding more validation features
- Improving error messages
- Enhancing documentation
- Performance optimizations
- Auto-optimization features
- Pydantic project for setting the standard in Python data validation and inspiring our API design
- msgspec project for demonstrating high-performance validation is achievable
- Rust community for providing the foundation for our performance
Note to Data Validation Library Authors: Feel free to incorporate our performance optimizations into your libraries! We believe in making the Python ecosystem faster for everyone. All we ask is for appropriate attribution to Satya under our Apache 2.0 license. Together, we can make data validation blazingly fast for all Python developers!
We welcome contributions of all kinds! Whether you're fixing bugs, improving documentation, or sharing new performance optimizations, here's how you can help:
- π Report issues and bugs
- π‘ Suggest new features or optimizations
- π Improve documentation
- π§ Submit pull requests
- π Share benchmarks and use cases
Check out our CONTRIBUTING.md for guidelines.
Apache 2.0
Note: Performance numbers are from comprehensive benchmarks and may vary based on use case and data structure complexity.
- GitHub Issues: Satya Issues
- Author: Rach Pradhan