AQEA: aqea-all-MiniLM-L6-v2-29x
MiniLM text embeddings compressed 29x while preserving 97.1% similarity ranking
π Performance
| Metric | Value |
|---|---|
| Compression Ratio | 29.5x |
| Spearman Ο | 97.1% |
| Source Dimension | 384D |
| Compressed Dimension | 13D |
| Storage Savings | 96.6% |
π Usage
from aqea import AQEACompressor
# Load pre-trained compressor
compressor = AQEACompressor.from_pretrained("nextxag/aqea-all-MiniLM-L6-v2-29x")
# Compress embeddings
embeddings = model.encode(texts) # 384D
compressed = compressor.compress(embeddings) # 13D
# Decompress for retrieval
reconstructed = compressor.decompress(compressed) # 384D
π Files
weights.aqwt- Binary weights (AQEA native format)config.json- Model configuration
π¬ How It Works
AQEA (Adaptive Quantized Embedding Architecture) uses learned linear projections with Pre-Quantify rotation to compress embeddings while maximally preserving pairwise similarity rankings (measured by Spearman correlation).
π Citation
@software{aqea2024,
title = {AQEA: Adaptive Quantized Embedding Architecture},
author = {AQEA Team},
year = {2024},
url = {https://huggingface.co/nextxag}
}
π License
Apache 2.0
- Downloads last month
- 10
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for nextxag/aqea-all-MiniLM-L6-v2-29x
Base model
sentence-transformers/all-MiniLM-L6-v2