I benchmarked computing sentence embeddings with e5.py script
small model = intfloat/e5-small-v2, result:

larger model=BAAI/bge-large-zh-v1.5, result:

Pure rust BERT example is still slower than huggingface transformer with BAAI/bge-large-zh-v1.5 model