In the previous update,
we mentioned HTTP/2, ONNX Runtime and factory.vespa.oath.cloud.
This month, we’re excited to share the following updates:
Vespa CLI is a zero-dependency tool built with Go, available for Linux, macOS and Windows –
it greatly simplifies interaction with a Vespa instance. Use Vespa CLI to:
- Clone Vespa sample applications
- Deploy an application to a Vespa installation running locally or remote
- Deploy an application to a dev zone in Vespa Cloud
- Feed and query documents
- Send custom requests with automatic authentication
Nearest neighbor search performance improvement
Exact nearest neighbor search without HNSW index improved serving performance by 20x in Vespa 7.457.52
when combined with query filters and using multiple threads per query.
HNSW index itself has a reduced memory footprint, too –
this enables applications to fit larger data sets for nearest neighbor use cases.
Reindexing an HNSW-index
is multithreaded since Vespa 7.436.31.
This makes it much faster to apply e.g. changes in what distance function is used depending on available CPU cores –
and is now as fast as regular multi-threaded updates to HNSW index.
Paged tensor attributes
Fields indexed as attributes are stored in memory which enables fast partial updates,
flexible match modes, grouping, range queries, sorting, parent/child imports, and direct use in ranking.
Dense tensor attributes can now be set as paged, which means they will mostly reside on disk rather than in memory.
This is useful for large tensors where fast access over many documents per query is not required.
Vespa now supports mTLS across all internal services and endpoints.
See the blog post
for an introduction and the reference documentation for setup instructions.
Many applications use Vespa for real-time partial update
rates in the 1000s per node per second.
Since Vespa 7.468.9, the Vespa Distributor uses multiple threads by default
with each thread handling a distinct set of the document bucket space.
Context switching is reduced by using async operations in the network threads.
The end-to-end feed throughput have increased significantly:
- 25-40% increase in throughput for partial updates
- 25% increase in throughput for puts of summary-only data
A common task in modern IR is to embed a document or query in a vector space for retrieval and/or ranking,
which often means turning a natural language text into a tensor.
Since 7.474.25, Vespa ships with a native implementation
a language agnostic and fast algorithm for this task.
You can use it by having it injected into your own Java code, or by:
- On the query side, passing tensors as “embed(some text)”
- On the indexing side, use the “embed” command to turn a text field into a tensor.
As part of this we added a genetic Embedder interface,
so applications can plug in any algorithm for this and use it in queries and indexing as described above.
See this system test
for an example using this.
About Vespa: Largely developed by Yahoo engineers,
Vespa is an open source big data processing and serving engine.
It’s in use by many products, such as Yahoo News, Yahoo Sports, Yahoo Finance, and the Yahoo Ad Platform.
Thanks to feedback and contributions from the community, Vespa continues to grow.