Lowest latency machine learning inference accelerator for the finance industry

Designed for finance

VOLLO™ is designed to achieve the lowest latency on financial neural network models, while maximizing throughput, quality and energy- and space-efficiency. Its success has been convincingly demonstrated by its performance in the STAC-ML™ Markets (Inference) benchmarks1 which represent such models.

Unrivalled low latency

Independently audited results show that VOLLO has less than half the latency of its nearest competitor.  Other competitors demonstrated up to 20x longer latencies than VOLLO.  Audited latencies as low as 5.1 microseconds were achieved by VOLLO for the neural network models defined in the STAC–ML benchmarks and subsequent improvements have reduced those latencies even further.

Although not audited by STAC, the compute latency (excluding off-chip communications) is of the order of just 1 microsecond. This means VOLLO is fast enough to open up new applications, such as inference in a NIC subsystem.

Works with your ML model

VOLLO is optimized for time-series inference of financial AI models.

It supports user-defined models in ONNX or PyTorch comprising layer types such as these.

Additional layer types may be added on request.

  • Fully Connected
  • LSTM
  • 1D Convolution

Quick to evaluate & deploy

Evaluate:

  • Virtual machine for performance estimation on your own models.  Does not require an FPGA card.

Deploy:

  • Standard FHFL or low profile PCIe card
  • FPGA netlist for integration into existing systems (mid-2024)

High throughput and power-efficiency

Designed to be installed in a server co-located in a stock exchange, VOLLO achieves very high throughput and low energy consumption in a 1U server. This significantly reduces the costs incurred in running co-located servers. Up to four PCIe accelerator cards will run in a 1U server at less than 650W.

Simple to program

Models can be trained in PyTorch or TensorFlow before being exported in ONNX format into the VOLLO tool suite, making it simple to program from your existing ML development environment.

Flexible for future-proofing

The flexibility of FPGA technology ensures that not only can VOLLO be software-configured with multiple user models today, but significant architectural innovations can also be adopted quickly with optimal compute resources3.