Zoho Corporation has announced that it will be leveraging the NVIDIA AI accelerated computing platform - which includes NVIDIA NeMo, part of NVIDIA AI Enterprise software – to build and deploy its large language models (LLMs) in its SaaS applications. 

COMMERCIAL BREAK
SCROLL TO CONTINUE READING

Once the LLMs are built and deployed, they will be available to Zoho Corporation's 700,000+ customers globally. Over the past year, the company has invested more than USD 10 million in NVIDIA's AI technology and GPUs, and plans to invest an additional USD 10 million in the coming year. The announcement was made during NVIDIA AI Summit in Mumbai. 

Zoho has been building its own AI technology for over a decade and adding it contextually to its wide portfolio of over 100 products across its ManageEngine and Zoho divisions. Its approach to AI is multi-modal, geared towards deriving contextual intelligence that can help users make business decisions. 

The company is building narrow, small and medium language models, which are distinct from LLMs. This provides options for using different size models in order to provide better results across a variety of use cases. Relying on multiple models also means that businesses that do not have a large amount of data can still benefit from AI. Privacy is also a core tenet in Zoho's AI strategy, and its LLM models will not be trained on customer data.

Through this collaboration, Zoho will be accelerating its LLMs on the NVIDIA accelerated computing platform with NVIDIA Hopper  GPUs, using the NVIDIA NeMo end-to-end platform for developing custom generative AI—including LLMs, multimodal, vision, and speech AI. Additionally, Zoho is testing NVIDIA TensorRT-LLM to optimize its LLMs for deployment, and has already seen a 60% increase in throughput and 35% reduction in latency compared with a previously used open-source framework. The company is also accelerating other workloads like speech-to-text on NVIDIA accelerated computing infrastructure.