3 links
tagged with all of: ai-models + hugging-face + inference
Click any tag below to further narrow down your results
Links
Hugging Face has launched a new deployment option for OpenAI's Whisper model on Inference Endpoints, offering up to 8x performance improvements for transcription tasks. The platform leverages advanced optimizations like PyTorch compilation and CUDA graphs, enhancing the efficiency and speed of audio transcriptions while maintaining high accuracy. Users can easily deploy their own ASR pipelines with minimal effort and access powerful hardware options.
Cohere has become a supported Inference Provider on the Hugging Face Hub, allowing users to access a variety of enterprise-focused AI models designed for tasks such as generative AI, embeddings, and vision-language applications. The article highlights several of Cohere's models, their features, and how to implement them using the Hugging Face platform, including serverless inference capabilities and integration with client SDKs.
Scaleway has been added as a new Inference Provider on the Hugging Face Hub, allowing users to easily access various AI models through a serverless API. The service features competitive pricing, low latency, and supports advanced functionalities like structured outputs and multimodal processing, making it suitable for production use. Users can manage their API keys and preferences directly within their accounts for seamless integration.