24 links
tagged with hugging-face
Click any tag below to further narrow down your results
Links
The Smol Training Playbook on Hugging Face provides a comprehensive guide for efficiently training machine learning models using the Hugging Face ecosystem. It emphasizes best practices and methodologies for optimizing training processes, making it accessible for both beginners and experienced practitioners. The playbook also includes practical examples and resources to enhance the learning experience.
Trackio is a new open-source experiment tracking library from Hugging Face that simplifies the process of tracking metrics during machine learning model training. It features a local dashboard, seamless integration with Hugging Face Spaces for easy sharing, and compatibility with existing libraries like wandb, allowing users to adopt it with minimal changes to their code.
Hugging Face has partnered with Cloudflare to provide FastRTC developers with access to enterprise-grade WebRTC infrastructure, enabling the creation of low-latency audio and video streams. This collaboration simplifies the development process by leveraging Cloudflare's global TURN network, allowing developers to focus on building applications without the burden of maintaining communication infrastructure. Free streaming is available for developers with a Hugging Face token, supporting various AI applications.
The webpage features the Deepsite application hosted on Hugging Face Spaces, showcasing its current running status and community interactions. Users can explore files and access the app, which fetches metadata from the HF Docker repository. The platform has garnered significant interest, reflected in its high like count.
Hugging Face has launched a new 3D-printed robotic arm priced starting at $100, aimed at making robotics more accessible to hobbyists and educators. The arm is designed for various applications, including prototyping and learning, emphasizing affordability and ease of use.
Hugging Face has unveiled two new humanoid robots designed to enhance human-robot interaction and improve AI capabilities. These robots are equipped with advanced language processing and learning abilities, aiming to bridge the gap between humans and machines in various applications.
Large diffusion models like Flux can generate impressive images but require substantial memory, making quantization an attractive option to reduce their size without significantly affecting output quality. The article discusses various quantization backends available in Hugging Face Diffusers, including bitsandbytes, torchao, and Quanto, and provides examples of how to implement these quantizations to optimize memory usage and performance in image generation tasks.
Parquet Content-Defined Chunking (CDC) is now integrated with PyArrow and Pandas, allowing efficient deduplication of Parquet files on content-addressable storage like Hugging Face's Xet storage layer. This feature significantly reduces data transfer and storage costs by only uploading or downloading modified data chunks, streamlining data workflows. Demonstrations highlight its effectiveness in various scenarios, including adding or removing columns and re-uploading identical tables without incurring additional data transfer.
The Hugging Face CLI has been renamed from `huggingface-cli` to `hf`, introducing a clearer and more organized command structure. Users can easily access commands grouped by resource, with new features like `hf jobs` for running scripts on Hugging Face infrastructure. The legacy CLI remains functional to aid in the transition to the new format.
Groq has been integrated as a new Inference Provider on the Hugging Face Hub, enhancing serverless inference capabilities for a variety of text and conversational models. Utilizing Groq's Language Processing Unit (LPU™), developers can achieve faster inference for Large Language Models with a pay-as-you-go API, while managing preferences and API keys directly from their user accounts on Hugging Face.
Featherless AI is now an Inference Provider on the Hugging Face Hub, enhancing serverless AI inference capabilities with a wide range of supported models. Users can easily integrate Featherless AI into their projects using client SDKs for both Python and JavaScript, with flexible billing options depending on their API key usage. PRO users receive monthly inference credits and access to additional features.
Hugging Face has launched a new deployment option for OpenAI's Whisper model on Inference Endpoints, offering up to 8x performance improvements for transcription tasks. The platform leverages advanced optimizations like PyTorch compilation and CUDA graphs, enhancing the efficiency and speed of audio transcriptions while maintaining high accuracy. Users can easily deploy their own ASR pipelines with minimal effort and access powerful hardware options.
The AI Agents Course offers a comprehensive journey from beginner to expert in understanding and building AI agents. It includes foundational units, hands-on practice with popular libraries, and opportunities for certification, all while fostering community engagement through Discord and collaborative assignments.
Hugging Face has announced the release of a free operator-like AI tool designed for enhancing agentic capabilities in AI applications. This tool aims to democratize access to advanced AI functionalities, enabling developers to create more intelligent and responsive systems. The initiative reflects Hugging Face's commitment to fostering innovation in the AI community.
Cohere has become a supported Inference Provider on the Hugging Face Hub, allowing users to access a variety of enterprise-focused AI models designed for tasks such as generative AI, embeddings, and vision-language applications. The article highlights several of Cohere's models, their features, and how to implement them using the Hugging Face platform, including serverless inference capabilities and integration with client SDKs.
Hugging Face has announced a new collaboration with NVIDIA called Training Cluster as a Service, aimed at providing accessible GPU clusters for research organizations globally. This initiative allows institutions to request GPU capacity for training AI models on-demand, addressing the growing compute gap in AI research.
Microsoft and Hugging Face have expanded their collaboration to make over 10,000 open models easily deployable on Azure, enhancing accessibility for developers while ensuring secure deployment alongside company data. The initiative aims to empower enterprises to build AI applications using a diverse range of models, with ongoing updates and support for various modalities.
Generating detailed images with AI has become more accessible by connecting Claude to Hugging Face Spaces, enabling users to leverage advanced models like FLUX.1 Krea and Qwen-Image. These models enhance image realism and text quality, allowing for creative projects such as posters and marketing materials. Users can easily configure and switch between these models to achieve desired results.
Scaleway has been added as a new Inference Provider on the Hugging Face Hub, allowing users to easily access various AI models through a serverless API. The service features competitive pricing, low latency, and supports advanced functionalities like structured outputs and multimodal processing, making it suitable for production use. Users can manage their API keys and preferences directly within their accounts for seamless integration.
ZeroGPU enables efficient use of Nvidia H200 hardware in Hugging Face Spaces by allowing users to avoid keeping GPUs locked during idle periods. The article discusses how ahead-of-time (AoT) compilation with PyTorch can significantly enhance performance, reducing processing time for generating images and videos with speedups of 1.3x to 1.8x. It also provides a guide on implementing AoT compilation in ZeroGPU Spaces, including advanced techniques like FP8 quantization.
HiDream-I1 is an open-source image generative foundation model boasting 17 billion parameters, delivering high-quality image generation in seconds. Its recent updates include the release of various models and integrations with popular platforms, enhancing its usability for developers and users alike. For full capabilities, users can explore additional resources and demos linked in the article.
Hugging Face is transforming its existing NLP course into the LLM course to better reflect advancements in AI, incorporating new chapters on fine-tuning LLMs and reasoning models. While classic NLP material will be updated, the course will also focus on making cutting-edge research accessible and include interactive exercises and live sessions for enhanced learning.
Hugging Face has launched AI Sheets, a no-code tool that simplifies the process of building, enriching, and transforming datasets using open AI models. The user-friendly interface allows users to easily experiment with datasets, generate synthetic data, and refine prompts by providing feedback directly within the tool. It supports both local and cloud deployment, making it accessible for various use cases.
SINQ is a fast and model-agnostic quantization technique that enables the deployment of large language models on GPUs with limited memory while maintaining accuracy. It significantly reduces memory requirements and quantization time, offering improved model quality compared to existing methods. The technique introduces dual scaling to enhance quantization stability, allowing users to quantize models quickly and efficiently.