7 links tagged with all of: image-generation + open-source
Click any tag below to further narrow down your results
Links
FLUX.2 is a new image generation and editing model that excels in creating high-quality images while maintaining consistency across multiple references. It supports detailed typography and complex prompts, making it suitable for various creative workflows. The model emphasizes open innovation, offering different versions for developers and teams.
STARFlow and STARFlow-V are open-source models designed for generating high-quality images and videos from text prompts. They combine autoregressive models with normalizing flows to achieve impressive results in both text-to-image and text-to-video tasks. Users can easily set up the models and start generating content with provided scripts and configurations.
GLM-Image is an open-source model that combines auto-regressive and diffusion techniques for high-quality image generation. It excels in generating detailed images from text prompts and supports various image editing tasks. The model uses a semantic-VQ tokenization strategy to enhance semantic understanding and visual fidelity.
The article introduces the FLUX.2 [klein] model family, which offers rapid image generation and editing capabilities in under half a second. It combines text-to-image and multi-reference generation in a compact architecture that runs efficiently on consumer hardware. Open weights are available for customization and fine-tuning.
The paper presents BLIP3-o, a family of fully open unified multimodal models that enhance both image understanding and generation. It introduces a diffusion transformer for generating CLIP image features, advocates for a sequential pretraining strategy, and proposes a high-quality dataset, BLIP3o-60k, to improve performance across various benchmarks. The models, along with code and datasets, are open-sourced to foster further research.
HunyuanImage-3.0 has been released as an open-source image generation model, featuring a unified multimodal architecture that integrates text and image understanding. It boasts the largest Mixture of Experts model with 80 billion parameters, enabling superior image generation capabilities while supporting extensive customization through various checkpoints and performance optimizations.
HiDream-I1 is an open-source image generative foundation model boasting 17 billion parameters, delivering high-quality image generation in seconds. Its recent updates include the release of various models and integrations with popular platforms, enhancing its usability for developers and users alike. For full capabilities, users can explore additional resources and demos linked in the article.