6 links tagged with all of: machine-learning + generative-models
Click any tag below to further narrow down your results
+ embeddings
(1)
+ feedback
(1)
+ optimization
(1)
+ reinforcement-learning
(1)
+ performance
(1)
+ scalability
(1)
+ recommendation-systems
(1)
+ toolkit
(1)
+ transformers
(1)
+ continual-learning
(1)
+ hierarchical-architecture
(1)
+ artificial-intelligence
(1)
+ world-models
(1)
+ autoencoders
(1)
+ neural-networks
(1)
Links
The article explains reinforcement learning through a psychological lens, focusing on feedback mechanisms in both humans and computers. It outlines how computer programs learn by receiving scores, updating their responses, and emphasizes a specific approach called Reformist RL, which simplifies implementation for generative models.
This article discusses the challenges and solutions in developing large-scale generative recommendation systems, particularly in managing user data and improving training efficiency. It highlights techniques like multi-modal item towers and sampled softmax to enhance performance while addressing issues like cold-start and latency.
Pingkit is a toolkit designed for training reproducible, capacity-aware models using transformer activations. It offers features for extracting embeddings, training neural architectures, and creating custom probes tailored to specific research needs. The toolkit is integrated with Hugging Face models and provides various utilities for data processing and model training.
The essay critiques various perspectives on world models, which are essential for developing virtual agents with artificial general intelligence. Drawing from sci-fi and psychology, it emphasizes that a world model should simulate all actionable possibilities of the real world for effective reasoning and action, and proposes a new hierarchical architecture for such models within a Physical, Agentic, and Nested (PAN) AGI framework.
+ world-models
+ artificial-intelligence
machine-learning ✓
+ hierarchical-architecture
generative-models ✓
Contemporary generative models leverage a two-stage approach, first extracting a latent representation of input signals via an autoencoder, then training a generative model on these latents. This method enhances efficiency by focusing on perceptually meaningful information while reducing the computational burden associated with processing raw pixel or waveform data. The article details the training process, the evolution of generative techniques, and the significance of latent representations in modern applications.
ContinualFlow is a novel framework designed for targeted unlearning in generative models, utilizing Flow Matching and an energy-based reweighting loss to effectively remove undesired data distribution regions without extensive retraining. The method demonstrates its effectiveness through various experiments and provides visualizations and quantitative evaluations to support its claims.