Category: Uncategorized
-
How Cells Resist the Pressure of the Deep Sea
Cell membranes from comb jellies reveal a new kind of adaptation to the deep sea: curvy lipids that conform to an ideal shape under pressure. Read more
-
Protesters Say Uber and Lyft Are Still Failing Their Blind Passengers
At a protest in San Francisco, blind Uber and Lyft users claim that the ride-hailing companies aren’t doing enough to prevent drivers from turning them away. Read more
-
CREAM: A New Self-Rewarding Method that Allows the Model to Learn more Selectively and Emphasize on Reliable Preference Data
One of the most critical challenges of LLMs is how to align these models with human values and preferences, especially in generated texts. Most generated text outputs by models are inaccurate, biased, or potentially harmful—for example, hallucinations. This misalignment limits the potential usage of LLMs in real-world applications across domains such as education, health, and… Read more
-
This AI Paper Explores If Human Visual Perception can Help Computer Vision Models Outperform in Generalized Tasks
Human beings possess innate extraordinary perceptual judgments, and when computer vision models are aligned with them, model’s performance can be improved manifold. Various attributes such as scene layout, subject location, camera pose, color, perspective, and semantics help us have a clear picture of the world and objects within. The alignment of vision models with visual… Read more
-
Differentiable Rendering of Robots (Dr. Robot): A Robot Self-Model Differentiable from Its Visual Appearance to Its Control Parameters
Visual and action data are interconnected in robotic tasks, forming a perception-action loop. Robots rely on control parameters for movement, while VFMs excel in processing visual data. However, a modality gap exists between visual and action data arising from the fundamental differences in their sensory modalities, abstraction levels, temporal dynamics, contextual dependence, and susceptibility to… Read more
-
Self-Data Distilled Fine-Tuning: A Solution for Pruning and Supervised Fine-tuning Challenges in LLMs
Large language models (LLMs) like GPT-4, Gemini, and Llama 3 have revolutionized natural language processing through extensive pre-training and supervised fine-tuning (SFT). However, these models come with high computational costs for training and inference. Structured pruning has emerged as a promising method to improve LLM efficiency by selectively removing less critical components. Despite its potential,… Read more
-
Rethinking Direct Alignment: Balancing Likelihood and Diversity for Better Model Performance
The problem of over-optimization of likelihood in Direct Alignment Algorithms (DAAs), such as Direct Preference Optimisation (DPO) and Identity Preference Optimisation (IPO), arises when these methods fail to improve model performance despite increasing the likelihood of preferred outcomes. These algorithms, which are alternatives to Reinforcement Learning from Human Feedback (RLHF), aim to align language models… Read more
-
Harnessing Introspection in AI: How Large Language Models Are Learning to Understand and Predict Their Behavior for Greater Accuracy
Large Language models (LLMs) have long been trained to process vast amounts of data to generate responses that align with patterns seen during training. However, researchers are exploring a more profound concept: introspection, the ability of LLMs to reflect on their behavior and gain knowledge that isn’t directly derived from their training data. This new… Read more
-
Meta AI Releases Cotracker3: A Semi-Supervised Tracker that Produces Better Results with Unlabelled Data and Simple Architecture
Point tracking is paramount in video; from 3d reconstruction to editing tasks, a precise approximation of points is necessary to achieve quality results. Over time, trackers have incorporated transformer and neural network-based designs to track individual and multiple points simultaneously. However, these neural networks could be fully exploited only with high-quality training data. Now, while… Read more
-
Nvidia AI Introduces the Normalized Transformer (nGPT): A Hypersphere-based Transformer Achieving 4-20x Faster Training and Improved Stability for LLMs
The rise of Transformer-based models has significantly advanced the field of natural language processing. However, the training of these models is often computationally intensive, requiring substantial resources and time. This research addresses the issue of improving the training efficiency of Transformer models without compromising their performance. Specifically, it seeks to explore whether the benefits of… Read more