Latest AI & Business News
Stay updated with the latest insights in AI and business, delivered directly to you.
-
Circuit Breakers for AI: Interrupting Harmful Outputs Through Representation Engineering
The adversarial attacks and defenses for LLMs encompass a wide range of techniques and strategies. Manually crafted and automated red teaming methods expose vulnerabilities, while white box access reveals potential for prefilling attacks. Defense approaches include RLHF, DPO, prompt optimization, and adversarial training. Inference-time defenses and representation engineering show promise but face limitations. The control…
-
This AI Paper from China Introduces a Reward-Robust Reinforcement Learning from Human Feedback RLHF Framework for Enhancing the Stability and Performance of Large Language Models
Reinforcement Learning from Human Feedback (RLHF) has emerged as a vital technique in aligning large language models (LLMs) with human values and expectations. It plays a critical role in ensuring that AI systems behave in understandable and trustworthy ways. RLHF enhances the capabilities of LLMs by training them based on feedback that allows models to…
-
JailbreakBench: An Open Sourced Benchmark for Jailbreaking Large Language Models (LLMs)
Large Language Models (LLMs) are vulnerable to jailbreak attacks, which can generate offensive, immoral, or otherwise improper information. By taking advantage of LLM flaws, these attacks go beyond the safety precautions meant to prevent offensive or hazardous outputs from being generated. Jailbreak attack evaluation is a very difficult procedure, and existing benchmarks and evaluation methods…
-
Salesforce AI Introduces SFR-Judge: A Family of Three Judge Models of 8-Billion Parameters 8B, 12B, and 70B Size, Built with Meta Llama 3 and Mistral NeMO
The advancement of large language models (LLMs) in natural language processing has significantly improved various domains. As more complex models are developed, evaluating their outputs accurately becomes essential. Traditionally, human evaluations have been the standard approach for assessing quality, but this process is time consuming and needs to be more scalable for the rapid pace…
-
SELMA: A Novel AI Approach to Enhance Text-to-Image Generation Models Using Auto-Generated Data and Skill-Specific Learning Techniques
Text-to-image (T2I) models have seen rapid progress in recent years, allowing the generation of complex images based on natural language inputs. However, even state-of-the-art T2I models need help accurately capture and reflect all the semantics in given prompts, leading to images that may miss crucial details, such as multiple subjects or specific spatial relationships. For…
-
Multi-View and Multi-Scale Alignment (MaMA): Advancing Mammography with Contrastive Learning and Visual-Language Pre-training
Multi-View and Multi-Scale Alignment for Mammography Contrastive Learning:Contrastive Language-Image Pre-training (CLIP) has shown potential in medical imaging, but its application to mammography faces challenges due to limited labeled data, high-resolution images, and imbalanced datasets. This study introduces the first full adaptation of CLIP to mammography through a new framework called Multi-view and Multi-scale Alignment (MaMA).…
-
Why AI is a know-it-all know nothing
Today’s AI systems can’t earn our trust by sharing the reasoning behind what they say, because there is no such reasoning.Read More
-
Practical Lossless Text Compression: FineZip Delivers 54x Speed Boost via Large Language Models
Although the connection between language modeling and data compression has been recognized for some time, current Large Language Models (LLMs) are not typically used for practical text compression due to their lengthy processing times. For example, LLMZip, a recent compression system based on the LLaMA3-8B model, requires 9.5 days to compress just 10 MB of…
-
IEEE’s Let’s Make Light Competition Returns to Tackle Illumination Gap
In economically advantaged countries, it’s hard to imagine a time when electric lighting wasn’t available in nearly every home, business, and public facility. But according to the World Economic Forum, the sun remains the primary light source for more than 1 billion people worldwide. Known as light poverty, the lack of access to reliable, adequate,…
-
AMD Releases AMD-135M: AMD’s First Small Language Model Series Trained from Scratch on AMD Instinct™ MI250 Accelerators Utilizing 670B Tokens
AMD has recently introduced its new language model, AMD-135M or AMD-Llama-135M, which is a significant addition to the landscape of AI models. Based on the LLaMA2 model architecture, this language model boasts a robust structure with 135 million parameters and is optimized for performance on AMD’s latest GPUs, specifically the MI250. This release marks a…