Pushing the Limits of Sparsity: A Bag of Tricks for Extreme Pruning

Pushing the Limits of Sparsity: A Bag of Tricks for Extreme Pruning

[Submitted on 20 Nov 2024 (v1), last revised 21 Nov 2024 (this version, v2)] View a PDF of the paper titled Pushing the Limits of Sparsity: A Bag of Tricks for Extreme Pruning, by Andy Li and 3 other authors View PDF HTML (experimental) Abstract:Pruning of deep neural networks has been an effective technique for reducing model size while preserving most of the performance of dense networks, crucial for deploying models on memory and power-constrained devices. While recent sparse learning methods have shown promising performance up to moderate sparsity levels such as 95% and 98%, accuracy quickly deteriorates when pushing…
Read More
Knowledge Graphs, Large Language Models, and Hallucinations: An NLP Perspective

Knowledge Graphs, Large Language Models, and Hallucinations: An NLP Perspective

arXiv:2411.14258v1 Announce Type: new Abstract: Large Language Models (LLMs) have revolutionized Natural Language Processing (NLP) based applications including automated text generation, question answering, chatbots, and others. However, they face a significant challenge: hallucinations, where models produce plausible-sounding but factually incorrect responses. This undermines trust and limits the applicability of LLMs in different domains. Knowledge Graphs (KGs), on the other hand, provide a structured collection of interconnected facts represented as entities (nodes) and their relationships (edges). In recent research, KGs have been leveraged to provide context that can fill gaps in an LLM understanding of certain topics offering a promising approach…
Read More
Overwatch’s new hero is Hazard, the spiky Scotsman you’ve been waiting for

Overwatch’s new hero is Hazard, the spiky Scotsman you’ve been waiting for

Overwatch 2’s newest hero is a tank with surprising mobility and brutal diving and brawling tactics. Hazard has a punk-rock aesthetic, Scottish brogue and a shotgun that can blast your vulnerable backline heroes to smithereens. But don’t be put off by his tough exterior: This anti-establishment rapscallion has a steadfast belief in bodily autonomy and a love of cute puppies.Playable during a limited-time trial, Hazard was initially conceived as a daunting cyberpunk villain called Spiker. But after Blizzard’s designers showcased their hulking scoundrel to the development team, it became clear he was destined to evolve into a charming rebel with…
Read More
Finding the Way: Backtracking Algorithm for Rat in a Maze

Finding the Way: Backtracking Algorithm for Rat in a Maze

Introduction Imagine a rat searching for cheese in a complex maze. Every path looks promising until it hits a dead end. How can it systematically explore every route without missing any possible solution? This is where the Backtracking Algorithm comes in, a powerful tool for solving intricate puzzles and real-world problems. Backtracking is a recursive algorithmic technique that incrementally builds solutions and abandons paths that don’t lead to a valid solution. Its significance lies in its simplicity and versatility, making it applicable in fields like AI, robotics, and optimization. In this blog, we’ll dive into how backtracking works, explore its…
Read More
Why this ‘Outlander’ character was recast for season 7

Why this ‘Outlander’ character was recast for season 7

"Outlander" season seven has seen one major member of Jamie Fraser's family recast.Jenny Murray was played by Laura Donnelly in earlier seasons.As of season seven part two, Kristin Atherton has taken over the role.Warning: Spoilers ahead for "Outlander" season seven, episode nine, "Unfinished Business."The second half of "Outlander" season seven has reintroduced viewers to a few characters not seen since the show's early days.As Jamie (Sam Heughan) and Claire (Caitríona Balfe) made their return to the Frasers' ancestral home, Lallybroch, in the midseason premiere, which aired on November 22, the couple reunited with a number of old acquaintances, including Ian…
Read More
Uncertainty-Guided Alignment for Unsupervised Domain Adaptation in Regression

Uncertainty-Guided Alignment for Unsupervised Domain Adaptation in Regression

[Submitted on 24 Jan 2024 (v1), last revised 21 Nov 2024 (this version, v3)] View a PDF of the paper titled Uncertainty-Guided Alignment for Unsupervised Domain Adaptation in Regression, by Ismail Nejjar and 3 other authors View PDF HTML (experimental) Abstract:Unsupervised Domain Adaptation for Regression (UDAR) aims to adapt models from a labeled source domain to an unlabeled target domain for regression tasks. Traditional feature alignment methods, successful in classification, often prove ineffective for regression due to the correlated nature of regression features. To address this challenge, we propose Uncertainty-Guided Alignment (UGA), a novel method that integrates predictive uncertainty into…
Read More
Decompose and Leverage Preferences from Expert Models for Improving Trustworthiness of MLLMs

Decompose and Leverage Preferences from Expert Models for Improving Trustworthiness of MLLMs

arXiv:2411.13697v1 Announce Type: new Abstract: Multimodal Large Language Models (MLLMs) can enhance trustworthiness by aligning with human preferences. As human preference labeling is laborious, recent works employ evaluation models for assessing MLLMs' responses, using the model-based assessments to automate preference dataset construction. This approach, however, faces challenges with MLLMs' lengthy and compositional responses, which often require diverse reasoning skills that a single evaluation model may not fully possess. Additionally, most existing methods rely on closed-source models as evaluators. To address limitations, we propose DecompGen, a decomposable framework that uses an ensemble of open-sourced expert models. DecompGen breaks down each response…
Read More
Natural Language Reinforcement Learning

Natural Language Reinforcement Learning

arXiv:2411.14251v1 Announce Type: cross Abstract: Reinforcement Learning (RL) mathematically formulates decision-making with Markov Decision Process (MDP). With MDPs, researchers have achieved remarkable breakthroughs across various domains, including games, robotics, and language models. This paper seeks a new possibility, Natural Language Reinforcement Learning (NLRL), by extending traditional MDP to natural language-based representation space. Specifically, NLRL innovatively redefines RL principles, including task objectives, policy, value function, Bellman equation, and policy iteration, into their language counterparts. With recent advancements in large language models (LLMs), NLRL can be practically implemented to achieve RL-like policy and value improvement by either pure prompting or gradient-based training.…
Read More
Threads is testing out advanced search features and AI summaries for trending topics

Threads is testing out advanced search features and AI summaries for trending topics

Threads is making more changes to address long-running complaints from users. This time, the company is testing out improvements to its search and trending topics feature in updates that Adam Mosseri described as “long-overdue improvements.”On search, Threads is testing the ability to search for posts within specific date ranges and account-specific searches. The changes are similar to some of X’s advanced search capabilities and could make it easier for users to look for a specific post they want to revisit. The lack of advanced search on Threads has long been frustrating and up to now, the most reliable way to…
Read More
Split Federated Learning Over Heterogeneous Edge Devices: Algorithm and Optimization

Split Federated Learning Over Heterogeneous Edge Devices: Algorithm and Optimization

arXiv:2411.13907v1 Announce Type: new Abstract: Split Learning (SL) is a promising collaborative machine learning approach, enabling resource-constrained devices to train models without sharing raw data, while reducing computational load and preserving privacy simultaneously. However, current SL algorithms face limitations in training efficiency and suffer from prolonged latency, particularly in sequential settings, where the slowest device can bottleneck the entire process due to heterogeneous resources and frequent data exchanges between clients and servers. To address these challenges, we propose the Heterogeneous Split Federated Learning (HSFL) framework, which allows resource-constrained clients to train their personalized client-side models in parallel, utilizing different cut…
Read More
No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.