Viral News

Achieving Dimension-Free Communication in Federated Learning via Zeroth-Order Optimization

Achieving Dimension-Free Communication in Federated Learning via Zeroth-Order Optimization

arXiv:2405.15861v1 Announce Type: new Abstract: Federated Learning (FL) offers a promising framework for collaborative and privacy-preserving machine learning across distributed data sources. However, the substantial communication costs associated with FL pose a significant challenge to its efficiency. Specifically, in each communication round, the communication costs scale linearly with the model's dimension, which presents a formidable obstacle, especially in large model scenarios. Despite various communication efficient strategies, the intrinsic dimension-dependent communication cost remains a major bottleneck for current FL implementations. In this paper, we introduce a novel dimension-free communication strategy for FL, leveraging zero-order optimization techniques. We propose a new algorithm,…
Read More
Rethinking the Elementary Function Fusion for Single-Image Dehazing

Rethinking the Elementary Function Fusion for Single-Image Dehazing

arXiv:2405.15817v1 Announce Type: new Abstract: This paper addresses the limitations of physical models in the current field of image dehazing by proposing an innovative dehazing network (CL2S). Building on the DM2F model, it identifies issues in its ablation experiments and replaces the original logarithmic function model with a trigonometric (sine) model. This substitution aims to better fit the complex and variable distribution of haze. The approach also integrates the atmospheric scattering model and other elementary functions to enhance dehazing performance. Experimental results demonstrate that CL2S achieves outstanding performance on multiple dehazing datasets, particularly in maintaining image details and color authenticity.…
Read More
SLIDE: A Framework Integrating Small and Large Language Models for Open-Domain Dialogues Evaluation

SLIDE: A Framework Integrating Small and Large Language Models for Open-Domain Dialogues Evaluation

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs. Source link lol
Read More
Spatio-temporal Value Semantics-based Abstraction for Dense Deep Reinforcement Learning

Spatio-temporal Value Semantics-based Abstraction for Dense Deep Reinforcement Learning

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs. Source link lol
Read More
From CNNs to Transformers in Multimodal Human Action Recognition: A Survey

From CNNs to Transformers in Multimodal Human Action Recognition: A Survey

[Submitted on 22 May 2024] View a PDF of the paper titled From CNNs to Transformers in Multimodal Human Action Recognition: A Survey, by Muhammad Bilal Shaikh and 2 other authors View PDF Abstract:Due to its widespread applications, human action recognition is one of the most widely studied research problems in Computer Vision. Recent studies have shown that addressing it using multimodal data leads to superior performance as compared to relying on a single data modality. During the adoption of deep learning for visual modelling in the last decade, action recognition approaches have mainly relied on Convolutional Neural Networks (CNNs).…
Read More
Enhancing Augmentative and Alternative Communication with Card Prediction and Colourful Semantics

Enhancing Augmentative and Alternative Communication with Card Prediction and Colourful Semantics

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs. Source link lol
Read More
Efficient Mitigation of Bus Bunching through Setter-Based Curriculum Learning

Efficient Mitigation of Bus Bunching through Setter-Based Curriculum Learning

arXiv:2405.15824v1 Announce Type: new Abstract: Curriculum learning has been growing in the domain of reinforcement learning as a method of improving training efficiency for various tasks. It involves modifying the difficulty (lessons) of the environment as the agent learns, in order to encourage more optimal agent behavior and higher reward states. However, most curriculum learning methods currently involve discrete transitions of the curriculum or predefined steps by the programmer or using automatic curriculum learning on only a small subset training such as only on an adversary. In this paper, we propose a novel approach to curriculum learning that uses a…
Read More
Sequence Length Scaling in Vision Transformers for Scientific Images on Frontier

Sequence Length Scaling in Vision Transformers for Scientific Images on Frontier

[Submitted on 17 Apr 2024] Authors:Aristeidis Tsaris, Chengming Zhang, Xiao Wang, Junqi Yin, Siyan Liu, Moetasim Ashfaq, Ming Fan, Jong Youl Choi, Mohamed Wahib, Dan Lu, Prasanna Balaprakash, Feiyi Wang View a PDF of the paper titled Sequence Length Scaling in Vision Transformers for Scientific Images on Frontier, by Aristeidis Tsaris and 11 other authors View PDF Abstract:Vision Transformers (ViTs) are pivotal for foundational models in scientific imagery, including Earth science applications, due to their capability to process large sequence lengths. While transformers for text has inspired scaling sequence lengths in ViTs, yet adapting these for ViTs introduces unique challenges.…
Read More
DuanzAI: Slang-Enhanced LLM with Prompt for Humor Understanding

DuanzAI: Slang-Enhanced LLM with Prompt for Humor Understanding

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website. Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them. Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs. Source link lol
Read More
Executive Overview: The Rise of Open Foundational Models

Executive Overview: The Rise of Open Foundational Models

Moving generative AI applications from the proof of concept stage into production requires control, reliability and data governance. Organizations are turning to open source foundation models in search of that control and the ability to better influence outputs by more tightly managing both the models and the data they are trained on.Databricks has assisted thousands of customers in evaluating use cases for generative AI and determining the most appropriate architecture for their organization.Our customers have shared with us the challenge of building and deploying production-quality AI models, which is often difficult and expensive. As a result, most CIOs are not…
Read More
No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.