stp2y

29120 Posts
Learning to Retrieve Iteratively for In-Context Learning

Learning to Retrieve Iteratively for In-Context Learning

arXiv:2406.14739v1 Announce Type: new Abstract: We introduce iterative retrieval, a novel framework that empowers retrievers to make iterative decisions through policy optimization. Finding an optimal portfolio of retrieved items is a combinatorial optimization problem, generally considered NP-hard. This approach provides a learned approximation to such a solution, meeting specific task requirements under a given family of large language models (LLMs). We propose a training procedure based on reinforcement learning, incorporating feedback from LLMs. We instantiate an iterative retriever for composing in-context learning (ICL) exemplars and apply it to various semantic parsing tasks that demand synthesized programs as outputs. By adding…
Read More
How Gradient created an open LLM with a million-token context window

How Gradient created an open LLM with a million-token context window

Don’t miss OpenAI, Chevron, Nvidia, Kaiser Permanente, and Capital One leaders only at VentureBeat Transform 2024. Gain essential insights about GenAI and expand your network at this exclusive three day event. Learn More In a recent collaboration, AI startup Gradient and cloud compute platform Crusoe extended the “context window” of Llama-3 models to 1 million tokens. The context window determines the number of input and output tokens a large language model (LLM) can process.  Big tech companies and frontier AI labs are locked in a race to extend the context windows of their LLMs. In a few months, models have…
Read More
An embattled oil tycoon is selling a mansion in Singapore for $32 million. It’s one of the city’s status-symbol houses that business leaders covet.

An embattled oil tycoon is selling a mansion in Singapore for $32 million. It’s one of the city’s status-symbol houses that business leaders covet.

A former oil tycoon convicted of cheating and instigating forgery is selling his mansion in Singapore for $43 million Singapore dollars, or $32 million.Lim Oon Kuin, the founder of the collapsed oil firm Hin Leong Trading, is putting his Good Class Bungalow in Tanglin Hill, one of the city's wealthy enclaves, for sale.A home on private property is incredibly rare in the land-scarce city — which spans a mere 274 square miles. The largest and most expensive type of landed home available in Singapore is known as a Good Class Bungalow, or GCB.There are only an estimated 2,800 GCBs in…
Read More
Probabilistic Emulation of a Global Climate Model with Spherical DYffusion

Probabilistic Emulation of a Global Climate Model with Spherical DYffusion

[Submitted on 21 Jun 2024] View a PDF of the paper titled Probabilistic Emulation of a Global Climate Model with Spherical DYffusion, by Salva R"uhling Cachay and 4 other authors View PDF HTML (experimental) Abstract:Data-driven deep learning models are on the verge of transforming global weather forecasting. It is an open question if this success can extend to climate modeling, where long inference rollouts and data complexity pose significant challenges. Here, we present the first conditional generative model able to produce global climate ensemble simulations that are accurate and physically consistent. Our model runs at 6-hourly time steps and is…
Read More
Substrate lands $8M funding to bring ‘Lego blocks’ approach to enterprise AI

Substrate lands $8M funding to bring ‘Lego blocks’ approach to enterprise AI

Don’t miss OpenAI, Chevron, Nvidia, Kaiser Permanente, and Capital One leaders only at VentureBeat Transform 2024. Gain essential insights about GenAI and expand your network at this exclusive three day event. Learn More Substrate, a startup founded by tech industry veterans Rob Cheung and Ben Guo, quietly emerged from stealth last week to launch its artificial intelligence development platform. The company also announced it has raised $8 million in a funding round led by Lightspeed Venture Partners to grow its team and expand its product offerings. Substrate aims to democratize AI by providing a unified platform for enterprises to build,…
Read More
PEANO-ViT: Power-Efficient Approximations of Non-Linearities in Vision Transformers

PEANO-ViT: Power-Efficient Approximations of Non-Linearities in Vision Transformers

arXiv:2406.14854v1 Announce Type: new Abstract: The deployment of Vision Transformers (ViTs) on hardware platforms, specially Field-Programmable Gate Arrays (FPGAs), presents many challenges, which are mainly due to the substantial computational and power requirements of their non-linear functions, notably layer normalization, softmax, and Gaussian Error Linear Unit (GELU). These critical functions pose significant obstacles to efficient hardware implementation due to their complex mathematical operations and the inherent resource count and architectural limitations of FPGAs. PEANO-ViT offers a novel approach to streamlining the implementation of the layer normalization layer by introducing a division-free technique that simultaneously approximates the division and square root…
Read More
New difficulty mod in Stardew Valley will purge your saves if you use a guide

New difficulty mod in Stardew Valley will purge your saves if you use a guide

A great number of us have played games in extra-difficult modes (or in the case of Kingdom Hearts, Proud Mode) to challenge ourselves. Now, a Stardew Valley player has created a “hardcore” option for the otherwise chill game, one that will delete the save files of any player who uses a guide while playing the game on PC.According to , software engineer Sylvie Nightshade created the high difficulty mod on June 21 after reading an article published the day before on the satirical website , the gaming version of The Onion. The article in question joked about a “hardcore mode”…
Read More
Dissecting the Ullman Variations with a SCALPEL: Why do LLMs fail at Trivial Alterations to the False Belief Task?

Dissecting the Ullman Variations with a SCALPEL: Why do LLMs fail at Trivial Alterations to the False Belief Task?

arXiv:2406.14737v1 Announce Type: new Abstract: Recent empirical results have sparked a debate about whether or not Large Language Models (LLMs) are capable of Theory of Mind (ToM). While some have found LLMs to be successful on ToM evaluations such as the False Belief task (Kosinski, 2023), others have argued that LLMs solve these tasks by exploiting spurious correlations -- not representing beliefs -- since they fail on trivial alterations to these tasks (Ullman, 2023). In this paper, we introduce SCALPEL: a technique to generate targeted modifications for False Belief tasks to test different specific hypotheses about why LLMs fail. We…
Read More
Set dtype with dtype argument functions and get it in PyTorch

Set dtype with dtype argument functions and get it in PyTorch

You can set dtype with the functions which have dtype arguments and get it with dtype and type() as shown below: *Memos: tensor(). *My post explains tensor(): import torch my_tensor = torch.tensor([0, 1, 2]) my_tensor = torch.tensor([0, 1, 2], dtype=torch.int64) my_tensor = torch.tensor([0, 1, 2], dtype=int) my_tensor, my_tensor.dtype, my_tensor.type() # (tensor([0, 1, 2]), torch.int64, 'torch.LongTensor') my_tensor = torch.tensor([0., 1., 2.], dtype=torch.float64) my_tensor = torch.tensor([0., 1., 2.], dtype=float) my_tensor, my_tensor.dtype, my_tensor.type() # (tensor([0., 1., 2.], dtype=torch.float64), # torch.float64, # 'torch.DoubleTensor') my_tensor = torch.tensor([0.+7.j, 1.+4.j, 2.+5.j], dtype=torch.complex32) my_tensor, my_tensor.dtype, my_tensor.type() # (tensor([0.+7.j, 1.+4.j, 2.+5.j], dtype=torch.complex32), # torch.complex32, # 'torch.ComplexHalfTensor') my_tensor = torch.tensor([True,…
Read More
No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.