AI

Using and Finetuning Pretrained Transformers

Using and Finetuning Pretrained Transformers

This week has been filled with developments, including exciting new AI research that I’ll be discussing in my usual end-of-month write-ups.Additionally, I am excited to announce the release of my new book, Machine Learning Q and AI, published by No Starch Press.If you’ve been searching for a resource following an introductory machine learning course, this might be the one. I’m covering 30 concepts that were slightly out of scope for the previous books and courses I’ve taught, and I’ve compiled them here in a concise question-and-answer format (including exercises).I believe it will also serve as a useful companion for preparing…
Read More
Google Taps AI to Show Shoppers How Clothes Fit Different Bodies

Google Taps AI to Show Shoppers How Clothes Fit Different Bodies

One of the new ad formats Google announced today will allow brands to link short-form videos they made—or ones they hired creators to film—to their advertisements in Google’s search engine. AI-generated text summaries of the clips will be included below. “I’ve got three Gen Z-ers at home, and watching them shop, it’s very video-based,” said Madrigal.Google also launched a tool that allows companies to create entirely new, AI-generated product images based on photos from earlier marketing campaigns and pictures that represent their brand identity. For example, a home goods brand could upload a picture of one of its candles and…
Read More
The Shift from Models to Compound AI Systems

The Shift from Models to Compound AI Systems

AI caught everyone’s attention in 2023 with Large Language Models (LLMs) that can be instructed to perform general tasks, such as translation or coding, just by prompting. This naturally led to an intense focus on models as the primary ingredient in AI application development, with everyone wondering what capabilities new LLMs will bring. As more developers begin to build using LLMs, however, we believe that this focus is rapidly changing: state-of-the-art AI results are increasingly obtained by compound systems with multiple components, not just monolithic models. For example, Google’s AlphaCode 2 set state-of-the-art results in programming through a carefully engineered…
Read More
How LotteON built a personalized recommendation system using Amazon SageMaker and MLOps | Amazon Web Services

How LotteON built a personalized recommendation system using Amazon SageMaker and MLOps | Amazon Web Services

This post is co-written with HyeKyung Yang, Jieun Lim, and SeungBum Shim from LotteON. LotteON aims to be a platform that not only sells products, but also provides a personalized recommendation experience tailored to your preferred lifestyle. LotteON operates various specialty stores, including fashion, beauty, luxury, and kids, and strives to provide a personalized shopping experience across all aspects of customers’ lifestyles. To enhance the shopping experience of LotteON’s customers, the recommendation service development team is continuously improving the recommendation service to provide customers with the products they are looking for or may be interested in at the right time.…
Read More
Meta’s answer to GPT-4o? Meet Chameleon

Meta’s answer to GPT-4o? Meet Chameleon

In the wake of OpenAI's recent announcement of GPT-4o, a new model that processes and generates text, audio, and images in real-time, it's clear that the race for multimodal AI is heating up. Not to be outdone, researchers at Meta AI (FAIR) have just released a fascinating new paper introducing Chameleon, a family of early-fusion foundation models that also seamlessly blend language and vision. In this post, we'll do a deep dive into the Chameleon paper, exploring how it pushes the boundaries of multimodal AI in different but equally exciting ways compared to GPT-4o. We’ll also speculate a bit on…
Read More

SambaNova Systems Welcomes Lip-Bu Tan as Executive Chairman

SambaNova Systems, makers of the fastest models and most efficient chips, embedded within a full-stack AI platform, announces the appointment of Lip-Bu Tan, a semiconductor visionary, as its new Executive Chairman. This announcement comes as SambaNova experiences rocketship growth amidst surging interest in generative AI. “It will be an honor to guide SambaNova as Executive Chairman,” said Lip-Bu Tan. “The potential of this company is immense — SambaNova has differentiated itself by providing a privatized, full-stack AI platform that enterprises can own; the fastest, open models, most efficient chips, software and services. SambaNova is committed to ensuring customers can own…
Read More
Messaging your AI pricing model

Messaging your AI pricing model

This article is based on Ismail Madni’s brilliant talk at the Product Marketing Summit in Austin, hosted by our sister community, Product Marketing Alliance. More and more AI capabilities are being added to product roadmaps every single day. Even companies that aren’t using true AI are still incorporating increasingly advanced capabilities into their product. This gives us all a golden opportunity to rethink how we price our offerings and the story behind that pricing – and that’s what I’m excited to talk to you about today.A brief history of software pricing modelsLet's start by taking a quick look at the history…
Read More
How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?

How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?

April 2024, what a month! My birthday, a new book release, spring is finally here, and four major open LLM releases: Mixtral, Meta AI's Llama 3, Microsoft's Phi-3, and Apple's OpenELM.This article reviews and discusses all four major transformer-based LLM model releases that have been happening in the last few weeks, followed by new research on reinforcement learning with human feedback methods for instruction finetuning using PPO and DPO algorithms.1. How Good are Mixtral, Llama 3, and Phi-3?2. OpenELM: An Efficient Language Model Family with Open-source Training and Inference Framework3. Is DPO Superior to PPO for LLM Alignment? A Comprehensive…
Read More
OpenAI’s Long-Term AI Risk Team Has Disbanded

OpenAI’s Long-Term AI Risk Team Has Disbanded

In July last year, OpenAI announced the formation of a new research team that would prepare for the advent of supersmart artificial intelligence capable of outwitting and overpowering its creators. Ilya Sutskever, OpenAI’s chief scientist and one of the company’s cofounders, was named as the colead of this new team. OpenAI said the team would receive 20 percent of its computing power.Now OpenAI’s “superalignment team” is no more, the company confirms. That comes after the departures of several researchers involved, Tuesday’s news that Sutskever was leaving the company, and the resignation of the team’s other colead. The group’s work will…
Read More
2024 BAIR Graduate Directory

2024 BAIR Graduate Directory

Every year, the Berkeley Artificial Intelligence Research (BAIR) Lab graduates some of the most talented and innovative minds in artificial intelligence and machine learning. Our Ph.D. graduates have each expanded the frontiers of AI research and are now ready to embark on new adventures in academia, industry, and beyond. These fantastic individuals bring with them a wealth of knowledge, fresh ideas, and a drive to continue contributing to the advancement of AI. Their work at BAIR, ranging from deep learning, robotics, and natural language processing to computer vision, security, and much more, has contributed significantly to their fields and has…
Read More
No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.