AI

Embedding secure generative AI in mission-critical public safety applications | Amazon Web Services

Embedding secure generative AI in mission-critical public safety applications | Amazon Web Services

This post is co-written with Lawrence Zorio III from Mark43. Public safety organizations face the challenge of accessing and analyzing vast amounts of data quickly while maintaining strict security protocols. First responders need immediate access to relevant data across multiple systems, while command staff require rapid insights for operational decisions. Mission-critical public safety applications require the highest levels of security and reliability when implementing technology capabilities. Mark43, a public safety technology company, recognized this challenge and embedded generative artificial intelligence (AI) capabilities into their application using Amazon Q Business to transform how law enforcement agencies interact with their mission-critical applications. By embedding…
Read More
Streamline RAG applications with intelligent metadata filtering using Amazon Bedrock | Amazon Web Services

Streamline RAG applications with intelligent metadata filtering using Amazon Bedrock | Amazon Web Services

Retrieval Augmented Generation (RAG) has become a crucial technique for improving the accuracy and relevance of AI-generated responses. The effectiveness of RAG heavily depends on the quality of context provided to the large language model (LLM), which is typically retrieved from vector stores based on user queries. The relevance of this context directly impacts the model’s ability to generate accurate and contextually appropriate responses. One effective way to improve context relevance is through metadata filtering, which allows you to refine search results by pre-filtering the vector store based on custom metadata attributes. By narrowing down the search space to the…
Read More
Nvidia Says Its Blackwell Chip Is Fine, Nothing to See Here

Nvidia Says Its Blackwell Chip Is Fine, Nothing to See Here

A large portion of Nvidia’s growth this quarter was driven by data center revenue, totaling $30.8 billion for the quarter, which was up 112 percent from last year. The company’s gross profit margin was 74.5 percent, essentially flat from a year ago. But analysts expect that Nvidia’s margins could shrink as the company shifts to producing more Blackwell chips, which cost more to make than their less advanced predecessors.Nvidia’s earnings reports are seen as an important bellwether for the AI industry as a whole. The chip architect’s advanced GPUs, which power complex neural network processing, are what made the current…
Read More
Automate Q&A email responses with Amazon Bedrock Knowledge Bases | Amazon Web Services

Automate Q&A email responses with Amazon Bedrock Knowledge Bases | Amazon Web Services

Email remains a vital communication channel for business customers, especially in HR, where responding to inquiries can use up staff resources and cause delays. The extensive knowledge required can make it overwhelming to respond to email inquiries manually. In the future, high automation will play a crucial role in this domain. Using generative AI allows businesses to improve accuracy and efficiency in email management and automation. This technology allows for automated responses, with only complex cases requiring manual review by a human, streamlining operations and enhancing overall productivity. The combination of retrieval augmented generation (RAG) and knowledge bases enhances automated…
Read More
Google Gemini gets a memory function for personalized replies

Google Gemini gets a memory function for personalized replies

Just a heads up, if you buy something through our links, we may get a small share of the sale. It’s one of the ways we keep the lights on here. Click here for more. Google has updated its Gemini AI to include a new feature that allows its Advanced subscription users to store and recall personal preferences, interests, and specific details indefinitely. This will improve the AI’s ability to deliver personalized and relevant responses.  Previously, users could share preferences with Gemini, but this context was limited to individual conversations and would reset once a new chat started. With the…
Read More
The Man Behind Amazon’s Robot Army Wants Everyone to Have an AI-Powered Helper

The Man Behind Amazon’s Robot Army Wants Everyone to Have an AI-Powered Helper

Unlike other robots, Proxie’s battery can be swapped out to avoid downtime charging. Cobot declined to say how much Proxie costs to buy or lease, but mobile robots often cost tens of thousands of dollars a piece.The robots work alongside humans, taking turns moving carts and navigating busy spaces without running into anyone. Porter says the idea is for the robots to level up as AI becomes more capable, allowing for more sophisticated manipulation and communication.Cobot has a version of Proxie that will respond to voice commands using a large language model to parse utterances, Porter says. When a worker…
Read More
Unify structured data in Amazon Aurora and unstructured data in Amazon S3 for insights using Amazon Q | Amazon Web Services

Unify structured data in Amazon Aurora and unstructured data in Amazon S3 for insights using Amazon Q | Amazon Web Services

In today’s data-intensive business landscape, organizations face the challenge of extracting valuable insights from diverse data sources scattered across their infrastructure. Whether it’s structured data in databases or unstructured content in document repositories, enterprises often struggle to efficiently query and use this wealth of information. In this post, we explore how you can use Amazon Q Business, the AWS generative AI-powered assistant, to build a centralized knowledge base for your organization, unifying structured and unstructured datasets from different sources to accelerate decision-making and drive productivity. The solution combines data from an Amazon Aurora MySQL-Compatible Edition database and data stored in…
Read More
How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instances | Amazon Web Services

How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instances | Amazon Web Services

Large language models (LLMs) are AI systems trained on vast amounts of text data, enabling them to understand, generate, and reason with natural language in highly capable and flexible ways. LLM training has seen remarkable advances in recent years, with organizations pushing the boundaries of what’s possible in terms of model size, performance, and efficiency. In this post, we explore how FP8 optimization can significantly speed up large model training on Amazon SageMaker P5 instances. LLM training using SageMaker P5 In 2023, SageMaker announced P5 instances, which support up to eight of the latest NVIDIA H100 Tensor Core GPUs. Equipped…
Read More
Penguin Solutions Accelerates Time to Value for AI Factories

Penguin Solutions Accelerates Time to Value for AI Factories

Complete high-performing AI architectures based on Penguin Solutions OriginAI infrastructure software and services and Dell AI-optimized solutions showcased at SC24 Penguin and Dell Technologies co-sponsor University of New Mexico team in SC24 student cluster competition Penguin Solutions, Inc. (Penguin Solutions, Nasdaq: PENG), today announced its collaboration with Dell Technologies to deliver complete AI solutions that accelerate customers’ time to value through validated, high-performing AI infrastructure. These end-to-end Penguin OriginAI® solutions combine Dell PowerEdge AI-optimized servers, Dell PowerScale storage and Dell PowerSwitch data center switches with Penguin’s validated solution architectures. When enhanced by Penguin’s ClusterWare® intelligent cluster management software and supported by Penguin’s expert AI and high-performance…
Read More
No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.