Digital Media Net - Your Gateway To Digital media Creation. News and information on Digital Video, VR, Animation, Visual Effects, Mac Based media. Post Production, CAD, Sound and Music
Categories: DITNewsVR

AWS Announces More Model Choice and Powerful New Capabilities in Amazon Bedrock to Securely Build and Scale Generative AI Applications

The latest models from Anthropic, Cohere, Meta, Stability AI, and Amazon expand customers’ choice of industry-leading models to support a variety of use cases

Model Evaluation on Amazon Bedrock helps customers evaluate, compare, and select the best model for their use case and business needs

Knowledge Bases for Amazon Bedrock makes it even easier to build generative AI applications that use proprietary data to deliver customized, up-to-date responses

Customers have more options to customize models in Amazon Bedrock with fine-tuning support for Cohere Command, Meta Llama 2, and Amazon Titan models, with Anthropic Claude coming soon

With Agents for Amazon Bedrock, customers can enable generative AI applications to plan and perform a wide variety of multistep business tasks securely and privately

Guardrails for Amazon Bedrock helps customers implement safeguards customized to their generative AI applications and aligned with their responsible AI policies

Blueshift, dentsu, Druva, GoDaddy, INRIX, MongoDB, OfferUp, Salesforce, SmartBots AI, and TTEC Digital are among the customers and partners using Amazon Bedrock to harness generative AI

LAS VEGAS–(BUSINESS WIRE)–At AWS re:Invent, Amazon Web Services, Inc. (AWS), an Amazon.com, Inc. company (NASDAQ: AMZN), today announced Amazon Bedrock innovations that expand model choice and deliver powerful capabilities, making it easier for customers to build and scale generative artificial intelligence (AI) applications customized to their businesses. Amazon Bedrock is a fully managed service that offers easy access to a choice of industry-leading large language models and other foundation models from AI21 Labs, Anthropic, Cohere, Meta, Stability AI, and Amazon, along with a broad set of capabilities that customers need to build generative AI applications—simplifying development while supporting privacy and security. These announcements further democratize access to generative AI by empowering customers with even more choice of industry-leading models and new capabilities to evaluate them, simplifying how they customize models with relevant and proprietary data, supplying tools to automate the execution of complex tasks, and equipping customers with safeguards to build and deploy applications responsibly. Together, these new additions to Amazon Bedrock transform how organizations of all sizes and across all industries can use generative AI to spark innovation and reinvent customer experiences. To get started with Amazon Bedrock, visit aws.amazon.com/bedrock.

Generative AI is poised to be the most transformational technology of our time, and we are inspired by how customers are applying it to new opportunities and tackling business challenges,” said Dr. Swami Sivasubramanian, vice president of Data and AI at AWS. “As customers incorporate generative AI into their businesses, they turn to Amazon Bedrock for its choice of leading models, customization features, agent capabilities, and enterprise-grade security and privacy in a fully managed experience. With even more tools at their fingertips, customers are using Amazon Bedrock to leverage the full potential of generative AI to reimagine user experiences, reinvent their businesses, and accelerate their generative AI journeys.”

Organizations want to use generative AI for a wide variety of use cases—like generating productivity gains, driving innovative user experiences, and reimagining work—but generative AI is evolving rapidly, with new options and innovations happening daily. With so much fluidity in this space, customers’ ability to adapt is arguably the most valuable tool of all. Organizations need to be able to experiment, deploy, iterate, and pivot using the latest and greatest models available, and be ready to immediately embrace what comes tomorrow. To address these challenges, AWS developed Amazon Bedrock to make building with—and moving between—a range of models as easy as an API call, to put the latest techniques for model customization in the hands of all developers, and to keep customers secure and their data private. This is why customers such as Alida, Automation Anywhere, Blueshift, BMW Group, Clariant, Coinbase, Cox Automotive, dentsu, Druva, Genesys, Gilead, GoDaddy, Hellmann Worldwide Logistics, INRIX, KONE, LexisNexis Legal & Professional, Lonely Planet, NatWest, Nexxiot, OfferUp, Omnicom, the PGA TOUR, Proofpoint, Salesforce, Siemens, Takenaka Corporation, and Verint have turned to Amazon Bedrock to help them harness the power of generative AI for their organizations. Today’s announcement introduces new models and capabilities that will make it even easier for customers to build and scale generative AI applications.

The latest models from Anthropic, Cohere, Meta, and Stability AI, as well as additions to the Amazon Titan family, expand model choice for customers

No single model is ideal for every use case. Models vary across capabilities, price, and performance. Customers need easy access to a variety of model choices, so they can try out different models, switch between them, and combine the best models for their needs. With Amazon Bedrock, customers can drive rapid innovation with the latest versions of models, including the newly available Anthropic Claude 2.1 and Meta Llama 2 70B, and the recently available Cohere Command Light, Cohere Embed English, Cohere Embed multilingual, Meta Llama 2 13B, and Stability AI Stable Diffusion XL 1.0—all accessible via an API. In addition to Amazon Titan Text Embeddings and Amazon Titan Text models (now generally available), AWS is introducing Amazon Titan Image Generator and Amazon Titan Multimodal Embeddings to give customers even more choice and flexibility to build generative AI applications with models. Exclusive to Amazon Bedrock, Amazon Titan models are created and pre-trained by AWS on large and diverse datasets for a variety of use cases, with built-in support for the responsible use of AI. And Amazon indemnifies customers against claims that generally available Amazon Titan models or their outputs infringe on third-party copyrights.

  • Anthropic’s Claude 2.1 in Amazon Bedrock: Anthropic, an AI safety and research company that builds reliable, interpretable, and steerable AI systems, has brought Claude 2.1, the latest version of their language model, to Amazon Bedrock. Claude 2.1 offers a 200K token context window and improved accuracy over long documents. Customers can now process text heavy documents like financial statements and internal datasets, and Claude 2.1 can summarize, perform Q&A, or contrast documents, and much more. Anthropic reports that Claude 2.1 has made significant gains in honesty with a 2x decrease in false statements compared to their previous model.
  • Meta Llama 2 70B in Amazon Bedrock: Llama 2 is the next generation of language models by Meta. Llama 2 was trained on 40% more data than Llama 1 and has double the context length. The Llama 2 70 billion-parameter model is now available in Amazon Bedrock, in addition to the recently announced Llama 2 13 billion-parameter model. Built on top of the pre-trained Llama model, Llama 2 is optimized for dialog use cases through fine-tuning with instruction datasets and more than 1 million human annotations. The models perform competitively against multiple external benchmarks, including reasoning, coding, proficiency, and knowledge tests, and offer a compelling combination of price and performance in Amazon Bedrock.
  • New Amazon Titan Image Generator available in preview: Amazon Titan Image Generator helps customers in industries like advertising, ecommerce, and media and entertainment produce studio-quality, realistic images or enhance existing images using natural language prompts, for rapid ideation and iteration on large volumes of images and at low cost. The model can understand complex prompts and generate relevant images with accurate object composition and limited distortions, reducing the generation of harmful content and mitigating the spread of misinformation. Customers can use the model in the Amazon Bedrock console either by submitting a natural language prompt to generate an image or by uploading an image for automatic editing, before configuring the dimensions and specifying the number of variations the model should generate. To edit, customers can isolate parts of an image to add or replace details (e.g., inserting a surfboard into a beach scene or replacing mountains with a forest in the background of a car advertisement), or they can extend an image’s borders with additional details in the same style as the original. Building on the commitments AWS made earlier this year at the White House, Amazon Titan applies an invisible watermark to all images it generates to help reduce the spread of misinformation by providing a discreet mechanism to identify AI-generated images and to promote the safe, secure, and transparent development of AI technology. AWS is among the first model providers to widely release built-in invisible watermarks that are integrated into the image outputs and are designed to be resistant to alterations.
  • New Amazon Titan Multimodal Embeddings generally available: Amazon Titan Multimodal Embeddings helps customers power more accurate and contextually relevant multimodal search and recommendation experiences for end users. The model converts images and short text into embeddings—numerical representations that allow the model to easily understand semantic meanings and relationships among data— which are stored in a customer’s vector database. End users can submit search queries using any combination of image and text prompts. The model will generate embeddings for the search query and match them to the stored embeddings to produce more accurate and relevant search and recommendation results for end users. For example, a stock photography company with hundreds of millions of images can use the model to power its search functionality, so users can search for images using a phrase, image, or a combination of image and text (e.g., “show me images similar to the provided image, but with sunny skies”). By default, the model generates vectors that are well suited for search experiences that require a high degree of accuracy and speed. However, customers can also generate smaller dimensions to optimize for speed and performance. Amazon Titan Multimodal Embeddings joins the existing Amazon Titan Text Embeddings model, which is used to convert text input like single words, phrases, or even large documents into embeddings for use cases like search and personalization.

New capability helps customers efficiently evaluate, compare, and select the best model for their use case and business needs

Today, organizations have a wide range of model options to power their generative AI applications. To strike the right balance of accuracy and performance for their use case, organizations must efficiently compare models and find the best option based on their preferred metrics. To compare models, organizations must first spend days identifying benchmarks, setting up evaluation tools, and running assessments, all of which requires deep expertise in data science. Furthermore, these tests are not useful for evaluating subjective criteria (e.g., brand voice, relevance, and style) that requires judgment through tedious, time-intensive, human-review workflows. The time, expertise, and resources required for these comparisons—for every new use case—make it difficult for organizations to choose the optimal model for a task, limiting their use of generative AI.

Now available in preview, Model Evaluation on Amazon Bedrock helps customers evaluate, compare, and select the best models for their specific use case, using either automatic or human evaluations. In the Amazon Bedrock console, customers choose the models they want to compare for a given task, such as question-answering or content summarization. For automatic evaluations, customers select predefined evaluation criteria (e.g., accuracy, robustness, and toxicity) and upload their own testing dataset or select from built-in, publicly available datasets. For subjective criteria or nuanced content requiring sophisticated judgment, customers can easily set up human-based evaluation workflows with just a few clicks. These workflows leverage a customer’s in-house workforce, or use a managed workforce provided by AWS, to evaluate model responses. During human-based evaluations, customers define use case-specific metrics (e.g., relevance, style, and brand voice). Once customers finish the setup process, Amazon Bedrock runs evaluations and generates a report, so customers can easily understand how the model performed across key criteria and can make optimal tradeoffs and quickly select the best models for their use cases.

New model customization capabilities help customers make the most of their data, privately and securely, on AWS

Organizations want to maximize the value of their rich data sources to deliver remarkable user experiences—at scale—that are uniquely customized to reflect the company’s style, voice, and services. New, purpose-built capabilities available in Amazon Bedrock help customers personalize models privately and securely with their own data to build differentiated generative AI-powered applications.

  • Knowledge Bases for Amazon Bedrock customizes model responses with contextual and relevant company data: Organizations want to supplement existing models with proprietary data to create more relevant and accurate responses. To equip the model with up-to-date information, organizations turn to retrieval augmented generation (RAG), a technique that allows customers to customize a model’s responses by augmenting prompts with data from multiple sources, including document repositories, databases, and APIs. Now generally available, Knowledge Bases for Amazon Bedrock securely connects models to proprietary data sources for RAG to deliver more accurate, context-specific responses for use cases like chatbots and question-answering systems. Knowledge bases are fully managed, so customers simply point to the location of their data. Then knowledge bases fetch the text documents and save the data to a vector database or set one up on the customer’s behalf. When a user query comes in, Amazon Bedrock orchestrates RAG by fetching text needed to augment a prompt, sending the prompt to the model, and returning the response. Knowledge Bases for Amazon Bedrock supports databases with vector capabilities, including Amazon OpenSearch, and other popular databases like Pinecone and Redis Enterprise Cloud, with Amazon Aurora and MongoDB coming soon.
  • Cohere Command, Meta Llama 2, and Amazon Titan models can now be fine-tuned in Amazon Bedrock, with support for Anthropic’s Claude 2 coming soon: In addition to RAG, organizations can also leverage fine-tuning to further train the model on a specific task (e.g., text generation), using labeled datasets to adapt the model’s parameters to their business, and extending its knowledge with the lexicon and terminology used by the organization and end users. For example, a retail customer could fine-tune a model on a dataset of product descriptions to help it understand the brand style and produce more accurate descriptions for the website. Amazon Bedrock now supports fully managed fine-tuning for Cohere Command and Meta Llama 2, along with Amazon Titan Text Express, Amazon Titan Text Lite, Amazon Titan Multimodal Embeddings, and Amazon Titan Image Generator (in preview), so customers can use labeled datasets to increase model accuracy for specific tasks. Additionally, AWS customers will soon be able to fine-tune Claude 2’s performance with their data sources. To fine-tune a model, customers start by selecting the model and using Amazon Bedrock to make a copy. Customers then point to labeled examples in Amazon Simple Storage Service (Amazon S3). Amazon Bedrock incrementally trains the model (augments the copied model with the new information) on these examples, and the result is a private, more accurate fine-tuned model that delivers more relevant, customized responses. Customer data is encrypted in transit and at rest, so all valuable customer data remains secure and private. AWS and third-party model providers will not use any inputs or outputs from Amazon Bedrock to train their base models.

With Agents for Amazon Bedrock, generative AI applications can help execute multistep tasks using company systems and data sources

While models are effective at conversing and creating new content, they deliver more value if equipped to take actions, solve problems, and interact with a range of systems to complete multistep tasks (e.g., booking travel or ordering replacement parts). However, this requires custom integrations to connect models with company data sources, APIs, and internal and external systems. Developers must write code to orchestrate the interactions between models, systems, and the user, so the application can execute a series of API calls in a logical order. To connect the model with data sources, developers must implement RAG, so the model can customize its responses to the task. Finally, developers must provision and manage the requisite infrastructure, as well as establish policies for data security and privacy. These steps are time-consuming and require expertise, slowing the development of generative AI applications.

Now generally available, fully managed Agents for Amazon Bedrock enables generative AI applications to execute multistep tasks using company systems and data sources. Agents can plan and perform most business tasks, such as answering questions about product availability or taking orders. Customers can create an agent using a simple setup process, first selecting the desired model, writing a few instructions in natural language (e.g., “you are a cheerful customer service agent” and “check product availability in the inventory system”) and providing access to the company’s enterprise systems and knowledge bases. Agents automatically analyze the request and break it down into a logical sequence, using the model’s reasoning capabilities to determine the information needed. The agent then takes action by identifying the APIs to call and deciding when to call them to fulfill the request. Agents also retrieve needed information from proprietary data sources to provide accurate and relevant responses. Agents securely and privately perform this process in the background each time, relieving customers from having to engineer prompts, manage the session context, or orchestrate systems manually. With Agents for Amazon Bedrock, customers can improve the accuracy and speed of development of their generative AI applications.

With Guardrails for Amazon Bedrock, customers can implement safeguards across models based on application requirements and responsible AI policies

Organizations recognize the need to manage interactions within generative AI applications for a relevant and safe user experience. While many models use built-in controls to filter undesirable and harmful content, organizations want to further customize interactions to remain on topics relevant to their business, align with company policies, and adhere to responsible AI principles. For example, a bank might want to configure its online assistant to refrain from providing investment advice, avoid queries about competitors, and limit harmful content. As another example, after a customer service call, personally identifiable information (PII) may need to be redacted from the call summary. Organizations may need to change models, use multiple models, or replicate policies across applications, and they want a simple way to consistently deploy their preferences across all these areas simultaneously. Deep expertise is required to build custom protection systems with these kinds of safeguards and integrate them into applications, and the processes can take months. Organizations want a streamlined way to enforce key policies and rules in generative AI applications to deliver relevant user experiences and support safer use of the technology.

Now available in preview, Guardrails for Amazon Bedrock empowers customers to implement safeguards for generative AI applications that are customized to their use cases and responsible AI principles, enhancing the safety and privacy of user interactions. Guardrails drive consistency in how models in Amazon Bedrock respond to undesirable and harmful content within applications. Customers can apply guardrails to all large language models in Amazon Bedrock, as well as to fine-tuned models and in combination with Agents for Amazon Bedrock. To create a guardrail in the Amazon Bedrock console, customers start with natural language descriptions to define the denied topics within the context of their application. Customers can also configure thresholds across hate speech, insults, sexualized language, and violence to filter out harmful content to their desired level. In early 2024, customers will also be able to redact PII in models’ responses, set profanity filters, and provide a list of custom words to block interactions between users and models. Guardrails automatically evaluate both user queries and model responses to detect and help prevent content that falls into restricted categories. Customers can create multiple guardrails to support different use cases and apply the same guardrails across multiple models. Guardrails for Amazon Bedrock empowers customers to innovate safely by providing a consistent user experience and standardizing safety and privacy controls across generative AI applications.

Blueshift provides brands with marketing automation and customer data platforms to deliver personalized, customer engagement across all communication channels and devices. “Product catalogs are rapidly evolving with new content being changed every minute, and we need to continuously update our embeddings to ensure recommendations for brand audiences remain relevant,” said Manyam Mallela, co-founder and chief AI officer at Blueshift. “Amazon Titan Multimodal Embeddings in Amazon Bedrock is outperforming older models from other providers that we used, offering more nuanced and contextually relevant recommendations without complex feature engineering. Our team has seen a 10% performance improvement using Amazon Titan Multimodal Embeddings. With the robust infrastructure, security, and collaboration offered by AWS, Blueshift is poised to seamlessly integrate cutting-edge embeddings, ensuring that our recommendation solutions remain state-of-the-art and lead to improved audience engagement.”

Dentsu is one of the world’s largest providers of integrated marketing and technology services. “We work at the convergence of marketing, technology, and consulting to drive people-centered transformations for brands that want to shape society for the better, and generative AI is changing our ability to deliver at scale and speed for clients, augmenting, not replacing, our 72,000-strong team around the world,” said Brian Klochkoff, executive vice president of Innovation & Emerging Technologies at dentsu. “Specifically, Amazon Bedrock gives us the enterprise control and ease-of-use to deploy third-party models for decentralized usage across our product and engineering teams. This allows our teams to innovate with the latest and greatest generative AI advancements in a safe and responsible space, while inventing cutting-edge opportunities for clients.

Contacts

Amazon.com, Inc.

Media Hotline

Amazon-pr@amazon.com
www.amazon.com/pr

Read full story here

Staff

Recent Posts

Digital Matter Completes Ubilite UBI206 Evaluation and Issues a Purchase Order for the UBI206 Ultra Low Power Wi-Fi Chips

SAN DIEGO--(BUSINESS WIRE)--Ubilite Inc. (“Ubilite”) announced UBI206 was selected by Digital Matter, Perth, AUS for…

3 days ago

Caesars Sportsbook Accepts Legalized Mobile Sports Wagers on Eastern Band of Cherokee Indians Tribal Lands in North Carolina; First Within State Borders

Wagering is now exclusively live via the Caesars Sportsbook app at Harrah’s Cherokee Casino Resort,…

3 days ago

Hyatt Announces That the Rio Hotel & Casino, Las Vegas has Joined World of Hyatt Amid Multi-Million-Dollar Renovation

Property-wide, multi-phase transformation is underway at the iconic 2,500-room hotel as it expands Hyatt’s brand…

3 days ago

Brown-Forman Commits Funds to Foundation and Dendrifund

LOUISVILLE, Ky.--(BUSINESS WIRE)--Brown-Forman Corporation (NYSE:BFA, BFB) today announced a commitment of $22.5 million to support…

3 days ago

Brown-Forman Announces Launch of Distribution Business in Italy

LOUISVILLE, Ky.--(BUSINESS WIRE)--Brown-Forman Corporation (NYSE:BFA,BFB), one of the largest American-owned spirits companies, announces today its…

3 days ago

Endeavor Announces Q1 2024 Cash Dividend

BEVERLY HILLS, Calif.--(BUSINESS WIRE)--Endeavor Group Holdings, Inc. (NYSE: EDR) (“Endeavor” or the “Company”), a global…

3 days ago