Import customized fashions in Amazon Bedrock (preview)

Import customized fashions in Amazon Bedrock (preview)
Import customized fashions in Amazon Bedrock (preview)


Voiced by Polly

With Amazon Bedrock, you will have entry to a selection of high-performing basis fashions (FMs) from main synthetic intelligence (AI) corporations that make it simpler to construct and scale generative AI functions. A few of these fashions present publicly out there weights that may be fine-tuned and customised for particular use instances. Nevertheless, deploying custom-made FMs in a safe and scalable approach will not be a straightforward activity.

Beginning at present, Amazon Bedrock provides in preview the aptitude to import customized weights for supported mannequin architectures (similar to Meta Llama 2, Llama 3, and Mistral) and serve the customized mannequin utilizing On-Demand mode. You’ll be able to import fashions with weights in Hugging Face safetensors format from Amazon SageMaker and Amazon Simple Storage Service (Amazon S3).

On this approach, you should use Amazon Bedrock with present custom-made fashions similar to Code Llama, a code-specialized model of Llama 2 that was created by additional coaching Llama 2 on code-specific datasets, or use your knowledge to fine-tune fashions in your personal distinctive enterprise case and import the ensuing mannequin in Amazon Bedrock.

Let’s see how this works in follow.

Bringing a customized mannequin to Amazon Bedrock
Within the Amazon Bedrock console, I select Imported fashions from the Basis fashions part of the navigation pane. Now, I can create a customized mannequin by importing mannequin weights from an Amazon Simple Storage Service (Amazon S3) bucket or from an Amazon SageMaker mannequin.

I select to import mannequin weights from an S3 bucket. In one other browser tab, I obtain the MistralLite mannequin from the Hugging Face web site utilizing this pull request (PR) that gives weights in safetensors format. The pull request is at the moment Able to merge, so it could be a part of the primary department while you learn this. MistralLite is a fine-tuned Mistral-7B-v0.1 language mannequin with enhanced capabilities of processing lengthy context as much as 32K tokens.

When the obtain is full, I add the information to an S3 bucket in the identical AWS Region the place I’ll import the mannequin. Listed below are the MistralLite mannequin information within the Amazon S3 console:

Console screenshot.

Again on the Amazon Bedrock console, I enter a reputation for the mannequin and preserve the proposed import job identify.

Console screenshot.

I choose Mannequin weights within the Mannequin import settings and browse S3 to decide on the situation the place I uploaded the mannequin weights.

Console screenshot.

To authorize Amazon Bedrock to entry the information on the S3 bucket, I choose the choice to create and use a brand new AWS Identity and Access Management (IAM) service position. I exploit the View permissions particulars hyperlink to examine what can be within the position. Then, I submit the job.

About ten minutes later, the import job is accomplished.

Console screenshot.

Now, I see the imported mannequin within the console. The record additionally reveals the mannequin Amazon Resource Name (ARN) and the creation date.

Console screenshot.

I select the mannequin to get extra data, such because the S3 location of the mannequin information.

Console screenshot.

Within the mannequin element web page, I select Open in playground to check the mannequin within the console. Within the textual content playground, I kind a query utilizing the immediate template of the mannequin:

<|prompter|>What are the primary challenges to assist an extended context for LLM?</s><|assistant|>

The MistralLite imported mannequin is fast to answer and describe a few of these challenges.

Console screenshot.

Within the playground, I can tune responses for my use case utilizing configurations similar to temperature and most size or add cease sequences particular to the imported mannequin.

To see the syntax of the API request, I select the three small vertical dots on the prime proper of the playground.

Console screenshot.

I select View API syntax and run the command utilizing the AWS Command Line Interface (AWS CLI):

aws bedrock-runtime invoke-model 
--model-id arn:aws:bedrock:us-east-1:123412341234:imported-model/a82bkefgp20f 
--body ">","max_tokens":512,"top_k":200,"top_p":0.9,"cease":[],"temperature":0.5" 
--cli-binary-format raw-in-base64-out 
--region us-east-1 
invoke-model-output.txt

The output is much like what I obtained within the playground. As you may see, for imported fashions, the mannequin ID is the ARN of the imported mannequin. I can use the mannequin ID to invoke the imported mannequin with the AWS CLI and AWS SDKs.

Issues to know
You’ll be able to convey your personal weights for supported mannequin architectures to Amazon Bedrock within the US East (N. Virginia) AWS Region. The mannequin import functionality is at the moment out there in preview.

When utilizing customized weights, Amazon Bedrock serves the mannequin with On-Demand mode, and also you solely pay for what you employ with no time-based time period commitments. For detailed data, see Amazon Bedrock pricing.

The power to import fashions is managed utilizing AWS Identity and Access Management (IAM), and you may enable this functionality solely to the roles in your group that have to have it.

With this launch, it’s now simpler to construct and scale generative AI functions utilizing customized fashions with safety and privateness in-built.

To be taught extra:

Danilo



Leave a Reply

Your email address will not be published. Required fields are marked *