iamtdb

joined 11 months ago
 

TL;DR: We’re looking to deploy a LLaMA 2 13B model on an AWS Service and want to know the best option. AWS Bedrock is not preferred as we need full control over the model.

Hello everyone,

The company I work for is exploring options to host a LLaMA 2 13B model on an AWS Service, but we’re unsure of the best choice. Our priorities are data privacy and maximum control over data processing. This is solely for internal use.

We’ve tried AWS Bedrock, but have decided against it as it doesn’t provide us with complete control over the model.

Currently, we’re testing SageMaker and considering other options like an EC2 instance, but we’re uncertain about the most effective and economical solution.

I would appreciate hearing about your experiences.

Thanks in advance.

 

Hi all,

I am running a LLaMA 2 7b model on a AWS Sagemaker instance. I need the model just for providing me summaries of long documents, and i am using Langchain to do a map reduce on the data and get a summary of it.

I want to know if there's a better way to do this or if you could share your personal experiences on summarizing efficiently.

I am not getting good results since the summarization includes too much information.

Thanks in advance.