标签: google-generativeai

从 Vertex AI Model Garden 对 Llama-2 进行推理

llama2-chat-13b我从模型花园部署。但是,我在尝试执行推理时遇到错误。

配置:

project="X";
endpoint_id="Y";
location="us-east1";
64 VCPUs, 57.6 GB RAM;
GPU= 4 T4;
Run Code Online (Sandbox Code Playgroud)

我尝试了三种方法,但它们都返回某种错误:

方法一:

from typing import Dict, List, Union

from google.cloud import aiplatform
from google.protobuf import json_format
from google.protobuf.struct_pb2 import Value


def predict_custom_trained_model_sample(
    project: str,
    endpoint_id: str,
    instances: Union[Dict, List[Dict]],
    location: str = "us-east1",
    api_endpoint: str = "us-east1-aiplatform.googleapis.com",
):
    """
    `instances` can be either single instance of type dict or a list
    of instances.
    """
    # The AI Platform services require regional API endpoints.
    client_options = {"api_endpoint": …
Run Code Online (Sandbox Code Playgroud)

google-cloud-platform google-cloud-vertex-ai google-generativeai llama

4
推荐指数
1
解决办法
370
查看次数