llama2-chat-13b我从模型花园部署。但是,我在尝试执行推理时遇到错误。
配置:
project="X";
endpoint_id="Y";
location="us-east1";
64 VCPUs, 57.6 GB RAM;
GPU= 4 T4;
Run Code Online (Sandbox Code Playgroud)
我尝试了三种方法,但它们都返回某种错误:
方法一:
from typing import Dict, List, Union
from google.cloud import aiplatform
from google.protobuf import json_format
from google.protobuf.struct_pb2 import Value
def predict_custom_trained_model_sample(
project: str,
endpoint_id: str,
instances: Union[Dict, List[Dict]],
location: str = "us-east1",
api_endpoint: str = "us-east1-aiplatform.googleapis.com",
):
"""
`instances` can be either single instance of type dict or a list
of instances.
"""
# The AI Platform services require regional API endpoints.
client_options = {"api_endpoint": …Run Code Online (Sandbox Code Playgroud) google-cloud-platform google-cloud-vertex-ai google-generativeai llama