開始使用 MedGemma 和 MedSigLIP API
先進的醫療文字和多模態 AI 模型
存取 MedGemma 強大的醫療文字分析、影像理解和臨床決策支援功能。
用於醫療影像和文字任務的輕量級模型
40 億參數專門用於醫療文字的大型語言模型
270 億參數用於複雜醫療任務的先進多模態模型
270 億參數用於分類和檢索的醫療影像文字編碼器
利用 MedSigLIP 高效的雙塔架構進行醫療影像分類、零樣本推理和語義檢索。
選擇最適合您需求的部署方法
使用 Hugging Face transformers 在本地執行模型
在 Google Cloud 上部署為 REST API 端點
使用 Vertex AI 批次作業處理大型資料集
Get started with implementation examples
from transformers import pipeline
# Load MedGemma model
pipe = pipeline(
"text-generation",
model="google/medgemma-27b-text-it",
torch_dtype="bfloat16",
device="cuda"
)
# Generate medical text
response = pipe(
"What are the symptoms of diabetes?",
max_length=200,
temperature=0.7
)
print(response[0]['generated_text'])
import requests
import json
# Vertex AI endpoint
endpoint_url = "https://your-endpoint.googleapis.com/v1/projects/your-project/locations/us-central1/endpoints/your-endpoint:predict"
# Request payload
payload = {
"instances": [{
"prompt": "What are the symptoms of diabetes?",
"max_tokens": 200,
"temperature": 0.7
}]
}
# Make API request
headers = {
"Authorization": "Bearer YOUR_ACCESS_TOKEN",
"Content-Type": "application/json"
}
response = requests.post(endpoint_url, json=payload, headers=headers)
result = response.json()
print(result['predictions'][0]['generated_text'])
from transformers import AutoModel, AutoProcessor
import torch
from PIL import Image
# Load MedSigLIP model
model = AutoModel.from_pretrained("google/medsiglip")
processor = AutoProcessor.from_pretrained("google/medsiglip")
# Load and process image
image = Image.open("medical_image.jpg")
text = "chest x-ray showing pneumonia"
# Process inputs
inputs = processor(
text=[text],
images=[image],
return_tensors="pt",
padding=True
)
# Get embeddings
with torch.no_grad():
outputs = model(**inputs)
image_embeds = outputs.image_embeds
text_embeds = outputs.text_embeds
# Calculate similarity
similarity = torch.cosine_similarity(image_embeds, text_embeds)
print(f"Similarity score: {similarity.item():.4f}")
import torch
from transformers import AutoModel, AutoProcessor
from PIL import Image
# Load model and processor
model = AutoModel.from_pretrained("google/medsiglip")
processor = AutoProcessor.from_pretrained("google/medsiglip")
# Define classification labels
labels = [
"normal chest x-ray",
"pneumonia chest x-ray",
"covid-19 chest x-ray",
"lung cancer chest x-ray"
]
# Load image
image = Image.open("chest_xray.jpg")
# Process inputs
inputs = processor(
text=labels,
images=[image] * len(labels),
return_tensors="pt",
padding=True
)
# Get predictions
with torch.no_grad():
outputs = model(**inputs)
logits = outputs.logits_per_image
probs = torch.softmax(logits, dim=-1)
# Get top prediction
top_idx = torch.argmax(probs, dim=-1)
confidence = probs[0, top_idx].item()
print(f"Prediction: {labels[top_idx]}")
print(f"Confidence: {confidence:.4f}")
完整的文件和範例