respect🫡
Collection
Everything about the paper "Retrospective Learning from Interactions" • 4 items • Updated
How to use lil-lab/respect with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("image-text-to-text", model="lil-lab/respect") # Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("lil-lab/respect", dtype="auto")How to use lil-lab/respect with vLLM:
# Install vLLM from pip:
pip install vllm
# Start the vLLM server:
vllm serve "lil-lab/respect"
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:8000/v1/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "lil-lab/respect",
"prompt": "Once upon a time,",
"max_tokens": 512,
"temperature": 0.5
}'docker model run hf.co/lil-lab/respect
How to use lil-lab/respect with SGLang:
# Install SGLang from pip:
pip install sglang
# Start the SGLang server:
python3 -m sglang.launch_server \
--model-path "lil-lab/respect" \
--host 0.0.0.0 \
--port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "lil-lab/respect",
"prompt": "Once upon a time,",
"max_tokens": 512,
"temperature": 0.5
}'docker run --gpus all \
--shm-size 32g \
-p 30000:30000 \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--env "HF_TOKEN=<secret>" \
--ipc=host \
lmsysorg/sglang:latest \
python3 -m sglang.launch_server \
--model-path "lil-lab/respect" \
--host 0.0.0.0 \
--port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "lil-lab/respect",
"prompt": "Once upon a time,",
"max_tokens": 512,
"temperature": 0.5
}'How to use lil-lab/respect with Docker Model Runner:
docker model run hf.co/lil-lab/respect
This repository contains the lil-lab/respect model, based on the ACL paper Retrospective Learning from Interactions. For more resources, please see https://lil-lab.github.io/respect and https://github.com/lil-lab/respect.
To get started with the model, follow these steps:
Prepare your conda environment:
conda create -n respect python=3.9.18
pip install -r requirements.txt
pip install -e .
from datasets import load_dataset
ds = load_dataset("lil-lab/respect", name="turn", split="train")
Download checkpoints and load the model using transformers and peft:
import torch
from transformers import Idefics2ForConditionalGeneration
from peft import PeftModel
checkpoint = "HuggingFaceM4/idefics2-8b"
model_id = 'lil-lab/respect'
model = Idefics2ForConditionalGeneration.from_pretrained(
checkpoint, torch_dtype=torch.bfloat16)
peft_model = PeftModel.from_pretrained(
model, model_id, adapter_name="r6_bp", revision="r6_bp")
To generate plots from the paper, run analysis/plots.ipynb in the GitHub repository.
Base model
HuggingFaceM4/idefics2-8b