Vision2Seq Backend
Vision2SeqBackend
Bases: Backend
Vision2SeqBackend backend that runs locally to generate robot actions.
Beware of the memory requirements of 7B+ parameter models like OpenVLA.
Attributes:
Name | Type | Description |
---|---|---|
model_id |
str
|
The model to use for the OpenVLA backend. |
device |
device
|
The device to run the model on. |
torch_dtype |
dtype
|
The torch data type to use. |
processor |
AutoProcessor
|
The processor for the model. |
model |
AutoModelForVision2Seq
|
The model for the OpenVLA backend. |
Source code in mbodied/agents/backends/vision2seq_backend.py
29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 |
|