You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+8-2Lines changed: 8 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -20,7 +20,7 @@ ome
20
20
21
21
## Key Features
22
22
-**Compatibility**: Designed for various multimodal models.
23
-
-**Integration**: Currently integrated with **GPT-4o, o1, Gemini Pro Vision, Claude 3, Qwen-VL and LLaVa.**
23
+
-**Integration**: Currently integrated with **GPT-4o, GPT-4.1, o1, Gemini Pro Vision, Claude 3, Qwen-VL and LLaVa.**
24
24
-**Future Plans**: Support for additional models.
25
25
26
26
## Demo
@@ -54,12 +54,18 @@ operate
54
54
55
55
#### OpenAI models
56
56
57
-
The default model for the project is gpt-4o which you can use by simply typing `operate`. To try running OpenAI's new `o1` model, use the command below.
57
+
The default model for the project is gpt-4o which you can use by simply typing `operate`. To try running OpenAI's new `o1` model, use the command below.
58
58
59
59
```
60
60
operate -m o1-with-ocr
61
61
```
62
62
63
+
To experiment with OpenAI's latest `gpt-4.1` model, run:
64
+
65
+
```
66
+
operate -m gpt-4.1-with-ocr
67
+
```
68
+
63
69
64
70
### Multimodal Models `-m`
65
71
Try Google's `gemini-pro-vision` by following the instructions below. Start `operate` with the Gemini model
0 commit comments