update readme
This commit is contained in:
1
.gitignore
vendored
1
.gitignore
vendored
@@ -4,3 +4,4 @@ weights/icon_detect/
|
|||||||
weights/icon_detect_v1_5/
|
weights/icon_detect_v1_5/
|
||||||
.gradio
|
.gradio
|
||||||
__pycache__/
|
__pycache__/
|
||||||
|
debug.ipynb
|
||||||
|
|||||||
@@ -12,7 +12,7 @@
|
|||||||
**OmniParser** is a comprehensive method for parsing user interface screenshots into structured and easy-to-understand elements, which significantly enhances the ability of GPT-4V to generate actions that can be accurately grounded in the corresponding regions of the interface.
|
**OmniParser** is a comprehensive method for parsing user interface screenshots into structured and easy-to-understand elements, which significantly enhances the ability of GPT-4V to generate actions that can be accurately grounded in the corresponding regions of the interface.
|
||||||
|
|
||||||
## News
|
## News
|
||||||
- [2024/11/26] We release an updated version, OmniParser V1.5 which features more fine grained/small icon detection. Examples in the demo.ipynb.
|
- [2024/11/26] We release an updated version, OmniParser V1.5 which features 1) more fine grained/small icon detection, 2) prediction of whether each screen element is interactable or not. Examples in the demo.ipynb.
|
||||||
- [2024/10] OmniParser was the #1 trending model on huggingface model hub (starting 10/29/2024).
|
- [2024/10] OmniParser was the #1 trending model on huggingface model hub (starting 10/29/2024).
|
||||||
- [2024/10] Feel free to checkout our demo on [huggingface space](https://huggingface.co/spaces/microsoft/OmniParser)! (stay tuned for OmniParser + Claude Computer Use)
|
- [2024/10] Feel free to checkout our demo on [huggingface space](https://huggingface.co/spaces/microsoft/OmniParser)! (stay tuned for OmniParser + Claude Computer Use)
|
||||||
- [2024/10] Both Interactive Region Detection Model and Icon functional description model are released! [Hugginface models](https://huggingface.co/microsoft/OmniParser)
|
- [2024/10] Both Interactive Region Detection Model and Icon functional description model are released! [Hugginface models](https://huggingface.co/microsoft/OmniParser)
|
||||||
|
|||||||
16
demo.ipynb
16
demo.ipynb
@@ -2,7 +2,7 @@
|
|||||||
"cells": [
|
"cells": [
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 12,
|
"execution_count": 1,
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [
|
"outputs": [
|
||||||
{
|
{
|
||||||
@@ -28,7 +28,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 3,
|
"execution_count": 2,
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [
|
"outputs": [
|
||||||
{
|
{
|
||||||
@@ -75,7 +75,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 15,
|
"execution_count": 3,
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [
|
"outputs": [
|
||||||
{
|
{
|
||||||
@@ -84,8 +84,8 @@
|
|||||||
"text": [
|
"text": [
|
||||||
"image size: (1919, 1079)\n",
|
"image size: (1919, 1079)\n",
|
||||||
"\n",
|
"\n",
|
||||||
"image 1/1 /home/yadonglu/OmniParser/imgs/word.png: 736x1280 138 0s, 7.5ms\n",
|
"image 1/1 /home/yadonglu/OmniParser/imgs/word.png: 736x1280 138 0s, 61.1ms\n",
|
||||||
"Speed: 5.4ms preprocess, 7.5ms inference, 5.1ms postprocess per image at shape (1, 3, 736, 1280)\n"
|
"Speed: 9.0ms preprocess, 61.1ms inference, 432.6ms postprocess per image at shape (1, 3, 736, 1280)\n"
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
],
|
],
|
||||||
@@ -129,16 +129,16 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 16,
|
"execution_count": 4,
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [
|
"outputs": [
|
||||||
{
|
{
|
||||||
"data": {
|
"data": {
|
||||||
"text/plain": [
|
"text/plain": [
|
||||||
"<matplotlib.image.AxesImage at 0x7f64287ae000>"
|
"<matplotlib.image.AxesImage at 0x7ff76a9f3110>"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
"execution_count": 16,
|
"execution_count": 4,
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"output_type": "execute_result"
|
"output_type": "execute_result"
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ if args.version == 'v1':
|
|||||||
torch.save({'model':model}, 'weights/icon_detect/best.pt')
|
torch.save({'model':model}, 'weights/icon_detect/best.pt')
|
||||||
elif args.version == 'v1_5':
|
elif args.version == 'v1_5':
|
||||||
print("Converting v1_5")
|
print("Converting v1_5")
|
||||||
tensor_dict = torch.load("weights/icon_detect_v1_5/model.safetensors")
|
tensor_dict = load_file("weights/icon_detect_v1_5/model.safetensors")
|
||||||
model = DetectionModel('weights/icon_detect_v1_5/model.yaml')
|
model = DetectionModel('weights/icon_detect_v1_5/model.yaml')
|
||||||
model.load_state_dict(tensor_dict)
|
model.load_state_dict(tensor_dict)
|
||||||
save_dict = {'model':model}
|
save_dict = {'model':model}
|
||||||
|
|||||||
Reference in New Issue
Block a user