зеркало из
https://github.com/ssciwr/AMMICO.git
synced 2025-10-30 05:26:05 +02:00
* add image summary notebook * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * pin deepface version to avoid bug with progress bar after update * update actions version for checkout and python * test ci without lavis * no lavis for ci test * merging * return lavis * change lavis to salesforce-lavis * change pycocotools install method * change pycocotools install method * fix_pycocotools * Downgrade Python * back to 3.9 and remove pycocotools dependance * instrucctions for windows * missing comma after merge * lavis only for ubuntu * use lavis package name in install instead of git * adding multimodal searching py and notebook * exclude lavis on windows * skip import on windows * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * reactivate lavis * Revert "reactivate lavis" This reverts commit ecdaf9d316e4b08816ba62da5e0482c8ff15b14e. * Change input format for multimodal search * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fix clip models * account for new interface in init imports * changed imports bec of lavis/windows * fix if-else, added clip ViT-L-14=336 model * fix code smells * add model change function to summary * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fixed new model in summary.py * fixed summary windget * moved some function to utils * fixed imort torch in utils * added test_summary.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fixed opencv version * added first test of multimodal_search.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fixed test * removed windows in CI and added test in multimodal search * change lavis from dependencies from pip ro git * fixed blip2 model in test_multimodal_search.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fixed test multimodal search on cpu and gpu machines * added test, fixed dependencies * add -vv to pytest command in CI * added test_multimodal_search tests * fixed tests in test_multimodal_search.py * fixed tests in test_summary * changed CI and fixed test_multimodel search * fixed ci * fixed error in test multimodal search, changed ci * added multimodal search test, added windows CI, added picture in test data * CI debuging * fixing tests in CI * fixing test in CI 2 * fixing CI 3 * fixing CI * added filtering function * Brought back all tests after CI fixing * changed CI one pytest by individual tests * fixed opencv problem * fix path for text, adjust result for new gcv * remove opencv * fixing cv2 error * added opencv-contrib, change objects_cvlib * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * fixing tests in CI * fixing CI testing * cleanup objects * fixing codecov in CI * fixing codecov in CI * run tests together; install opencv last * update requirements for opencv dependencies * moved lavis functions from utils to summary * Remove lavis from utils.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * add missing jupyter --------- Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: iulusoy <inga.ulusoy@uni-heidelberg.de>
167 строки
5.7 KiB
Python
167 строки
5.7 KiB
Python
import os
|
|
from torch import device, cuda
|
|
from lavis.models import load_model_and_preprocess
|
|
import misinformation.summary as sm
|
|
|
|
images = [
|
|
"./test/data/d755771b-225e-432f-802e-fb8dc850fff7.png",
|
|
"./test/data/IMG_2746.png",
|
|
"./test/data/IMG_2750.png",
|
|
"./test/data/IMG_2805.png",
|
|
"./test/data/IMG_2806.png",
|
|
"./test/data/IMG_2807.png",
|
|
"./test/data/IMG_2808.png",
|
|
"./test/data/IMG_2809.png",
|
|
"./test/data/IMG_3755.jpg",
|
|
"./test/data/IMG_3756.jpg",
|
|
"./test/data/IMG_3757.jpg",
|
|
"./test/data/pic1.png",
|
|
]
|
|
|
|
|
|
def test_analyse_image():
|
|
mydict = {}
|
|
for img_path in images:
|
|
id_ = os.path.splitext(os.path.basename(img_path))[0]
|
|
mydict[id_] = {"filename": img_path}
|
|
|
|
for key in mydict:
|
|
mydict[key] = sm.SummaryDetector(mydict[key]).analyse_image()
|
|
keys = list(mydict.keys())
|
|
assert len(mydict) == 12
|
|
for key in keys:
|
|
assert len(mydict[key]["3_non-deterministic summary"]) == 3
|
|
|
|
const_image_summary_list = [
|
|
"a river running through a city next to tall buildings",
|
|
"a crowd of people standing on top of a tennis court",
|
|
"a crowd of people standing on top of a field",
|
|
"a room with a desk and a chair",
|
|
"a table with plastic containers on top of it",
|
|
"a view of a city with mountains in the background",
|
|
"a view of a city street from a window",
|
|
"a busy city street with cars and pedestrians",
|
|
"a close up of an open book with writing on it",
|
|
"a book that is open on a table",
|
|
"a yellow book with green lettering on it",
|
|
"a person running on a beach near a rock formation",
|
|
]
|
|
|
|
for i in range(len(const_image_summary_list)):
|
|
assert mydict[keys[i]]["const_image_summary"] == const_image_summary_list[i]
|
|
|
|
del sm.SummaryDetector.summary_model, sm.SummaryDetector.summary_vis_processors
|
|
cuda.empty_cache()
|
|
|
|
summary_device = device("cuda" if cuda.is_available() else "cpu")
|
|
summary_model, summary_vis_processors, _ = load_model_and_preprocess(
|
|
name="blip_caption",
|
|
model_type="base_coco",
|
|
is_eval=True,
|
|
device=summary_device,
|
|
)
|
|
|
|
for key in mydict:
|
|
mydict[key] = sm.SummaryDetector(mydict[key]).analyse_image(
|
|
summary_model, summary_vis_processors
|
|
)
|
|
keys = list(mydict.keys())
|
|
|
|
assert len(mydict) == 12
|
|
for key in keys:
|
|
assert len(mydict[key]["3_non-deterministic summary"]) == 3
|
|
|
|
const_image_summary_list2 = [
|
|
"a river running through a city next to tall buildings",
|
|
"a crowd of people standing on top of a tennis court",
|
|
"a crowd of people standing on top of a field",
|
|
"a room with a desk and a chair",
|
|
"a table with plastic containers on top of it",
|
|
"a view of a city with mountains in the background",
|
|
"a view of a city street from a window",
|
|
"a busy city street with cars and pedestrians",
|
|
"a close up of an open book with writing on it",
|
|
"a book that is open on a table",
|
|
"a yellow book with green lettering on it",
|
|
"a person running on a beach near a rock formation",
|
|
]
|
|
|
|
for i in range(len(const_image_summary_list2)):
|
|
assert mydict[keys[i]]["const_image_summary"] == const_image_summary_list2[i]
|
|
|
|
del summary_model, summary_vis_processors
|
|
cuda.empty_cache()
|
|
|
|
summary_model, summary_vis_processors, _ = load_model_and_preprocess(
|
|
name="blip_caption",
|
|
model_type="large_coco",
|
|
is_eval=True,
|
|
device=summary_device,
|
|
)
|
|
|
|
for key in mydict:
|
|
mydict[key] = sm.SummaryDetector(mydict[key]).analyse_image(
|
|
summary_model, summary_vis_processors
|
|
)
|
|
keys = list(mydict.keys())
|
|
assert len(mydict) == 12
|
|
for key in keys:
|
|
assert len(mydict[key]["3_non-deterministic summary"]) == 3
|
|
|
|
const_image_summary_list3 = [
|
|
"a river running through a town next to tall buildings",
|
|
"a crowd of people standing on top of a track",
|
|
"a group of people standing on top of a track",
|
|
"a desk and chair in a small room",
|
|
"a table that has some chairs on top of it",
|
|
"a view of a city from a window of a building",
|
|
"a view of a city from a window",
|
|
"a city street filled with lots of traffic",
|
|
"an open book with german text on it",
|
|
"a close up of a book on a table",
|
|
"a book with a green cover on a table",
|
|
"a person running on a beach near the ocean",
|
|
]
|
|
|
|
for i in range(len(const_image_summary_list2)):
|
|
assert mydict[keys[i]]["const_image_summary"] == const_image_summary_list3[i]
|
|
|
|
|
|
def test_analyse_questions():
|
|
mydict = {}
|
|
for img_path in images:
|
|
id_ = os.path.splitext(os.path.basename(img_path))[0]
|
|
mydict[id_] = {"filename": img_path}
|
|
|
|
list_of_questions = [
|
|
"How many persons on the picture?",
|
|
"What happends on the picture?",
|
|
]
|
|
for key in mydict:
|
|
mydict[key] = sm.SummaryDetector(mydict[key]).analyse_questions(
|
|
list_of_questions
|
|
)
|
|
|
|
keys = list(mydict.keys())
|
|
assert len(mydict) == 12
|
|
|
|
list_of_questions_ans = [2, 100, "many", 0, 0, "none", "two", 5, 0, 0, 0, 1]
|
|
|
|
list_of_questions_ans2 = [
|
|
"flood",
|
|
"festival",
|
|
"people are flying kites",
|
|
"no one's home",
|
|
"chair is being moved",
|
|
"traffic jam",
|
|
"day time",
|
|
"traffic jam",
|
|
"nothing",
|
|
"nothing",
|
|
"nothing",
|
|
"running",
|
|
]
|
|
|
|
for i in range(len(list_of_questions_ans)):
|
|
assert mydict[keys[i]][list_of_questions[1]] == str(list_of_questions_ans2[i])
|