зеркало из
https://github.com/ssciwr/AMMICO.git
synced 2025-10-28 20:54:14 +02:00
minor changes for revision 2 (#212)
* fix typos * add buttons for google colab everywhere * update readme, separate out FAQ * add privacy disclosure statement * do not install using uv * update docs notebook * explicit install of libopenblas * explicit install of libopenblas * explicit install of libopenblas * try to get scipy installed using uv * use ubuntu 24.04 * go back to pip * try with scipy only * try with a few others * use hatchling * wording changes, install all requirements * fix offending spacy version * run all tests * include faq in documentation, fix link * make readme links point to documentation * load model safely * correct edit on GH link and bump version * remove comments
Этот коммит содержится в:
родитель
174054f465
Коммит
65531c6204
4
FAQ.md
4
FAQ.md
@ -98,7 +98,9 @@ Some features of ammico require internet access; a general answer to this questi
|
||||
Due to well documented biases in the detection of minorities with computer vision tools, and to the ethical implications of such detection, these parts of the tool are not directly made available to users. To access these capabilities, users must first agree with a ethical disclosure statement that reads:
|
||||
|
||||
"DeepFace and RetinaFace provide wrappers to trained models in face recognition and emotion detection. Age, gender and race/ethnicity models were trained on the backbone of VGG-Face with transfer learning.
|
||||
ETHICAL DISCLOSURE STATEMENT:
|
||||
|
||||
ETHICAL DISCLOSURE STATEMENT:
|
||||
|
||||
The Emotion Detector uses DeepFace and RetinaFace to probabilistically assess the gender, age and race of the detected faces. Such assessments may not reflect how the individuals identify. Additionally, the classification is carried out in simplistic categories and contains only the most basic classes (for example, “male” and “female” for gender, and seven non-overlapping categories for ethnicity). To access these probabilistic assessments, you must therefore agree with the following statement: “I understand the ethical and privacy implications such assessments have for the interpretation of the results and that this analysis may result in personal and possibly sensitive data, and I wish to proceed.”
|
||||
|
||||
This disclosure statement is included as a separate line of code early in the flow of the Emotion Detector. Once the user has agreed with the statement, further data analyses will also include these assessments.
|
||||
|
||||
10
README.md
10
README.md
@ -39,22 +39,22 @@ The `AMMICO` package can be installed using pip:
|
||||
```
|
||||
pip install ammico
|
||||
```
|
||||
This will install the package and its dependencies locally. If after installation you get some errors when running some modules, please follow the instructions in the [FAQ](FAQ.md).
|
||||
This will install the package and its dependencies locally. If after installation you get some errors when running some modules, please follow the instructions in the [FAQ](https://ssciwr.github.io/AMMICO/build/html/faq_link.html).
|
||||
|
||||
## Usage
|
||||
|
||||
The main demonstration notebook can be found in the `notebooks` folder and also on google colab: [].
|
||||
The main demonstration notebook can be found in the `notebooks` folder and also on google colab: [](https://colab.research.google.com/github/ssciwr/ammico/blob/main/ammico/notebooks/DemoNotebook_ammico.ipynb).
|
||||
|
||||
There are further sample notebooks in the `notebooks` folder for the more experimental features:
|
||||
1. Topic analysis: Use the notebook `get-text-from-image.ipynb` to analyse the topics of the extraced text.\
|
||||
**You can run this notebook on google colab: [**
|
||||
**You can run this notebook on google colab: [](https://colab.research.google.com/github/ssciwr/ammico/blob/main/ammico/notebooks/get-text-from-image.ipynb)**
|
||||
Place the data files and google cloud vision API key in your google drive to access the data.
|
||||
1. To crop social media posts use the `cropposts.ipynb` notebook.
|
||||
**You can run this notebook on google colab: [**
|
||||
**You can run this notebook on google colab: [](https://colab.research.google.com/github/ssciwr/ammico/blob/main/ammico/notebooks/cropposts.ipynb)**
|
||||
|
||||
## Features
|
||||
### Text extraction
|
||||
The text is extracted from the images using [google-cloud-vision](https://cloud.google.com/vision). For this, you need an API key. Set up your google account following the instructions on the google Vision AI website or as described [here](docs/source/set_up_credentials.md).
|
||||
The text is extracted from the images using [google-cloud-vision](https://cloud.google.com/vision). For this, you need an API key. Set up your google account following the instructions on the google Vision AI website or as described [here](https://ssciwr.github.io/AMMICO/build/html/create_API_key_link.html).
|
||||
You then need to export the location of the API key as an environment variable:
|
||||
```
|
||||
export GOOGLE_APPLICATION_CREDENTIALS="location of your .json"
|
||||
|
||||
@ -287,7 +287,7 @@ class MultimodalSearch(AnalysisMethod):
|
||||
Returns:
|
||||
features_image_stacked (torch.Tensor): tensors of images features.
|
||||
"""
|
||||
features_image_stacked = torch.load(name)
|
||||
features_image_stacked = torch.load(name, weights_only=True)
|
||||
return features_image_stacked
|
||||
|
||||
def extract_text_features(self, model, text_input: str) -> torch.Tensor:
|
||||
|
||||
@ -15,7 +15,7 @@ sys.path.insert(0, os.path.abspath("../../ammico/"))
|
||||
project = "AMMICO"
|
||||
copyright = "2022, Scientific Software Center, Heidelberg University"
|
||||
author = "Scientific Software Center, Heidelberg University"
|
||||
release = "0.0.1"
|
||||
release = "0.2.2"
|
||||
|
||||
# -- General configuration ---------------------------------------------------
|
||||
# https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration
|
||||
@ -31,7 +31,7 @@ html_context = {
|
||||
"github_user": "ssciwr", # Username
|
||||
"github_repo": "AMMICO", # Repo name
|
||||
"github_version": "main", # Version
|
||||
"conf_py_path": "/source/", # Path in the checkout to the docs root
|
||||
"conf_py_path": "/docs/source/", # Path in the checkout to the docs root
|
||||
}
|
||||
|
||||
templates_path = ["_templates"]
|
||||
|
||||
@ -4,7 +4,7 @@ build-backend = "hatchling.build"
|
||||
|
||||
[project]
|
||||
name = "ammico"
|
||||
version = "0.2.1"
|
||||
version = "0.2.2"
|
||||
description = "AI Media and Misinformation Content Analysis Tool"
|
||||
readme = "README.md"
|
||||
maintainers = [
|
||||
|
||||
Загрузка…
x
Ссылка в новой задаче
Block a user