1
0
Fork 0
mirror of https://github.com/immich-app/immich.git synced 2024-12-28 22:51:59 +00:00
immich/machine-learning
Daniele Ricci c7d53a5006
docs: remove obsolete environment variables (#3926)
* Obsolete environment variables

* Review fixes

* Review fixes

* Review fixes
2023-09-01 18:05:45 +07:00
..
app docs: remove obsolete environment variables (#3926) 2023-09-01 18:05:45 +07:00
.dockerignore feat: facial recognition (#2180) 2023-05-17 12:07:17 -05:00
.gitignore feat: facial recognition (#2180) 2023-05-17 12:07:17 -05:00
Dockerfile chore(ml): memory optimisations (#3934) 2023-08-31 18:30:53 -05:00
load_test.sh added locustfile (#2926) 2023-06-25 13:20:45 -05:00
locustfile.py added locustfile (#2926) 2023-06-25 13:20:45 -05:00
log_conf.json chore(ml): memory optimisations (#3934) 2023-08-31 18:30:53 -05:00
poetry.lock chore(ml): memory optimisations (#3934) 2023-08-31 18:30:53 -05:00
pyproject.toml chore(ml): memory optimisations (#3934) 2023-08-31 18:30:53 -05:00
README.md added locustfile (#2926) 2023-06-25 13:20:45 -05:00
README_es_ES.md Add Spanish translations of Readme (#3511) 2023-08-02 06:51:08 -05:00
requirements.txt feat(ml)!: switch image classification and CLIP models to ONNX (#3809) 2023-08-25 06:28:51 +02:00
start.sh chore(ml): memory optimisations (#3934) 2023-08-31 18:30:53 -05:00

Immich Machine Learning

  • Image classification
  • CLIP embeddings
  • Facial recognition

Setup

This project uses Poetry, so be sure to install it first. Running poetry install --no-root --with dev will install everything you need in an isolated virtual environment.

To add or remove dependencies, you can use the commands poetry add $PACKAGE_NAME and poetry remove $PACKAGE_NAME, respectively. Be sure to commit the poetry.lock and pyproject.toml files to reflect any changes in dependencies.

Load Testing

To measure inference throughput and latency, you can use Locust using the provided locustfile.py. Locust works by querying the model endpoints and aggregating their statistics, meaning the app must be deployed. You can run load_test.sh to automatically deploy the app locally and start Locust, optionally adjusting its env variables as needed.

Alternatively, for more custom testing, you may also run locust directly: see the documentation. Note that in Locust's jargon, concurrency is measured in users, and each user runs one task at a time. To achieve a particular per-endpoint concurrency, multiply that number by the number of endpoints to be queried. For example, if there are 3 endpoints and you want each of them to receive 8 requests at a time, you should set the number of users to 24.