Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

Interface for Image-Related Deep Learning Models (E.g. NSFW, MAYBE and SFW)

License

NotificationsYou must be signed in to change notification settings

qeeqbox/image-analyzer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Image analyzer is an interface that simplifies interaction with image-related deep learning models. The interface has built-in features that get adjusted based on the models. You need to set the destination folder that contains your models with specific pattern names. The built-in features will use the names.

The interface was initially part of an internal project that detects abusive content with the possibility of tracing the subjects.

Structure

Interface

How to run?

sudo apt-get install -y python3-opencvpip3 install image-analyzer
fromimageanalyzerimportrun_serverrun_server(settings={'input_shape':[224,224],'percentage':0.90,'options':[],'weights': {'safe':50,'maybe':75,'unsafe':100},'verbose':True},port=8989)

Name structue

  • [Name] Name of the model
  • [Info] Description
  • [Categories]Your model categories (if any)
  • [Number] Model place
  • .h5 Model extention

E.g.

The following are examples of the model generated automatically by QeeqBox Automated Deep Learning System for large files. The examples are included in this project

  • [Model 3.224x][Find safe, maybe and unsafe images 3.224][safe unsafe][1].h5
  • [Model 2.224x][Find safe and unsafe images 2.224][safe maybe unsafe][2].h5

Other Projects

Releases

No releases published

Sponsor this project

 

Packages

No packages published

[8]ページ先頭

©2009-2025 Movatter.jp