diff --git a/generate_reconstructions.ipynb b/generate_reconstructions.ipynb index 82f75eb..0d13f1b 100644 --- a/generate_reconstructions.ipynb +++ b/generate_reconstructions.ipynb @@ -26,9 +26,9 @@ "import numpy as np\n", "from PIL import Image\n", "from matplotlib import pyplot as plt\n", - "from huggingface_mae import MAEModel\n", + "from maes_microscopy.huggingface_mae import MAEModel\n", "from skimage.transform import resize\n", - "from mae_utils import unflatten_tokens" + "from maes_microscopy.mae_utils import unflatten_tokens" ] }, { diff --git a/maes_microscopy/__init__.py b/maes_microscopy/__init__.py new file mode 100644 index 0000000..e69de29 diff --git a/huggingface_mae.py b/maes_microscopy/huggingface_mae.py similarity index 97% rename from huggingface_mae.py rename to maes_microscopy/huggingface_mae.py index 3bbefae..30ea542 100644 --- a/huggingface_mae.py +++ b/maes_microscopy/huggingface_mae.py @@ -5,11 +5,11 @@ from transformers import PretrainedConfig, PreTrainedModel -from loss import FourierLoss -from normalizer import Normalizer -from mae_modules import CAMAEDecoder, MAEDecoder, MAEEncoder -from mae_utils import flatten_images -from vit import ( +from maes_microscopy.loss import FourierLoss +from maes_microscopy.normalizer import Normalizer +from maes_microscopy.mae_modules import CAMAEDecoder, MAEDecoder, MAEEncoder +from maes_microscopy.mae_utils import flatten_images +from maes_microscopy.vit import ( generate_2d_sincos_pos_embeddings, sincos_positional_encoding_vit, vit_small_patch16_256, diff --git a/loss.py b/maes_microscopy/loss.py similarity index 100% rename from loss.py rename to maes_microscopy/loss.py diff --git a/mae_modules.py b/maes_microscopy/mae_modules.py similarity index 98% rename from mae_modules.py rename to maes_microscopy/mae_modules.py index a09f2c9..f958613 100644 --- a/mae_modules.py +++ b/maes_microscopy/mae_modules.py @@ -7,8 +7,8 @@ from timm.models.helpers import checkpoint_seq from timm.models.vision_transformer import Block, Mlp, VisionTransformer -from masking import transformer_random_masking -from vit import channel_agnostic_vit +from maes_microscopy.masking import transformer_random_masking +from maes_microscopy.vit import channel_agnostic_vit # If interested in training new MAEs, combine an encoder and decoder into a new module, and you should # leverage the flattening and unflattening utilities as needed from mae_utils.py. diff --git a/mae_utils.py b/maes_microscopy/mae_utils.py similarity index 100% rename from mae_utils.py rename to maes_microscopy/mae_utils.py diff --git a/masking.py b/maes_microscopy/masking.py similarity index 100% rename from masking.py rename to maes_microscopy/masking.py diff --git a/normalizer.py b/maes_microscopy/normalizer.py similarity index 100% rename from normalizer.py rename to maes_microscopy/normalizer.py diff --git a/test_huggingface_mae.py b/maes_microscopy/test_huggingface_mae.py similarity index 95% rename from test_huggingface_mae.py rename to maes_microscopy/test_huggingface_mae.py index 1de4b1d..c09ea10 100644 --- a/test_huggingface_mae.py +++ b/maes_microscopy/test_huggingface_mae.py @@ -1,7 +1,7 @@ import pytest import torch -from huggingface_mae import MAEModel +from maes_microscopy.huggingface_mae import MAEModel huggingface_openphenom_model_dir = "." # huggingface_modelpath = "recursionpharma/OpenPhenom" diff --git a/vit.py b/maes_microscopy/vit.py similarity index 100% rename from vit.py rename to maes_microscopy/vit.py diff --git a/vit_encoder.py b/maes_microscopy/vit_encoder.py similarity index 100% rename from vit_encoder.py rename to maes_microscopy/vit_encoder.py diff --git a/pyproject.toml b/pyproject.toml index 8d1d750..41ceeb8 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -30,4 +30,4 @@ dependencies = [ ] [tool.setuptools] -py-modules = [] \ No newline at end of file +packages=["maes_microscopy"]