Code Monkey home page Code Monkey logo

secmi's People

Contributors

jinhaoduan avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar

secmi's Issues

Code for GAN-Leaks

Hi, I am doing some research on the privacy of diffusion models. Could I get the code you use to test GAN-Leaks?

Reconfirmation of Training (Fine-tuning) Details on Stable diffusion

Though you've explained it in issue6, I'd still like to double-check this detail cause I'm confused about the experiement result.

Referring to the official code for stable diffusion fine-tuning and the .sh example, there are two parameters for data transform during fine-tuning, namely args.center_crop and args.random_flip, representing centerCrop / randomCrop, no-Flip / randomFlip respectively.

If using the default parameter settings, it's actually a combination of randomCrop and no-Flip. Here's the code:

# Preprocessing the datasets.
train_transforms = transforms.Compose(
    [
        transforms.Resize(args.resolution, interpolation=transforms.InterpolationMode.BILINEAR),
        transforms.CenterCrop(args.resolution) if args.center_crop else transforms.RandomCrop(args.resolution),
        transforms.RandomHorizontalFlip() if args.random_flip else transforms.Lambda(lambda x: x),
        transforms.ToTensor(),
        transforms.Normalize([0.5], [0.5]),
    ]
)

I'd like to know which parameter combination you used?

I conducted experiments on various data-augmentation combination based on your methods (COCO dataset, 2500/2500 split, 150,000 steps, as described in your paper).

The experimental results align with yours only for randomCrop+ no-Flip (ASR/AUC: 0.8334/0.9105). And centerCrop +no-Flip yields higher results (ASR≈0.90). On the other hand, randomCrop+randomFlip results in much lower performance (ASR≈0.75).

So, did you use the combination of randomCrop and no-Flip?

Thank you for your time, and I look forward to your response!

Evaluation of conditional diffusion models

Thank you for your amazing work!

I'd like to inquire if there are evaluation codes about conditional diffusion model in this repository (Sec. 5.5 and Sec. 5.6 in paper) ?

Inquiries regarding the details of the shadow experiment.

Thank you for your inspiring work!

But I have one more question that arises from the closed issue:

You can assume we are in a shadow environment: training a shadow model with a known member/nonmember splitting and determining a threshold/NN according to this known splitting. After we get the threshold/NN, we could apply them to any real victim models (which we don't know the membership splitting).

We have some results to show that the threshold obtained from the shadow environment can be transferred to the real victim models:

Method Attack Target AUC ASR
SecMI_stat Shadow Model 0.881 0.811
SecMI_stat Victim Model - 0.804
SecMI_NNs Shadow Model 0.951 0.888
SecMI_NNs Victim Model - 0.893

Could you please describe how to construct the training data for the "shadow models"? When training a shadow model, do you assume that the adversary already knows the distribution of the dataset? (Or can obtain some of the training data?)

question about the evaluation

Hi,

Thanks for your great work! I have a question about the evaluation process. In my understanding, current implementations directly use the member/non-member labels to get the threshold for $SecMI_{stat}$ and for NN training $SecMI_{NNs}$, should not the attackers only have access to the data distribution $D$ instead of $D_M$ or $D_H$?

_, _, member_loader, nonmember_loader = load_member_data(dataset_root=dataset_root, dataset_name=dataset, batch_size=batch_size,

some questions on t_c, x_t_tilde

Hi, thank you for your paper and for sharing the code. May I ask some questions?

epsilon = model(x, t_c)

When you calculate epsilon, initially t_c is 0, but during DDPM inference, time step t_c should be equal or great than 1. So can we set t_c to 0 here?

  1. In Equation (13) in the paper, the $\tilde{x_t}$ is calculated many steps iteratively. Can we just calculate $\tilde{x_t}$ based on Equation (4) in one step?

Thanks.

Inquiry Regarding Experimental Setup in LDM SecMI

Thank you for your great work and open-source code, which inspires me a lot.

During replication, there was a slight disparity between my results and yours (with ASR, AUC even higher on the Pokémon dataset than yours). So I want to know which differing settings led to my higher results.

My setting:
Pokmon train-test split: 416, 417.
Training steps: 15000, Batch size: 1, Gradient_accumulation_steps: 4, LR: 1e-5.
Without Crop and Flip. (Did you use crop and flip during training?)

My result
ASR 0.90, AUC 0.9391 with Prompt (higher than yours: 0.821, 0.891)

Trying to keep the settings consistent with the paper, but I still obtained different results. Looking forward to your response!

Request for Dataset Statistics: CIFAR100 and Tiny-ImageNet

Dear Authors,

Thanks for your excellent work. I've noted the statistics for CIFAR10 in ./stats/cifar10.train.npz. Would it be feasible for you also to release the statistics for CIFAR100 and Tiny-ImageNet?

Thank you for your time and consideration.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.