Giter VIP home page Giter VIP logo

sample-imageinpainting-hifill's People

Contributors

ascendhuawei avatar duxingren14 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

sample-imageinpainting-hifill's Issues

Attention Computing Branch architecture

Hi, thanks a lot for your research and your detailed explanations !

I still have question about the architecture of the attention computing branch. What does [P] stands for after the downsampling step ? It is described as a high-level feature map (compared to [P^l=3]) with several additional layers in Fig. 2, but not mentioned in the Appendix.

Trouble on larger masks?

Hi,

Thanks for publishing this source code! It makes testing your model easy :)

I'm currently testing the model on center-crop masks; however, I find that the model is struggling with larger center crops?
The images I tested were 512x512, with a center mask of 256x256.

Here is an example image that I tested, where I got strange corruption in the final result.
Ground truth image:
118
Input image:
118
Generated image:
118

Do you know what might be the cause of this? Or are center-crops difficult for the model?

If you would like to reproduce the result, here is the original mask (256x256 center crop):
118

If the images are compressed, I can give you a link to the raw images :)

How to train my own data set?

First of all, thank you for your great work? Can I train my own dataset on this network? Will you open your network code?

pretrainned the model

thanks for your contribution
I am asking how can I fine tunning your model
is there instruction
thanks

Training code

Hello, Great work!
Can you share the full pipeline including the training code?

some bad example?

Places365_val_00000327_masked
22 png_inpainted
thansks for your excellent work!when I test the example images you provide with your code, It perform very well. However, if we test the image shown above (400x600), it perform very bad. The result of completion will be a little blurry. I noticed that the minimum rate of change you tested in the experiment was 512x512. Does that mean that the algorithm needs to be on large images to get good results?

Is there anyway to run in 1.7x version Ascend ADK?

Hello,

I've been trying to use this project, but I use the 20.1 CANN with 1.7x. It changed from the 1.3x version and we haven't a hiai library anymore, everything is made using ACL. I can't see a hiai--acl documentation with more information about this transition. I'd like to know if there's anything being made to upgrade this application.

Thanks and Congratulations for the job.

Source of high resolution test samples

Hi, thanks for your excellent work.

I've read your paper and I'm really impressed by the results in your paper. But I am having some question when trying to test the model. I would appreciate it if you could provide some help.
In the experiment part, it's mentioned that the model is tested in several different resolutions, from 512512 to 8K. However, when I was testing the model, I found that the images in Places2 dataset are at most 512 pixels, and there seems to have no ground truth image whose size is really larger than 512. If the high resolution samples are produced by resizing the 512512 images, it seems that the evaluations on high resolution images are actually equivalent to that on 512*512 because each part of the model is essentially working on the 512 bases (including the ground truth). Am I misunderstanding something or there is another source of private testing samples?

Thanks for any help in advance :).

Convolution for mask and attention score training

I noticed that a 7x7 kernel is used in the model to process the mask. Why do you want to do this operation? And how to train this module? In addition, attention score didn't work very well in my own training, because the score of each patch was always average. Are there any suggestions you can give me?

attention score

Hi! Thanks for your great work! But when I implementing the post process, I have no idea what the attention score is in your demo. In the paper, it said only one patch outside mask and one inside will have similarity. So I make all others in zero and do softmax to get the attention score. But it seems not the one in your demo, do you have any advices for me?

How to train on 512*512 images with CRA?

Dear author, in the paper, the model was trained on 512x512 images. However, both the input and output of the generator are 512x512, which I think there is no need to resize the output to a high resolution to get the residuals. So my question is how do you train your CRA module on 512x512 images? Even there is no weights to be trained in CRA, it does affect the other part of the generator. And is the loss calculated on the generator output? Or the postprocess output (end-to-end training)?

most inpainting results suck

Thanks for sharing the paper and model.
But we cannot get good results on our normal test cases
Most inpainting results have artifacts even in very low resolution cases especially when the hole is big.
Is there any update?

How to do inference in batch?

Hi, thanks for the great work. I would like to integrate your model on my task. SoI have a question about doing inference with batch data. The pretrained model seems only to provide inference for one image, i.e., the shape of the input is [1, 3, 512, 512] instead of [?, 3, 512, 512]. It means that I cannot pass batch data to the model. As a result, it took 10s to have the result for a batch input, e.g., BATCH_SIZE=16. Is there any solution to fasten inference? Thanks in advance!

[BUG] Running Time

Thanks for sharing code of your novel and inspiring work. However, when I runing your code, I got the following error:

F tensorflow/core/framework/tensor_shape.cc:44] Check failed: NDIMS == dims() (2 vs. 4)Asking for tensor of 2 dimensions from a tensor of 4 dimensions 

I'm using tensorflow 1.6 with TitanXp.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.