Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

One or two questions about the running of experimental code #5

Open
15922534139 opened this issue Aug 13, 2022 · 6 comments
Open

One or two questions about the running of experimental code #5

15922534139 opened this issue Aug 13, 2022 · 6 comments

Comments

@15922534139
Copy link

Hello, I'm honored to read your paper. I want to ask some questions. First, how can I download the experimental data set in the paper? Need to climb over the wall? Second, is the system running in the whole experiment on windows or Linux?

@tanyuqian
Copy link
Owner

Thanks a lot for reading our paper!

(1) We provide the code to download dataset (see README -- Download Data). The CNN data needs access to Google Drive.
(2) All under Linux.

@15922534139
Copy link
Author

15922534139 commented Aug 19, 2022 via email

@tanyuqian
Copy link
Owner

Oh sorry for missing this message.
(1) Yes, VPN might be necessary because the data is downloaded from google drive.
(2) About the GPUs, I'm not sure if your hardware setting is run-able. You can check if PyTorch can detect your gpus by torch.cuda.device_count().

@15922534139
Copy link
Author

Hello, the author. I also want to ask you about this configuration. Due to the limited hardware resources, can you change the code structure to run the bart and gpt models on a GPU. Because I don't have such high experimental conditions as 2 GPUs or 4 GPUs at present, please reply. Thank you very much

@15922534139
Copy link
Author

Hello, the author. My current configuration of the local laptop is an independent video card RTX3060, and the video memory is 6GB. Do you think such a configuration has the basic conditions for completing your experiment of progressive long text generation? Thank you for your reply. Thank you very much!

@tanyuqian
Copy link
Owner

Hi! Our model is BART-Large with max_length=1024 on both encoder and decoder sides, and it takes more than 10GB GPU memory (roughly 15 GB, if I remember correctly), so I distributed the parameters on 2 1080Ti GPUS. A single 6GB GPU is not able to run the code.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants