Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can we work this DeepQA for Chinese dataset? #201

Open
balagopal24 opened this issue Aug 25, 2018 · 4 comments
Open

Can we work this DeepQA for Chinese dataset? #201

balagopal24 opened this issue Aug 25, 2018 · 4 comments

Comments

@balagopal24
Copy link

Hello,

I'm trying to include my own chinese dataset by using conversation file <name>.txt, copy it in this repository and launch the program with the option --corpus lightweight --datasetTag <name>

But I could'nt get any responses in chinese. Can anybody help me to solve this issue. Does this chatbot support for chinese language?

Regards,
Bala

@ghostyoona
Copy link

Hello Bala, I'm glad to tell you that DeepQA indeed support Chinese. Please prepare you own data for example file named "my_own_data.txt", then put it in dir 'lightweight'. And when you train your model, you should use the command python main.py --corpus lightweight --datasetTag my_own_data. If the command doesn't work, please remove the data in "data/samples" dir.
Good luck!

@xxllp
Copy link

xxllp commented Dec 17, 2018

good ,show the data cut in word ?

@lhuang9703
Copy link

Hi,Do we need to cut the word if we want to use Chinese? Anyone used with Chinese data successfully?

@Zhouziyi828
Copy link

I have tried according to what you said, but the answer I got was still in English.Like this:
Q: 香港 还 卖 这么 有 爱的 冰棍 ?
A: And the jews killed me.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants