Giter VIP home page Giter VIP logo

tf_seq2seq_chatbot's Introduction

tensorflow seq2seq chatbot

Note: the repository is not maintained. Feel free to PM me if you'd like to take up the maintainance.

Build a general-purpose conversational chatbot based on a hot seq2seq approach implemented in tensorflow. Since it doesn't produce good results so far, also consider other implementations of seq2seq.

The current results are pretty lousy:

hello baby	        - hello
how old are you ?   - twenty .
i am lonely	        - i am not
nice                - you ' re not going to be okay .
so rude	            - i ' m sorry .

Disclaimer:

  • the answers are hand-picked (it looks cooler that way)
  • chatbot has no power to follow the conversation line so far; in the example above it's a just a coincidence (hand-picked one)

Everyone is welcome to investigate the code and suggest the improvements.

Actual deeds

  • realise how to diversify chatbot answers (currently the most probable one is picked and it's dull)

Papers

Nice picture

seq2seq

Curtesy of this article.

Setup

git clone [email protected]:nicolas-ivanov/tf_seq2seq_chatbot.git
cd tf_seq2seq_chatbot
bash setup.sh

Run

Train a seq2seq model on a small (17 MB) corpus of movie subtitles:

python train.py

(this command will run the training on a CPU... GPU instructions are coming)

Test trained trained model on a set of common questions:

python test.py

Chat with trained model in console:

python chat.py

All configuration params are stored at tf_seq2seq_chatbot/configs/config.py

GPU usage

If you are lucky to have a proper gpu configuration for tensorflow already, this should do the job:

python train.py

Otherwise you may need to build tensorflow from source and run the code as follows:

cd tensorflow  # cd to the tensorflow source folder
cp -r ~/tf_seq2seq_chatbot ./  # copy project's code to tensorflow root
bazel build -c opt --config=cuda tf_seq2seq_chatbot:train  # build with gpu-enable option
./bazel-bin/tf_seq2seq_chatbot/train  # run the built code

Requirements

tf_seq2seq_chatbot's People

Contributors

donothingloop avatar nicolas-ivanov avatar rodart avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

tf_seq2seq_chatbot's Issues

ImportError help!

Traceback (most recent call last):
File "train.py", line 8, in
from tf_seq2seq_chatbot.lib.train import train
File "/home/vinicius_torrecillas/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 14, in
from tf_seq2seq_chatbot.lib.seq2seq_model_utils import create_model
File "/home/vinicius_torrecillas/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 11, in
from tf_seq2seq_chatbot.lib import seq2seq_model
File "/home/vinicius_torrecillas/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 28, in
from tensorflow.models.rnn.translate import data_utils
ImportError: No module named models.rnn.translate

how can i fix this?

About training data order.

Thank you for your useful repository about seq2seq chatbot.

I'm wondering why utterance of your learning data are reverse order.
For example, in your data:

Okay... then how 'bout we try out some French cuisine.  Saturday?  Night?
Not the hacking and gagging and spitting part.  Please.
Well, I thought we'd start with pronunciation, if that's okay with you.
Can we make this quick?  Roxanne Korrine and Andrew Barrett are having an incredibly horrendous public break- up on the quad.  Again.

But actual utterance starts from “Can we make this quick?...”
Are there any reason?

I've tried to train the tensorflow's translation model with [Cornell Movie-Dialogs Corpus](Cornell Movie-Dialogs Corpus) in normal order and got outputs like macournoyer/neuralconvo's ones.

TypeError: unsupported operand type(s) for /=: 'Tensor' and 'Tensor'

I get this error message when I try to run train.py:

Traceback (most recent call last): File "train.py", line 15, in <module> tf.app.run() File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/platform/default/_app.py", line 11, in run sys.exit(main(sys.argv)) File "train.py", line 12, in main train() File "/home/urja/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 28, in train model = create_model(sess, forward_only=False) File "/home/urja/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 27, in create_model forward_only=forward_only) File "/home/urja/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 149, in __init__ softmax_loss_function=softmax_loss_function) File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 651, in model_with_buckets File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 457, in sequence_loss File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 424, in sequence_loss_by_example TypeError: unsupported operand type(s) for /=: 'Tensor' and 'Tensor'

if the answer is very long

For this model, if the answer is very long, the answer is not good。
In chat.py, the answer is up to 15 words.
How to modify the program?

ImportError: No module named rnn_enhancement

It seems _get_predicted_sentence is still using rnn_enhancement.decoding_enhanced's sample_with_temperature.

Will be happy to submit a PR pushing that function to seq2seq_model_utils.py.

Error when run "python train.py"

Preparing dialog data in /var/lib/tf_seq2seq_chatbot/data
Creating vocabulary /var/lib/tf_seq2seq_chatbot/data/vocab20000.in from data /var/lib/tf_seq2seq_chatbot/data/chat.in
Traceback (most recent call last):
File "train.py", line 15, in
tf.app.run()
File "/home/minh/opt/anaconda3/envs/ml/lib/python3.5/site-packages/tensorflow/python/platform/app.py", line 30, in run
sys.exit(main(sys.argv))
File "train.py", line 12, in main
train()
File "/home/minh/source/machine-learing/chatbot/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 22, in train
train_data, dev_data, _ = data_utils.prepare_dialog_data(FLAGS.data_dir, FLAGS.vocab_size)
File "/home/minh/source/machine-learing/chatbot/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/data_utils.py", line 200, in prepare_dialog_data
create_vocabulary(vocab_path, train_path + ".in", vocabulary_size)
File "/home/minh/source/machine-learing/chatbot/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/data_utils.py", line 70, in create_vocabulary
for line in f:
File "/home/minh/opt/anaconda3/envs/ml/lib/python3.5/site-packages/tensorflow/python/platform/gfile.py", line 176, in next
return next(self._fp)
File "/home/minh/opt/anaconda3/envs/ml/lib/python3.5/codecs.py", line 321, in decode
(result, consumed) = self._buffer_decode(data, self.errors, final)
UnicodeDecodeError: 'utf-8' codec can't decode byte 0xad in position 2329: invalid start byte

TypeError: Using a `tf.Tensor` as a Python `bool` is not allowed. -----tensorflow0.8

envy@ub1404:/os_pri/github/tf_seq2seq_chatbot$ python train.py
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcublas.so locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcudnn.so locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcufft.so locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcuda.so.1 locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcurand.so locally
You are using the decoding_enhanced library!
Initializer Library -- Will add to this if necessary.
Unitary Linear has been imported
Preparing dialog data in /var/lib/tf_seq2seq_chatbot/data
Creating vocabulary /var/lib/tf_seq2seq_chatbot/data/vocab80000.in from data /var/lib/tf_seq2seq_chatbot/data/chat.in
processing line 100000
processing line 200000
processing line 300000
Tokenizing data in /var/lib/tf_seq2seq_chatbot/data/chat.in
tokenizing line 100000
tokenizing line 200000
tokenizing line 300000
Tokenizing data in /var/lib/tf_seq2seq_chatbot/data/chat_test.in
I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:900] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
I tensorflow/core/common_runtime/gpu/gpu_init.cc:102] Found device 0 with properties:
name: GeForce GTX 950M
major: 5 minor: 0 memoryClockRate (GHz) 1.124
pciBusID 0000:01:00.0
Total memory: 4.00GiB
Free memory: 3.70GiB
I tensorflow/core/common_runtime/gpu/gpu_init.cc:126] DMA: 0
I tensorflow/core/common_runtime/gpu/gpu_init.cc:136] 0: Y
I tensorflow/core/common_runtime/gpu/gpu_device.cc:755] Creating TensorFlow device (/gpu:0) -> (device: 0, name: GeForce GTX 950M, pci bus id: 0000:01:00.0)
Creating 4 layers of 2048 units.
Traceback (most recent call last):
File "train.py", line 15, in
tf.app.run()
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/platform/app.py", line 30, in run
sys.exit(main(sys.argv))
File "train.py", line 12, in main
train()
File "/home/envy/os_pri/github/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 28, in train
model = create_model(sess, forward_only=False)
File "/home/envy/os_pri/github/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 27, in create_model
forward_only=forward_only)
File "/home/envy/os_pri/github/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 149, in init
softmax_loss_function=softmax_loss_function)
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 629, in model_with_buckets
File "/home/envy/os_pri/github/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 148, in
lambda x, y: seq2seq_f(x, y, False),
File "/home/envy/os_pri/github/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 112, in seq2seq_f
feed_previous=do_decode)
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 341, in embedding_attention_seq2seq
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 259, in embedding_attention_decoder
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 165, in attention_decoder
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 112, in attention
File "build/bdist.linux-x86_64/egg/rnn_enhancement/linear_enhanced.py", line 42, in linear
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/framework/ops.py", line 475, in nonzero
raise TypeError("Using a tf.Tensor as a Python bool is not allowed. "
TypeError: Using a tf.Tensor as a Python bool is not allowed. Use if t is not None: instead of if t: to test if a tensor is defined, and use the logical TensorFlow ops to test the value of a tensor.
envy@ub1404:
/os_pri/github/tf_seq2seq_chatbot$

Error when training

I get the following error when i attempt to train the network. Any ideas? I'm pretty sure all the dependencies are properly installed since the setup bash script worked.

1482-sraval:tf_seq2seq_chatbot sraval$ python train.py
You are using the decoding_enhanced library!
Initializer Library -- Will add to this if necessary.
Unitary Linear has been imported
Preparing dialog data in /var/lib/tf_seq2seq_chatbot/data
I tensorflow/core/common_runtime/local_device.cc:40] Local device intra op parallelism threads: 4
I tensorflow/core/common_runtime/direct_session.cc:58] Direct session inter op parallelism threads: 4
Creating 4 layers of 2048 units.
use aggregation_method=2
/Library/Python/2.7/site-packages/tensorflow/python/ops/gradients.py:87: UserWarning: Converting sparse IndexedSlices to a dense Tensor with 163840000 elements. This may consume a large amount of memory.
  "This may consume a large amount of memory." % num_elements)
Created model with fresh parameters.
Traceback (most recent call last):
  File "train.py", line 15, in <module>
    tf.app.run()
  File "/Library/Python/2.7/site-packages/tensorflow/python/platform/default/_app.py", line 30, in run
    sys.exit(main(sys.argv))
  File "train.py", line 12, in main
    train()
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 28, in train
    model = create_model(sess, forward_only=False)
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 35, in create_model
    session.run(tf.initialize_all_variables())
  File "/Library/Python/2.7/site-packages/tensorflow/python/client/session.py", line 368, in run
    results = self._do_run(target_list, unique_fetch_targets, feed_dict_string)
  File "/Library/Python/2.7/site-packages/tensorflow/python/client/session.py", line 444, in _do_run
    e.code)
tensorflow.python.framework.errors.InvalidArgumentError: Cannot assign a device to node 'embedding_attention_seq2seq/embedding_attention_decoder/attention_decoder/MultiRNNCell/Cell3/GRUCell/Candidate/Linear/Bias': Could not satisfy explicit device specification '/gpu:3'
     [[Node: embedding_attention_seq2seq/embedding_attention_decoder/attention_decoder/MultiRNNCell/Cell3/GRUCell/Candidate/Linear/Bias = Variable[container="", dtype=DT_FLOAT, shape=[2048], shared_name="", _device="/gpu:3"]()]]
Caused by op u'embedding_attention_seq2seq/embedding_attention_decoder/attention_decoder/MultiRNNCell/Cell3/GRUCell/Candidate/Linear/Bias', defined at:
  File "train.py", line 15, in <module>
    tf.app.run()
  File "/Library/Python/2.7/site-packages/tensorflow/python/platform/default/_app.py", line 30, in run
    sys.exit(main(sys.argv))
  File "train.py", line 12, in main
    train()
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 28, in train
    model = create_model(sess, forward_only=False)
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 27, in create_model
    forward_only=forward_only)
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 149, in __init__
    softmax_loss_function=softmax_loss_function)
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/seq2seq_enhanced.py", line 629, in model_with_buckets
    bucket_decoder_inputs) #nick pay attention here -- you added bucket_states
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 148, in <lambda>
    lambda x, y: seq2seq_f(x, y, False),
  File "/Users/sraval/Desktop/yooo2/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 112, in seq2seq_f
    feed_previous=do_decode)
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/seq2seq_enhanced.py", line 341, in embedding_attention_seq2seq
    temperature_decode = temperature_decode, temperature = temperature)
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/seq2seq_enhanced.py", line 259, in embedding_attention_decoder
    temperature = temperature)
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/seq2seq_enhanced.py", line 159, in attention_decoder
    cell_output, new_state = cell(x, hidden_state_input) #nick, changed this to your hidden state input
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/rnn_cell_enhanced.py", line 1149, in __call__
    cur_inp, new_state = cell(cur_inp, cur_state) #this replaces cur_inp to a differen input
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/rnn_cell_enhanced.py", line 520, in __call__
    c = tf.tanh(linear.linear([inputs, r * state], self._num_units, True))
  File "build/bdist.macosx-10.10-intel/egg/rnn_enhancement/linear_enhanced.py", line 68, in linear
    initializer=tf.constant_initializer(bias_start))
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/variable_scope.py", line 242, in get_variable
    trainable, collections)
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/variable_scope.py", line 175, in get_variable
    self.reuse, trainable, collections)
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/variable_scope.py", line 119, in get_variable
    collections=collections)
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/variables.py", line 210, in __init__
    name=name)
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/state_ops.py", line 116, in variable_op
    container=container, shared_name=shared_name)
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/gen_state_ops.py", line 334, in _variable
    name=name)
  File "/Library/Python/2.7/site-packages/tensorflow/python/ops/op_def_library.py", line 664, in apply_op
    op_def=op_def)
  File "/Library/Python/2.7/site-packages/tensorflow/python/framework/ops.py", line 1834, in create_op
    original_op=self._default_original_op, op_def=op_def)
  File "/Library/Python/2.7/site-packages/tensorflow/python/framework/ops.py", line 1043, in __init__
    self._traceback = _extract_stack()

Using a `tf.Tensor` as a Python `bool` is not allowed

When training the model by running the train.py command, I encounter the following issue:

python train.py
I tensorflow/stream_executor/dso_loader.cc:108] successfully opened CUDA library libcublas.so locally
I tensorflow/stream_executor/dso_loader.cc:108] successfully opened CUDA library libcudnn.so locally
I tensorflow/stream_executor/dso_loader.cc:108] successfully opened CUDA library libcufft.so locally
I tensorflow/stream_executor/dso_loader.cc:108] successfully opened CUDA library libcuda.so locally
I tensorflow/stream_executor/dso_loader.cc:108] successfully opened CUDA library libcurand.so locally
You are using the decoding_enhanced library!
Initializer Library -- Will add to this if necessary.
Unitary Linear has been imported
Preparing dialog data in /var/lib/tf_seq2seq_chatbot/data
E tensorflow/stream_executor/cuda/cuda_driver.cc:491] failed call to cuInit: CUDA_ERROR_NO_DEVICE
I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:140] kernel driver does not appear to be running on this host (instance-1): /proc/driver/nvidia/version does not exist
I tensorflow/core/common_runtime/gpu/gpu_init.cc:81] No GPU devices available on machine.
Creating 4 layers of 2048 units.
Traceback (most recent call last):
host (instance-1): /proc/driver/nvidia/version does not exist
File "train.py", line 15, in
tf.app.run()
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/platform/app.py", line 30, in run
sys.exit(main(sys.argv))
File "train.py", line 12, in main
train()
File "/home/vibhu_vandita/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 28, in train
model = create_model(sess, forward_only=False)
File "/home/vibhu_vandita/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 27, in cre
ate_model
forward_only=forward_only)
File "/home/vibhu_vandita/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 149, in init
softmax_loss_function=softmax_loss_function)
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 629, in model_with_buckets
File "/home/vibhu_vandita/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 148, in
lambda x, y: seq2seq_f(x, y, False),
File "/home/vibhu_vandita/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 112, in seq2seq_
f
feed_previous=do_decode)
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 341, in embedding_attention_se
q2seq
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 259, in embedding_attention_de
coder
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 165, in attention_decoder
File "build/bdist.linux-x86_64/egg/rnn_enhancement/seq2seq_enhanced.py", line 112, in attention
File "build/bdist.linux-x86_64/egg/rnn_enhancement/linear_enhanced.py", line 42, in linear
File "/usr/local/lib/python2.7/dist-packages/tensorflow/python/framework/ops.py", line 528, in nonzero
raise TypeError("Using a tf.Tensor as a Python bool is not allowed. "
TypeError: Using a tf.Tensor as a Python bool is not allowed. Use if t is not None: instead of if t: to test if a tensor is defined, and use the logical TensorFlow ops to test the value of a tensor.

What can be done in order to resolve this issue?

How long should I train in CPU?

global step 2200 learning rate 0.5000 step-time 2.00 perplexity 37.43
  eval: bucket 0 perplexity 20.10
  eval: bucket 1 perplexity 33.75
  eval: bucket 2 perplexity 34.75
  eval: bucket 3 perplexity 43.44

result:

how old are you ?
i ' m not .

ModuleNotFoundError: No module named 'tensorflow.models'

How can I fix this error while running python train.py

Traceback (most recent call last):
File "train.py", line 8, in
from tf_seq2seq_chatbot.lib.train import train
File "/home/laoma/nlp_Hujiang/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/train.py", line 14, in
from tf_seq2seq_chatbot.lib.seq2seq_model_utils import create_model
File "/home/laoma/nlp_Hujiang/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model_utils.py", line 11, in
from tf_seq2seq_chatbot.lib import seq2seq_model
File "/home/laoma/nlp_Hujiang/tf_seq2seq_chatbot/tf_seq2seq_chatbot/lib/seq2seq_model.py", line 28, in
from tensorflow.models.rnn.translate import data_utils
ModuleNotFoundError: No module named 'tensorflow.models'

How train this my own data?

hi i am confusing how can i traind this my own data like hi rep: Hi2 etc ,problem is that where place i file ?where i give this file direct?and how put data in file question anser pattren?

bash setup.sh access denied

tf_seq2seq_chatbot son$ sh setup.sh
Cloning into 'Seq2Seq_Upgrade_TensorFlow'...
Permission denied (publickey).
fatal: Could not read from remote repository.


bash setup.sh returns above error

I don't know what is the problem

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.