Comments (7)
Basically, you just need to put use_tensorflow = True
in your config.
Note that not all layers are implemented currently in the TF backend. So far, the linear/forward layer and the rec layer with LSTM should work. Other layers are limited and also might have slightly different options, e.g. the conv layer. Maybe post your config here so I can have a quick look.
from returnn.
I'm using the config_real in the MDLSTM demo for IAM
from returnn.
So that config file I guess.
Most of the layers are not yet available in TF, i.e. there is no 1Dto2D, conv2, mdlstm.
1Dto2D and conv2 should be easy to replace with the existing TF layers, I think (or it would not require much work otherwise).
mdlstm is not so simple to adapt to TF and would require some work. This is planned but no-one currently works on it.
from returnn.
It's a pity that this functionality is not available yet, it would be very interesting having MDLSTM supported by TF. I look forward to this new feature.
from returnn.
I want to use tensorflow backend, but i only see config for 'train' . how to build config file, eval data for 'forward'?
from returnn.
@lannp Where are you looking at? Here are configs for training (config_real
) as well as forwarding (config_fwd
). Anyway, TensorFlow is not supported at all right now for MDLSTM. Feel free to contribute!
from returnn.
This is not about any or the official pure TF implementation. This is about our own CUDA kernel, which should be much faster.
from returnn.
Related Issues (20)
- PyTorch collect model statistics
- PyTorch recover after CUDA OOM with restart does not work with CUDA HOT 3
- PyTorch distributed training CPU OOM with sync_on_cpu HOT 1
- Support `torch.compile` for RF
- RF backend: PyTorch code
- Different effective learning rate reported over gpus HOT 11
- CUDA error: initialization error HOT 3
- MultiProcDataset inside PyTorch DataLoader with num_workers>0, multiple issues HOT 4
- RuntimeError: CUDA error: unspecified launch failure HOT 2
- NonDaemonicSpawnProcess hangs at exit HOT 2
- High memory usage with datasets (specifically when multi procs are used)
- Hang at exit in TDL worker in multiprocessing `_run_finalizers`, deadlock in `_wait_for_tstate_lock`? HOT 6
- Hang HOT 2
- Returnn Native after using different apptainer uses old compilation HOT 6
- MetaDataset with sequence list filter file
- HDFDataset (or generic dataset) post processing HOT 15
- Dataset batching like ESPnet support
- torch.nn.functional.conv2d: RuntimeError: GET was unable to find an engine to execute this computation HOT 1
- TensorFlow 2.14 degradation in WER HOT 2
- Updates for recent TensorFlow version
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from returnn.