Comments (3)
You can just provide an all-zero input vector for the GO annotations - this will tell the model that you don't provide it with GO annotations and that it will have to infer them on its own. That's in fact also what we did when fine-tuning and testing the model so it's not an issue. ProteinBERT treats GO annotations has a bonus, but it's not really necessary.
from protein_bert.
Thank you for your explanation. I have another question, for my own protein sequence task. I want to compare the performance of pre-training on the results. How to directly train a end-to-end model without pre-training?
from protein_bert.
If you want a non-pretrained model, then instead of using load_pretrained_model
you can do something along these lines:
from proteinbert import PretrainingModelGenerator, FinetuningModelGenerator, InputEncoder
from proteinbert import conv_and_global_attention_model
n_annotations = 1 # So long as you don't plan to provide annotations as input, it doesn't really matter what number you choose
output_spec = ...
uninitialized_pretraining_model_generator = PretrainingModelGenerator(conv_and_global_attention_model.create_model, n_annotations)
model_generator = FinetuningModelGenerator(uninitialized_pretraining_model_generator, output_spec)
input_encoder = InputEncoder(n_annotations)
from protein_bert.
Related Issues (20)
- Failing to get the weights from the dedicated github repo HOT 5
- Use ProteinBERT with Own Dataset HOT 3
- Original h5 file HOT 5
- loss plot during pretraining HOT 1
- signal peptide detection HOT 1
- KeyError: "Unable to open object (object 'test_set_mask' doesn't exist)" HOT 6
- How to extract the embedding of an amino acid? HOT 10
- Graph execution error HOT 6
- Extract local and global representation using finetune model HOT 1
- Running Benchmarks HOT 4
- Evaluation on larger data set HOT 6
- Using vector representations in the "weights" parameter in the "embedding" section of an LSTM model after fine-tuning my own data HOT 1
- Failing to extract global embedding (1,15599) -> (1,512) HOT 1
- What do the settings mean? HOT 3
- Error when trying to run the finetuning code given in the jupyter notebook HOT 2
- ValueError, set_weights error
- model_generation.py list is not callable error HOT 2
- GO annotations during fine tuning HOT 1
- Missing MajorPTMs train CSV file HOT 1
- Can't get proteinBERT to run on GPU HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from protein_bert.