Giter VIP home page Giter VIP logo

airs's Introduction

logo
Logo by Zhao Xu

The Data Integration, Visualization, and Exploration (DIVE) Laboratory at Texas A&M University is led by Dr. Shuiwang Ji and conducts foundational research in machine learning and deep learning and applies machine learning methods to solve challenging real-world problems in biology, chemistry, neuroscience and medicine.

Highlighted Work

airs's People

Contributors

alexandermath avatar alliesaxton03 avatar congffu avatar divelab avatar eltociear avatar floatlazer avatar hongyiling avatar hyanan16 avatar jacobhelwig avatar jacoblau0513 avatar kruskallin avatar limei0307 avatar lyzustc avatar mengliu1998 avatar montgomerybohde avatar oceanusity avatar wangyucheng1234 avatar ycremar avatar ykq98 avatar zoexu119 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

airs's Issues

nothing

nothing,i have some mistake。

QH9: Hamiltonians for the same molecule are very similar.

The MD trajectories almost don't change (see attached video). The QHNet baseline gets 70[μHa] MAE. For the first molecule, the MAE between all 60 Hamiltonians is 45[μHa] (see attached image for all pair differences). This might be caused by the small ~2 attosecond time step. In your reply to reviewer wfG7 you mention a dataset with larger time steps. Any chance you'll publish this within a few days?

Please don't hesitate to let me know if I'm misunderstanding something.

video: https://github.com/divelab/AIRS/assets/8614529/63fd3109-e88f-4b4a-ae4b-eb0a416a87eb
image

Strange Results

Hello, Thanks for your nice paper! I use main_gen.py to generate ligands for the data in [https://github.com/pengxingang/Pocket2Mol/blob/main/data/test_list.tsv], which is also in the crossdocked2020 dataset, but lots of the ligands I generate are not in the protein pocket (Figure 1 shows the structure of protein pocket with the reference ligand, and Figure 2 shows the structure of protein pocket with the generated ligands) .I would like to ask if these results are correct?
1
2

PotNet enviroment set up

Hi, I meet an error when running conda env create -f environment.yml.

Here is the error message

ERROR: Could not find a version that satisfies the requirement torch-cluster==1.6.0+pt112cu116 (from versions: 0.1.1, 0.2.3, 0.2.4, 1.0.1, 1.0.3, 1.1.1, 1.1.2, 1.1.3, 1.1.4, 1.1.5, 1.2.1, 1.2.2, 1.2.3, 1.2.4, 1.3.0, 1.4.0, 1.4.1, 1.4.2, 1.4.3a1, 1.4.3, 1.4.4, 1.4.5, 1.5.2, 1.5.3, 1.5.4, 1.5.5, 1.5.6, 1.5.7, 1.5.8, 1.5.9, 1.6.0, 1.6.1)

ERROR: No matching distribution found for torch-cluster==1.6.0+pt112cu116

failed

CondaEnvException: Pip failed

A question about irreps in QHNet

Hi. I have a question about irreps in QHNet. In QHNet, it sets self.hidden_irrep = o3.Irreps(f'{self.hs}x0e + {self.hs}x1o + {self.hs}x2e + {self.hs}x3o + {self.hs}x4e') and self.hidden_irrep_base = o3.Irreps(f'{self.hs}x0e + {self.hs}x1e + {self.hs}x2e + {self.hs}x3e + {self.hs}x4e'). These layers are defined as

self.e3_gnn_layer.append(ConvNetLayer(
irrep_in_node=input_irrep,
irrep_hidden=self.hidden_irrep,
irrep_out=self.hidden_irrep,
edge_attr_dim=self.radius_embed_dim,
node_attr_dim=self.hs,
sh_irrep=self.sh_irrep,
resnet=True,
use_norm_gate=True if i != 0 else False
))
if i > self.start_layer:
self.e3_gnn_node_layer.append(SelfNetLayer(
irrep_in_node=self.hidden_irrep_base,
irrep_bottle_hidden=self.hidden_irrep_base,
irrep_out=self.hidden_irrep_base,
sh_irrep=self.sh_irrep,
edge_attr_dim=self.radius_embed_dim,
node_attr_dim=self.hs,
resnet=True,
))
self.e3_gnn_node_pair_layer.append(PairNetLayer(
irrep_in_node=self.hidden_irrep_base,
irrep_bottle_hidden=self.hidden_irrep_base,
irrep_out=self.hidden_irrep_base,
sh_irrep=self.sh_irrep,
edge_attr_dim=self.radius_embed_dim,
node_attr_dim=self.hs,
invariant_layers=self.num_fc_layer,
invariant_neurons=self.hs,
resnet=True,
))

And in forward process:
node_attr = layer(data, node_attr)
if layer_idx > self.start_layer:
fii = self.e3_gnn_node_layer[layer_idx-self.start_layer-1](data, node_attr, fii)
fij = self.e3_gnn_node_pair_layer[layer_idx-self.start_layer-1](data, node_attr, fij)

The node_attr outputs from e3_gnn_layer will be the inputs of e3_gnn_node_layer and e3_gnn_node_pair_layer. I'm confused about why the output irreps of e3_gnn_layer is different from the input irreps of e3_gnn_node_layer and e3_gnn_node_pair_layer?

the training loss of QHNet did not drop

Hi! I was running the QHNet model using the QH9 dataset these days, but the training loss and validation loss did drop after a few steps. The mae for hamiltonian is 0.05485373 at the begining but it only dropped to 0.01 at the 40,000th step, which is quite weird.

I didn't change any of the code in the original codebase and the dataset was directly downloaded from the Google Drive. I run the main.py directly, and I couldn't find the reason why I can't reproduce the experimental results from the paper.

I wonder if you guys have any comments about that. Any help would be much appreciated!

Jarvis version update

Error happen for the model PotNet

for the version jarvis-tools-2022.9.16, we will meet an error when running
** from jarvis.db.figshare import data
dft_3d = data(dataset='dft_3d')**

Update to the latest version jarvis-tools-2024.4.10 will solve it

Error info :
Obtaining 3D dataset 55k ...
Reference:https://www.nature.com/articles/s41524-020-00440-1
Loading the zipfile...
Traceback (most recent call last):
File "/home/workspace/AIRS/OpenMat/PotNet/main.py", line 23, in
train_prop_model(data, data_root=args.data_root, checkpoint=args.checkpoint, testing=args.testing)
File "/home/workspace/AIRS/OpenMat/PotNet/train_prop.py", line 479, in train_prop_model
result = train_pyg(config, data_root=data_root, file_format=file_format, checkpoint=checkpoint, testing=testing)
File "/home/workspace/AIRS/OpenMat/PotNet/train_prop.py", line 208, in train_pyg
) = get_train_val_loaders(
File "/home/workspace/AIRS/OpenMat/PotNet/data.py", line 507, in get_train_val_loaders
d = jdata(dataset)
File "/home/anaconda3/envs/potnet/lib/python3.9/site-packages/jarvis/db/figshare.py", line 350, in data
dat = get_request_data(js_tag=js_tag, url=url)
File "/home/anaconda3/envs/potnet/lib/python3.9/site-packages/jarvis/db/figshare.py", line 303, in get_request_data
data = json.loads(zipfile.ZipFile(path).read(js_tag))
File "/home/anaconda3/envs/potnet/lib/python3.9/zipfile.py", line 1268, in init
self._RealGetContents()
File "/home/anaconda3/envs/potnet/lib/python3.9/zipfile.py", line 1335, in _RealGetContents
raise BadZipFile("File is not a zip file")
zipfile.BadZipFile: File is not a zip file

SMILES of QH9 entries

I have loaded QH9Stable using this code:
https://github.com/divelab/AIRS/blob/main/OpenDFT/QHBench/QH9/datasets.py

I observe the entries are PyG data objects.

Data(pos=[5, 3], atoms=[5, 1], diagonal_hamiltonian=[5, 14, 14], non_diagonal_hamiltonian=[20, 14, 14], diagonal_hamiltonian_mask=[5, 14, 14], non_diagonal_hamiltonian_mask=[20, 14, 14], edge_index_full=[2, 20])

How do I get the SMILES of a sample?
I want to join (in the database sense) the QH9 records with QM9 records by SMILES.
Thanks!

QH9: Reproducing Hamiltonians with PySCF gives 2[μHa] error.

I tried reproducing the Hamiltonians using PySCF. This gave me a MAE of 2[μHa].

Question 0. Any chance you could release the code to reproduce the dataset? (I likely just made a mistake)

Question 1. Did you see similar errors when creating the dataset?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.