Skip to content

Commit

Permalink
deploy: c162617
Browse files Browse the repository at this point in the history
  • Loading branch information
lbluque committed Jan 21, 2025
1 parent fcbf3a5 commit db4d8d1
Show file tree
Hide file tree
Showing 449 changed files with 292,947 additions and 47 deletions.
208 changes: 208 additions & 0 deletions _downloads/5fdddbed2260616231dbf7b0d94bb665/train.txt

Large diffs are not rendered by default.

140 changes: 140 additions & 0 deletions _downloads/819e10305ddd6839cd7da05935b17060/mass-inference.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,140 @@
2025-01-21 19:37:23 /home/runner/work/fairchem/fairchem/src/fairchem/core/_cli.py:131: (INFO): Running in local mode without elastic launch (single gpu only)
2025-01-21 19:37:23 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/utils.py:1023: (INFO): Setting env PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True
2025-01-21 19:37:23 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/utils.py:358: (INFO): Project root: /home/runner/work/fairchem/fairchem/src/fairchem
/home/runner/work/fairchem/fairchem/src/fairchem/core/models/scn/spherical_harmonics.py:23: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
_Jd = torch.load(os.path.join(os.path.dirname(__file__), "Jd.pt"))
/home/runner/work/fairchem/fairchem/src/fairchem/core/models/equiformer_v2/wigner.py:10: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
_Jd = torch.load(os.path.join(os.path.dirname(__file__), "Jd.pt"))
/home/runner/work/fairchem/fairchem/src/fairchem/core/models/escn/so3.py:23: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
_Jd = torch.load(os.path.join(os.path.dirname(__file__), "Jd.pt"))
2025-01-21 19:37:24 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:208: (INFO): amp: false
cmd:
checkpoint_dir: /home/runner/work/fairchem/fairchem/docs/core/checkpoints/2025-01-21-19-37-36
commit: c162617
identifier: ''
logs_dir: /home/runner/work/fairchem/fairchem/docs/core/logs/tensorboard/2025-01-21-19-37-36
print_every: 10
results_dir: /home/runner/work/fairchem/fairchem/docs/core/results/2025-01-21-19-37-36
seed: 0
timestamp_id: 2025-01-21-19-37-36
version: 0.1.dev1+gc162617
dataset: {}
evaluation_metrics:
metrics:
energy:
- mae
forces:
- forcesx_mae
- forcesy_mae
- forcesz_mae
- mae
- cosine_similarity
- magnitude_error
misc:
- energy_forces_within_threshold
primary_metric: forces_mae
gp_gpus: null
gpus: 0
logger: tensorboard
loss_functions:
- energy:
coefficient: 1
fn: mae
- forces:
coefficient: 1
fn: l2mae
model:
activation: silu
cbf:
name: spherical_harmonics
cutoff: 6.0
direct_forces: true
emb_size_atom: 512
emb_size_bil_trip: 64
emb_size_cbf: 16
emb_size_edge: 512
emb_size_rbf: 16
emb_size_trip: 64
envelope:
exponent: 5
name: polynomial
extensive: true
max_neighbors: 50
name: gemnet_t
num_after_skip: 2
num_atom: 3
num_before_skip: 1
num_blocks: 3
num_concat: 1
num_radial: 128
num_spherical: 7
otf_graph: true
output_init: HeOrthogonal
rbf:
name: gaussian
regress_forces: true
optim:
batch_size: 16
clip_grad_norm: 10
ema_decay: 0.999
energy_coefficient: 1
eval_batch_size: 16
eval_every: 5000
force_coefficient: 1
loss_energy: mae
loss_force: atomwisel2
lr_gamma: 0.8
lr_initial: 0.0005
lr_milestones:
- 64000
- 96000
- 128000
- 160000
- 192000
max_epochs: 80
num_workers: 2
optimizer: AdamW
optimizer_params:
amsgrad: true
warmup_steps: -1
outputs:
energy:
level: system
forces:
eval_on_free_atoms: true
level: atom
train_on_free_atoms: true
relax_dataset: {}
slurm: {}
task:
prediction_dtype: float32
test_dataset:
a2g_args:
r_energy: false
r_forces: false
format: ase_db
select_args:
selection: natoms>5,xc=PBE
src: data.db
trainer: ocp
val_dataset: {}

2025-01-21 19:37:24 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:557: (INFO): Loading model: gemnet_t
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:568: (INFO): Loaded GemNetT with 31671825 parameters.
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/logger.py:157: (WARNING): log_summary for Tensorboard not supported
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/datasets/base_dataset.py:94: (WARNING): Could not find dataset metadata.npz files in '[PosixPath('data.db')]'
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/data_parallel.py:164: (WARNING): Disabled BalancedBatchSampler because num_replicas=1.
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/data_parallel.py:177: (WARNING): Failed to get data sizes, falling back to uniform partitioning. BalancedBatchSampler requires a dataset that has a metadata attributed with number of atoms.
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/data_parallel.py:88: (INFO): rank: 0: Sampler created...
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/data_parallel.py:196: (INFO): Created BalancedBatchSampler with sampler=<fairchem.core.common.data_parallel.StatefulDistributedSampler object at 0x7fcddcc00b90>, batch_size=16, drop_last=False
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/tasks/task.py:38: (INFO): Attemping to load user specified checkpoint at /tmp/fairchem_checkpoints/gndt_oc22_all_s2ef.pt
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:606: (INFO): Loading checkpoint from: /tmp/fairchem_checkpoints/gndt_oc22_all_s2ef.pt
/home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:607: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
checkpoint = torch.load(checkpoint_path, map_location=map_location)
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:636: (INFO): Overwriting scaling factors with those loaded from checkpoint. If you're generating predictions with a pretrained checkpoint, this is the correct behavior. To disable this, delete `scale_dict` from the checkpoint.
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/modules/scaling/compat.py:77: (WARNING): Scale factor comment not found in model
2025-01-21 19:37:26 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/ocp_trainer.py:445: (INFO): Predicting on test.
device 0: 0%| | 0/3 [00:00<?, ?it/s]device 0: 33%|███████████▋ | 1/3 [00:03<00:06, 3.19s/it]device 0: 67%|███████████████████████▎ | 2/3 [00:06<00:03, 3.08s/it]device 0: 100%|███████████████████████████████████| 3/3 [00:06<00:00, 1.99s/it]device 0: 100%|███████████████████████████████████| 3/3 [00:06<00:00, 2.30s/it]
2025-01-21 19:37:33 /home/runner/work/fairchem/fairchem/src/fairchem/core/trainers/base_trainer.py:1003: (INFO): Writing results to /home/runner/work/fairchem/fairchem/docs/core/results/2025-01-21-19-37-36/ocp_predictions.npz
2025-01-21 19:37:33 /home/runner/work/fairchem/fairchem/src/fairchem/core/common/utils.py:1112: (INFO): Total time taken: 7.040971517562866
Elapsed time = 14.4 seconds
Expand Down
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
18 changes: 3 additions & 15 deletions _sources/autoapi/core/datasets/base_dataset/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,6 @@ Classes

.. autoapisummary::

core.datasets.base_dataset.DatasetMetadata
core.datasets.base_dataset.BaseDataset
core.datasets.base_dataset.Subset

Expand All @@ -50,17 +49,6 @@ Module Contents

.. py:data:: T_co
.. py:class:: DatasetMetadata
Bases: :py:obj:`NamedTuple`


.. py:attribute:: natoms
:type: numpy.typing.ArrayLike | None
:value: None



.. py:exception:: UnsupportedDatasetError
Bases: :py:obj:`ValueError`
Expand Down Expand Up @@ -100,14 +88,14 @@ Module Contents
.. py:property:: _metadata
:type: DatasetMetadata
:type: dict[str, numpy.typing.ArrayLike]



.. py:method:: get_metadata(attr, idx)
.. py:class:: Subset(dataset: BaseDataset, indices: collections.abc.Sequence[int], metadata: DatasetMetadata | None = None)
.. py:class:: Subset(dataset: BaseDataset, indices: collections.abc.Sequence[int], metadata: dict[str, numpy.typing.ArrayLike])
Bases: :py:obj:`torch.utils.data.Subset`, :py:obj:`BaseDataset`

Expand All @@ -128,7 +116,7 @@ Module Contents
.. py:property:: _metadata
:type: DatasetMetadata
:type: dict[str, numpy.typing.ArrayLike]



Expand Down
3 changes: 0 additions & 3 deletions _sources/autoapi/core/datasets/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -404,8 +404,5 @@ Package Contents
.. py:method:: __del__()
.. py:method:: sample_property_metadata(num_samples: int = 100)
.. py:function:: data_list_collater(data_list: list[torch_geometric.data.data.BaseData], otf_graph: bool = False, to_dict: bool = False) -> torch_geometric.data.data.BaseData | dict[str, torch.Tensor]
3 changes: 0 additions & 3 deletions _sources/autoapi/core/datasets/lmdb_dataset/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -81,8 +81,5 @@ Module Contents
.. py:method:: __del__()
.. py:method:: sample_property_metadata(num_samples: int = 100)
.. py:function:: data_list_collater(data_list: list[torch_geometric.data.data.BaseData], otf_graph: bool = False, to_dict: bool = False) -> torch_geometric.data.data.BaseData | dict[str, torch.Tensor]
Original file line number Diff line number Diff line change
Expand Up @@ -201,6 +201,9 @@ Module Contents
.. py:method:: train(disable_eval_tqdm=False)
Run model training iterations.



.. py:method:: _compute_loss(out, batch)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ Module Contents
Scales up the energy and then scales down the forces
to prevent NaNs and infs in calculations using AMP.
Inspired by torch.cuda.amp.GradScaler.
Inspired by torch.GradScaler("cuda", args...).


.. py:attribute:: scale_factor
Expand Down
Loading

0 comments on commit db4d8d1

Please sign in to comment.