You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* Consolidate metadata during op conversion
* Unmark gpt2 and mnist test models to expect passing
* Disable conversion from aten._to_copy
* Pass device for all from_torch ops
* Replace aten.full op to a literal scalar for certain cases
* Compare only Tensor types for dictionary outputs
* Replace aten.view with aten.reshape
* Unmark bloom, llama, and yolos from xfail
* Add conversion for aten.min
* Add exception to aten.eq conversion
* Fix reusing ttnn data movement op if mixed with aten ops
* Convert all inputs to ttnn.bfloat16 when moving data in
* Skip unsqueeze transformation if last dim of input is not the same as last dim of output
* Add exception to aten.expand conversion when last dimension of input is 1
* Support list type arguments
* Check layout change for ttnn reshape and embedding op
* Freeze encoder for llama model
* Add workaround for ttnn.permute when dim 0 is 1 for rank 3
* Reconvert int64 types from metadata when mixing ttnn and aten ops
* Check for valid page size for ops that decompose to ttnn.full
* Delete aten.expand op if output has the exact same shape
* Mark GPT-2 model as xfail
* Update README with new model stats
* Fix output type of aten.arange unit test to match output of original
* Disable to_copy unit test to re-evaluate conversion
* Lower pcc for addmm slightly
* Change input shapes of some unit test to match exceptions in current state of lowering
* Fix page size validation for conversions involving ttnn.full ops
* Update README
* Revert changes to GPT-2 since it isn't working in this PR
* Remove commented out code
* Revert "Pass device for all from_torch ops"
This reverts commit 775fb9f.
@@ -8,17 +8,17 @@ This project allows to run PyTorch code on [Tenstorrent](https://tenstorrent.com
8
8
9
9
The table below summarizes the results of running various ML models through our TTNN compiler. For each model, we track whether the run was successful, the number of operations before and after conversion, the number of `to_device` and `from_device` operations, performance metrics, and accuracy.
10
10
11
-
| Model | Run Success | Torch Ops Before (Unique Ops) | Torch Ops Remain (Unique Ops) | To/From Device Ops| Original Run Time (ms) | Compiled Run Time (ms) | Accuracy (%) |
| Model | Run Success | Torch Ops Before (Unique Ops) | Torch Ops Remain (Unique Ops) |To/From Device Ops | Original Run Time (ms) | Compiled Run Time (ms) | Accuracy (%) |
0 commit comments