Skip to content
This repository was archived by the owner on Aug 7, 2024. It is now read-only.

Commit 7633925

Browse files
vkuzofacebook-github-bot
authored andcommitted
move dynamic linear to main directory (#180)
Summary: perf and accuracy are great, so makes more sense to have this at the same level of importance as `Float8Linear` Pull Request resolved: #180 Test Plan: ``` ./test/test_everything.sh ``` Reviewed By: drisspg Differential Revision: D52648574 Pulled By: vkuzo fbshipit-source-id: cc304c56360731cc8fe4b8bae5418683be75c684
1 parent dd0c596 commit 7633925

File tree

4 files changed

+2
-8
lines changed

4 files changed

+2
-8
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -143,7 +143,7 @@ We are working on a new FSDP implementation that uses per-parameter sharding tha
143143
# code tips
144144

145145
* `float8_experimental/float8_linear.py` - `Float8Linear` (main user facing entry point for delayed scaling)
146-
* `float8_experimental/dynamic_linear/dynamic_linear.py` - `Float8DynamicLinear` (main user facing entry point for dynamic scaling)
146+
* `float8_experimental/float8_dynamic_linear.py` - `Float8DynamicLinear` (main user facing entry point for dynamic scaling)
147147
* `float8_experimental/float8_tensor.py` - `Float8Tensor`, which allows `Float8Linear` to abide by the `x.dtype == x.grad.dtype` restriction
148148
* `float8_experimental/tp_linear.py` - `Float8ColumnParallelLinear` / `Float8RowParallelLinear` (TP/SP versions of float8 linear)
149149

float8_experimental/dynamic_linear/__init__.py

Lines changed: 0 additions & 6 deletions
This file was deleted.

float8_experimental/float8_linear_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88

99
import torch
1010
import torch.distributed as dist
11-
from float8_experimental.dynamic_linear import Float8DynamicLinear
11+
from float8_experimental.float8_dynamic_linear import Float8DynamicLinear
1212
from float8_experimental.float8_linear import Float8Linear
1313

1414
from float8_experimental.float8_utils import amax_history_to_scale, tensor_to_amax

0 commit comments

Comments
 (0)