File tree 3 files changed +21
-3
lines changed
3 files changed +21
-3
lines changed Original file line number Diff line number Diff line change @@ -219,11 +219,9 @@ CUDA_VISIBLE_DEVICES=${DEVICES} torchrun --nproc_per_node=2 --master_port ${MAST
219
219
220
220
# Citation
221
221
222
- If you use the code or data in your research, please use the following BibTex entry:
223
-
224
222
```
225
223
@inproceedings{
226
- munoz2024eftnas ,
224
+ eftnas2024 ,
227
225
title={Searching for Efficient Language Models in First-Order Weight-Reordered Super-Networks},
228
226
author={J. Pablo Munoz and Yi Zheng and Nilesh Jain},
229
227
booktitle={The 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation},
Original file line number Diff line number Diff line change @@ -119,4 +119,18 @@ CUDA_VISIBLE_DEVICES=${DEVICES} python run_math.py \
119
119
--lora \
120
120
--lora_weights trained_super_adapter/unified_math_10k/lonas-llama-7b \
121
121
--nncf_config nncf_config/unified_math_10k/nncf_lonas_llama_7b.json
122
+ ```
123
+
124
+
125
+ # Citation
126
+
127
+ ```
128
+ @inproceedings{
129
+ lonas2024,
130
+ title={LoNAS: Elastic Low-Rank Adapters for Efficient Large Language Models},
131
+ author={J. Pablo Munoz and Jinjie Yuan and Yi Zheng and Nilesh Jain},
132
+ booktitle={The 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation},
133
+ year={2024},
134
+ url={}
135
+ }
122
136
```
Original file line number Diff line number Diff line change @@ -10,6 +10,12 @@ EZNAS is a genetic programming-driven methodology for automatically discovering
10
10
11
11
BootstrapNAS automates the generation of weight-sharing super-networks using the Neural Network Compression Framework (NNCF).
12
12
13
+ ### [ LoNAS: Elastic Low-Rank Adapters for Efficient Large Language Models] ( ./LoNAS/README.md )
14
+
15
+ Initial exploration of using NAS on large language models by exploring a search space of elastic low-rank adapters
16
+ while reducing memory and compute requirements of full-scale NAS, resulting in high-performing compressed models obtained from
17
+ weight-sharing super-networks. We investigate the benefits and limitations of this method, motivating follow-up work.
18
+
13
19
### [ EFTNAS: Searching for Efficient Language Models in First-Order Weight-Reordered Super-Networks] ( ./EFTNAS/README.md )
14
20
15
21
Integrating neural architecture search (NAS) and network pruning techniques, we effectively generate and train
You can’t perform that action at this time.
0 commit comments