Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
Discussion options

Can I use the Autotuning feature while enabling AutoTP in the configuration file?

I previously attempted to use both AutoTP and autotuning together. It appeared to work when I set "mp_size": 1 in the autotuning configuration. However, it failed when mp_size was set to match autotp_size. The error seemed related to a train batch size mismatch.

Even when it did run, the results were questionable. For instance, with the same train_micro_batch_size_per_gpu, I observed higher throughput when "autotp_size": 2" compared to no tensor parallelism at all, which seemed counterintuitive.

Additionally, it looks like the initial model profiling phase in autotuning uses ZeRO Stage 3 to estimate the minimum memory requirement. But AutoTP isn’t compatible with ZeRO Stage 3, right? I worked around this limitation by copying the profiling results from a configuration without TP, tricking autotuning into thinking the model had successfully completed profiling. However, this workaround isn’t ideal.

You must be logged in to vote

Replies: 2 comments

Comment options

@inkcherry @delock @tjruwase @loadams

You must be logged in to vote
0 replies
Comment options

I didn't used Autotuning before but it looks like autotuning focus on configuration file so it should support autotp. From your description, the problem should be caused by 'mp_size' not compatible with 'autotp_size'. @inkcherry can confirm whether this is true.

For the second question, is it possible to set autotp_size to 1 during the profile stage? If the purpose is find minimum memory usage, then zero3 without autotp should be a pretty good estimation.

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
🙏
Q&A
Labels
None yet
2 participants
Morty Proxy This is a proxified and sanitized view of the page, visit original site.