Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] Shouldnt the LoRA be the same float as the base model? #796

Open
2 tasks done
jackswl opened this issue Oct 17, 2024 · 1 comment
Open
2 tasks done

[BUG] Shouldnt the LoRA be the same float as the base model? #796

jackswl opened this issue Oct 17, 2024 · 1 comment
Labels
bug Something isn't working stale

Comments

@jackswl
Copy link

jackswl commented Oct 17, 2024

Prerequisites

  • I have read the documentation.
  • I have checked other issues for similar problems.

Backend

Local

Interface Used

CLI

CLI Command

No response

UI Screenshots & Parameters

No response

Error Logs

    logger.info("Loading adapter...")
    base_model = AutoModelForCausalLM.from_pretrained(
        base_model_path,
        torch_dtype=torch.float16,
        low_cpu_mem_usage=True,
        trust_remote_code=ALLOW_REMOTE_CODE,
        token=token,
    )

This is the code for tools for merge_adapters

if my base model is in bfloat16, the merging should be bfloat16 too. but here, i cannot specify the base model torch_dtype. Wouldnt this be an issue? @abhishekkrthakur

Additional Information

No response

@jackswl jackswl added the bug Something isn't working label Oct 17, 2024
Copy link

This issue is stale because it has been open for 30 days with no activity.

@github-actions github-actions bot added the stale label Nov 17, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working stale
Projects
None yet
Development

No branches or pull requests

1 participant