Skip to content

bf16 scale/bias for INT4 (#5595)#5595

Closed
jeetkanjani7 wants to merge 1 commit intopytorch:mainfrom
jeetkanjani7:export-D95859348
Closed

bf16 scale/bias for INT4 (#5595)#5595
jeetkanjani7 wants to merge 1 commit intopytorch:mainfrom
jeetkanjani7:export-D95859348

Conversation

@jeetkanjani7
Copy link
Copy Markdown
Contributor

@jeetkanjani7 jeetkanjani7 commented Apr 8, 2026

Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: https://github.com/facebookresearch/FBGEMM/pull/2551

Reviewed By: zhaozhul

Differential Revision: D95859348

@meta-codesync
Copy link
Copy Markdown
Contributor

meta-codesync Bot commented Apr 8, 2026

@jeetkanjani7 has exported this pull request. If you are a Meta employee, you can view the originating Diff in D95859348.

@meta-cla meta-cla Bot added the cla signed label Apr 8, 2026
@meta-codesync meta-codesync Bot changed the title bf16 scale/bias for INT4 bf16 scale/bias for INT4 (#5595) Apr 13, 2026
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 13, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 13, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 13, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: https://github.com/facebookresearch/FBGEMM/pull/2551

Pull Request resolved: pytorch#5595

Differential Revision: D95859348
@jeetkanjani7 jeetkanjani7 force-pushed the export-D95859348 branch 2 times, most recently from 39d74bc to b65c26e Compare April 13, 2026 06:28
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 13, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: https://github.com/facebookresearch/FBGEMM/pull/2551

Pull Request resolved: pytorch#5595

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 21, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 23, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 23, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Differential Revision: D95859348
@jeetkanjani7 jeetkanjani7 force-pushed the export-D95859348 branch 2 times, most recently from eb5b190 to d0aa8bb Compare April 24, 2026 17:48
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: https://github.com/facebookresearch/FBGEMM/pull/2551

Pull Request resolved: pytorch#5595

Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: https://github.com/facebookresearch/FBGEMM/pull/2551

Pull Request resolved: pytorch#5595

Reviewed By: zhaozhul

Differential Revision: D95859348
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 24, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
@jeetkanjani7 jeetkanjani7 force-pushed the export-D95859348 branch 2 times, most recently from 4db39c2 to 3c73db9 Compare April 27, 2026 21:37
jeetkanjani7 added a commit to jeetkanjani7/FBGEMM-1 that referenced this pull request Apr 27, 2026
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
Summary:
Add bf16 scale/bias support for INT4/INT2 fused N-bit rowwise quantization in FBGEMM and SilverTorch.

Previously, fused N-bit rowwise quantization only supported fp16 scale/bias - storing scale/bias in bf16 avoids precision loss from fp16 truncation during quantization round-trips.

X-link: facebookresearch/FBGEMM#2551


Reviewed By: zhaozhul

Differential Revision: D95859348
@meta-codesync
Copy link
Copy Markdown
Contributor

meta-codesync Bot commented Apr 29, 2026

This pull request has been merged in 939f2da.

@facebook-github-tools
Copy link
Copy Markdown

This pull request has been reverted by 5e1dde6.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant