Skip to content

Conversation

@ggggxm
Copy link
Contributor

@ggggxm ggggxm commented May 9, 2025

PR Category

Operator Mechanism

PR Types

Bug fixes

Description

  • 修复如下算子的相关shape type
    • paddle.Flatten
    • paddle.Split 因为paddle.vector_to_parameters会调用split api,因此它也被修复
  • 性能分析
    • 修改前后波动在1%以内

@paddle-bot
Copy link

paddle-bot bot commented May 9, 2025

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@paddle-bot paddle-bot bot added the contributor External developers label May 9, 2025
@lshpku lshpku changed the title 【PHI】 Fix Shape Int overflow for flatten/split/expand_as [PHI] Fix shape int overflow for flatten/split/expand_as May 9, 2025
DenseTensor* out) {
std::vector<int> target_shape = target_shape_t;
template <typename T, typename Context, typename ShapeType>
void ExpandAsKernelInner(const Context& ctx,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

感觉这个叫ExpandAsKernelImpl好一点,虽然Inner也有人用,但是我还是觉得Impl更贴切

Comment on lines 98 to 101
bool use_int64 =
std::any_of(target_shape_t.begin(), target_shape_t.end(), [](int64_t v) {
return v > static_cast<int64_t>(std::numeric_limits<int32_t>::max());
});
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

如果每个shape[i]都不超过,但是乘在一起超过了会怎么样?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

expand_as只需要对target_shape的形状逐个进行判断,再进行broadcast操作,应该不影响?单个测试用例
x = paddle.ones([1, 1], dtype="bfloat16")
y = paddle.ones([1281701379, 3], dtype="bfloat16")
paddle.Tensor.expand_as(x, y)
是没有问题的

@ggggxm ggggxm changed the title [PHI] Fix shape int overflow for flatten/split/expand_as [PHI] Fix shape int overflow for flatten and split May 16, 2025
@lshpku lshpku merged commit 9a6ab46 into PaddlePaddle:develop May 19, 2025
51 of 52 checks passed
wanghuancoder pushed a commit to wanghuancoder/Paddle that referenced this pull request May 27, 2025
wanghuancoder added a commit that referenced this pull request Jun 3, 2025
* refine forrange (#72360)

* refine forrange

* refine forrange

* reduce support big tensor (#71970)

* reduce support big tensor

* [PHI] Fix gridDim limit for reduce kernel (#72507)

* [API] isclose support bigtensor (#72516)

* isclose support bigtensor

* refine

* [API] isnan isinf isfinite support bigtensor (#72517)

* isnan isinf isfinite support bigtensor

* refine

* [PHI] Fix cum kernel for big tensor (#72562)

* [PHI] Preliminary fix for elementwise broadcast int32 shape overflow (#72584)

* [PHI] Align linalg.solve kernel with torch (#72608)

* Update strided copy kernel (#72662)

* [PHI] Fix grid sample kernel for big tensor (#72628)

* [PHI] Fix argsort big tensor bug (#72712)

* [PHI] Fixed argsort big tensor bug

* [PHI] Fixed shape mismatch problem.

* [PHI] Fix contiguous kernel for big tensor (#72705)

* [PHI] Fix flatten and split kernel for big tensor (#72634)

* [PHI] Fix out-of-bound issue of paddle.take_along_axis (#72757)

* [PHI] fix paddle.diag with big tensor (#72638)

* [API] fix paddle.cross with big tensor (#72652)

* [PHI] Fix paddle.where api for big tensor (#72717)

* [PHI] Fix bincount kernel for big tensor (#72706)

* fix bincount kernel for big tensor

* use HostAlloc to alloc memory

* add cpu test case

* [PHI] Fix full_like kernel for big tensor (#72831)

* [API] Fix int overflow and float16 support for paddle.frac (#72815)

* [PHI] Align paddle.inner with torch in matmul logic (#72843)

* [PHI] Fix paddle.var & paddle.std float16 overflow (#72650)

* [PHI] Fix logsumexp precision problem (#72681)

* [PHI] Debug for logsumexp, bug source found

* [PHI] Removed GetNumBlocks func to get correct logsumexp

* [PHI] Removed redundant debug VLOG

* [PHI] Elegant grid bounded solution

* [Accuracy diff No.55-56、76-77] Fix accuracy diff for var&std API (#72879)

* [Accuracy diff No.21] Fix accuracy diff for heaviside API (#72894)

---------

Co-authored-by: Shuhao Liang <[email protected]>
Co-authored-by: Qianyue He <[email protected]>
Co-authored-by: Lei Ding <[email protected]>
Co-authored-by: ggggxm <[email protected]>
Co-authored-by: xkkkkkk23 <[email protected]>
Co-authored-by: Zx <[email protected]>
Co-authored-by: huangjiyi <[email protected]>
Co-authored-by: ooo oo <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

contributor External developers

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants