Skip to content

[R][CI] Some R CI jobs seem unable to access some S3 files on arrow-datasets bucket #49622

@raulcd

Description

@raulcd

Describe the bug, including details regarding any error messages, version, and platform.

Some R jobs

have started failing with:

══ Failed tests ════════════════════════════════════════════════════════════════
── Error ('test-s3.R:49:5'): read/write Feather on S3 ──────────────────────────
Error: IOError: When initiating multiple part upload for key '1774919416.26534/test.feather' in bucket 'arrow-datasets': AWS Error ACCESS_DENIED during CreateMultipartUpload operation: User: arn:aws:iam::855673865593:user/crossbow is not authorized to perform: s3:PutObject on resource: "arn:aws:s3:::arrow-datasets/1774919416.26534/test.feather" because no identity-based policy allows the s3:PutObject action (Request ID: 63QMV381T6QZFXAC)
Backtrace:
    ▆
 1. └─arrow::write_feather(example_data, bucket_uri(now, "test.feather")) at test-s3.R:49:5
 2.   └─arrow:::make_output_stream(sink)
 3.     └─fs_and_path$fs$OpenOutputStream(fs_and_path$path)
 4.       └─arrow:::fs___FileSystem__OpenOutputStream(self, clean_path_rel(path))
── Error ('test-s3.R:55:5'): read/write Parquet on S3 ──────────────────────────
Error: IOError: When initiating multiple part upload for key '1774919416.26534/test.parquet' in bucket 'arrow-datasets': AWS Error ACCESS_DENIED during CreateMultipartUpload operation: User: arn:aws:iam::855673865593:user/crossbow is not authorized to perform: s3:PutObject on resource: "arn:aws:s3:::arrow-datasets/1774919416.26534/test.parquet" because no identity-based policy allows the s3:PutObject action (Request ID: 63QRS2B11WH1QXR2)
Backtrace:
    ▆
 1. └─arrow::write_parquet(example_data, bucket_uri(now, "test.parquet")) at test-s3.R:55:5
 2.   └─arrow:::make_output_stream(sink)
 3.     └─fs_and_path$fs$OpenOutputStream(fs_and_path$path)
 4.       └─arrow:::fs___FileSystem__OpenOutputStream(self, clean_path_rel(path))
── Error ('test-s3.R:60:5'): RandomAccessFile$ReadMetadata() works for S3FileSystem ──
Error: IOError: Path does not exist 'arrow-datasets/1774919416.26534/test.parquet'. Detail: [errno 2] No such file or directory
Backtrace:
    ▆
 1. └─bucket$OpenInputFile(paste0(now, "/", "test.parquet")) at test-s3.R:60:5
 2.   └─arrow:::fs___FileSystem__OpenInputFile(self, clean_path_rel(path))
── Error ('test-s3.R:44:1'): (code run outside of `test_that()`) ───────────────
Error: IOError: Path does not exist 'arrow-datasets/1774919416.26534/'. Detail: [errno 2] No such file or directory
Backtrace:
    ▆
 1. └─bucket$DeleteDir(now) at test-s3.R:44:1
 2.   └─arrow:::fs___FileSystem__DeleteDir(self, clean_path_rel(path))

[ FAIL 4 | WARN 1 | SKIP 30 | PASS 8075 ]
Error:
! Test failures.
Warning messages:
1: In x :
  closing unused connection 4 (/tmp/RtmpWY0ydv/file8f9b1d551c62/part-0.csv)
2: In for (t in seq_along(app$themes)) { :
  closing unused connection 5 (/tmp/RtmpWY0ydv/file8f9b3c3f3a1/part-0.tsv)
3: In for (t in seq_along(app$themes)) { :
  closing unused connection 4 (/tmp/RtmpWY0ydv/file8f9ba15a480/part-0.csv)
Execution halted

Component(s)

Continuous Integration, R

Metadata

Metadata

Assignees

Type

No type
No fields configured for issues without a type.

Projects

No projects

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions