You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-issues@hadoop.apache.org by "Alessandro Passaro (Jira)" <ji...@apache.org> on 2022/10/18 15:35:00 UTC

[jira] [Commented] (HADOOP-18183) s3a audit logs to publish range start/end of GET requests in audit header

    [ https://issues.apache.org/jira/browse/HADOOP-18183?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17619658#comment-17619658 ] 

Alessandro Passaro commented on HADOOP-18183:
---------------------------------------------

+1 for adding the whole range header. This is also how the AWS SDK v2 handles ranges (rather than explicitly exposing start & end), so it will simplify the upgrade.

> s3a audit logs to publish range start/end of GET requests in audit header
> -------------------------------------------------------------------------
>
>                 Key: HADOOP-18183
>                 URL: https://issues.apache.org/jira/browse/HADOOP-18183
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: 3.3.2
>            Reporter: Steve Loughran
>            Assignee: Ankit Saurabh
>            Priority: Minor
>
> we don't get the range of ranged get requests in s3 server logs, because the AWS s3 log doesn't record that information. we can see it's a partial get from the 206 response, but the length of data retrieved is lost.
> LoggingAuditor.beforeExecution() would need to recognise a ranged GET and determine the extra key-val pairs for range start and end (rs & re?)
> we might need to modify {{HttpReferrerAuditHeader.buildHttpReferrer()}} to take a map of <string, string> so it can dynamically create a header for each request; currently that is not in there.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-issues-help@hadoop.apache.org