log4j - If you don't use a Socket appender, you're good. Otherwise, you
can replace the log4j jars in lib/ with a newer version. You could also
upgrade to 1.11.1 which uses log4j2.
guava - We do not use Guava for serialization AFAIK. We also do not use
Java serialization for records.
commons-io - We do not use it for serialization.
protobuf - Only used by specific components for exchanging records. If
an attacker is able to introduce arbitrary data in places, then they in
any case wreak havoc in all sorts of ways, be it by introducing false
data invalidating your pipeline or sending obscenely large messages
eating up all memory.
all things contained in flink-shaded-hadoop - Please evaluate yourself
whether this is problematic and for, and use a newer Hadoop version if
required. Flink 1.11.0 also came with support for Hadoop 3.
Flink-shaded-hadoop is no longer supported, and was just a collection of
Hadoop dependencies that we package for convenience.
TL;DR: Please use your own judgement for issues regarding Hadoop, and
reach out to the Hadoop project for guidance. As for the remaining
issues, it is unlikely that these vulnerabilities make your setup any
less secure.
On 06/08/2020 14:06, V N, Suchithra (Nokia - IN/Bangalore) wrote:
Hello,
We are using Apache Flink 1.10.1 version. During our security scans following
issues are reported by our scan tool.
Please let us know your comments on these dependency vulnerabilities.
Thanks,
Suchithra
-----Original Message-----
From: m...@gsuite.cloud.apache.org <m...@gsuite.cloud.apache.org> On Behalf Of
Apache Security Team
Sent: Thursday, August 6, 2020 1:08 PM
To: V N, Suchithra (Nokia - IN/Bangalore) <suchithra....@nokia.com>
Cc: Jash, Shaswata (Nokia - IN/Bangalore) <shaswata.j...@nokia.com>; Prabhala, Anuradha
(Nokia - IN/Bangalore) <anuradha.prabh...@nokia.com>; Badagandi, Srinivas B. (Nokia -
IN/Bangalore) <srinivas.b.badaga...@nokia.com>
Subject: Re: Security vulnerabilities with Apache Flink 1.10.1 version
Hi,
Outdated dependencies are not always security issues. A project would only be
affected if a dependency was used in such a way that the affected underlying
code is used and the vulnerabilities were exposed.
We typically get reports sent to us from scanning tools that looks at
dependencies out of context on how they are actually used in the projects. As
such we reject these reports and suggest you either a) show how the product is
affected by the dependency vulnerabilities, or
b) simply mention this as a normal bug report to that project. Since
dependency vulnerabilities are quite public, there is no need to use this
private reporting mechanism for them.
Regards, Mark
On Thu, Aug 6, 2020 at 6:04 AM V N, Suchithra (Nokia - IN/Bangalore)
<suchithra....@nokia.com> wrote:
Hello,
We are using Apache Flink 1.10.1 version. During our security scans following
issues are reported by our scan tool.
1.Package : log4j-1.2.17
Severity: CRITICAL
Fix version: 2.8.2
Description:
Apache Log4j contains a flaw that is triggered as the SocketServer class
accepts log data from untrusted network traffic, which it then insecurely
deserializes. This may allow a remote attacker to potentially execute arbitrary
code.
Path:
/opt/flink/lib/log4j-1.2.17.jar
/opt/flink/bin/bash-java-utils.jar:log4j
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2019-17571
https://seclists.org/oss-sec/2019/q4/167
https://logging.apache.org/log4j/1.2/
2.Package: guava-14.0.1
Severity: HIGH
Fix version: 25.0, 24.1.1
Description:
Google Guava contains a flaw in the
CompoundOrdering_CustomFieldSerializer::deserialize() function in
com/google/common/collect/CompoundOrdering_CustomFieldSerializer.java that is
triggered when deserializing Java objects. With specially crafted serialized
data, a context-dependent can exhaust available memory, resulting in a denial
of service.
References:
https://github.com/google/guava/wiki/CVE-2018-10237
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237
3.Package: guava-14.0.1,18.0
Severity: HIGH
Fix version: 25.0, 24.1.1
Description:
Google Guava contains a flaw in the AtomicDoubleArray::readObject() function in
com/google/common/util/concurrent/AtomicDoubleArray.java that is triggered when
deserializing Java objects. With specially crafted serialized data, a
context-dependent can cause a process linked against the library to exhaust
available memory.
References:
https://github.com/google/guava/wiki/CVE-2018-10237
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237
4. Package: guava-19.0
Severity: HIGH
Fix version: 25.0, 24.1.1, 23.6.1
Description:
Google Guava contains a flaw in the
CompoundOrdering_CustomFieldSerializer::deserialize() function in
com/google/common/collect/CompoundOrdering_CustomFieldSerializer.java that is
triggered when deserializing Java objects. With specially crafted serialized
data, a context-dependent can exhaust available memory, resulting in a denial
of service.
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-10237
https://github.com/google/guava/wiki/CVE-2018-10237
Path:
/opt/flink/lib/flink-table-blink_2.11-1.10.1.jar:guava
/opt/flink/lib/flink-table_2.11-1.10.1.jar:guava
/opt/flink/lib/flink-dist_2.11-1.10.1.jar:guava
/opt/flink/examples/streaming/Twitter.jar:guava
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:guava
/opt/flink/lib/flink-table-blink_2.11-1.10.1.jar:guava
/opt/flink/lib/flink-table_2.11-1.10.1.jar:guava
/opt/flink/lib/flink-dist_2.11-1.10.1.jar:guava
/opt/flink/examples/streaming/Twitter.jar:guava
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:guava
5.Package: commons_io-2.4
Severity: HIGH
Fix version: 2.5
Description:
Apache Commons IO contains a flaw that is due to the program failing to
restrict which class can be serialized. This may allow a remote attacker to
execute arbitrary Java code via deserialization methods.
Path:
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:commons-io
/opt/flink/lib/flink-dist_2.11-1.10.1.jar:commons-io
References:
https://issues.apache.org/jira/browse/IO-487
http://commons.apache.org/proper/commons-io/
6.Package: commons_beanutils-1.9.3
Severity: HIGH
Fix version: 1.9.3
Description:
Commons BeanUtils contains a flaw that is due to it failing to restrict the
setting of Class Loader attributes via the class attribute of an ActionForm
object. This may allow a remote attacker to manipulate the loaded Classloader
and execute arbitrary code.
Path:
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:commons-beanut
ils
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2014-0114
https://seclists.org/bugtraq/2014/Apr/177
7.Package: hadoop-2.6.5
Severity: HIGH
Fix version: 3.1.1, 3.0.3, 2.9.2, 2.8.5
Description:
Apache Hadoop contains a flaw that allows traversing outside of a restricted
path. The issue is due to the FileUtil::unpackEntries() functions in
fs/FileUtil.java not properly sanitizing user input, specifically path
traversal style attacks (e.g. '../') supplied via filenames. With a specially
crafted ZIP archive, a context-dependent attacker can write to arbitrary files.
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8009
https://github.com/snyk/zip-slip-vulnerability
8.Package: hadoop-2.6.5
Severity: HIGH
Fix version: 3.1.1, 3.0.3, 2.9.2, 2.8.5
Description:
Apache Hadoop contains a flaw that allows traversing outside of a restricted
path. The issue is due to the FileUtil::unZip() functions in fs/FileUtil.java
not properly sanitizing user input, specifically path traversal style attacks
(e.g. '../') supplied via filenames. With a specially crafted ZIP archive, a
context-dependent attacker can write to arbitrary files.
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8009
https://github.com/snyk/zip-slip-vulnerability
9.Package: hadoop-2.6.5
Severity: HIGH
Fix version: 3.1.1, 2.9.2, 2.8.5
Description:
Apache Hadoop contains an unspecified flaw which may allow a local attacker
with yarn privileges to gain elevated root privileges and execute arbitrary
commands. No further details have been provided.
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-8029
https://seclists.org/oss-sec/2019/q2/132
10.Package: hadoop-2.6.5
Severity: HIGH
Fix version: 2.7.7
Description:
Apache Hadoop contains an unspecified flaw that may allow a local attacker who
is able to escalate their privileges to yarn user to gain root privileges. No
further details have been provided.
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2018-11766
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2016-6811
11.Package: hadoop-2.6.5
Severity: HIGH
Fix version: 2.7.0
Description:
Apache Hadoop contains a flaw in a servlet on the DataNode related to the HDFS
namespace browsing functionality. The issue is triggered as a query parameter
is not properly validated when handling NameNode information. This may allow a
remote attacker to have an unspecified impact.
References:
https://cve.mitre.org/cgi-bin/cvename.cgi?name=2017-3162
https://seclists.org/oss-sec/2017/q2/126
12.Package: hadoop-2.6.5
Severity: HIGH
Description:
In Apache Hadoop 3.1.0 to 3.1.1, 3.0.0-alpha1 to 3.0.3, 2.9.0 to 2.9.1, and
2.0.0-alpha to 2.8.4, the user/group information can be corrupted across
storing in fsimage and reading back from fsimage.
References:
https://nvd.nist.gov/vuln/detail/CVE-2018-11768
Paths:
Hadoop paths
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-co
mmon
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-cl
ient
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-yarn-ap
i
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-mapredu
ce-client-core
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-hdfs
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-common
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-auth
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:hadoop-annotat
ions
13.Package: protobuf- 2.5.0, 2.6.1
Severity: HIGH
Description:
protobuf allows remote authenticated attackers to cause a heap-based buffer
overflow.
Path:
/opt/flink/lib/flink-dist_2.11-1.10.1.jar:protobuf-java
/opt/flink/lib/flink-shaded-hadoop-2-uber-2.6.5-7.0.jar:protobuf-java
References:
https://nvd.nist.gov/vuln/detail/CVE-2015-5237
Please let us know your comments on these issues and fix plans.
Regards,
Suchithra