Hello,
we are implementing a text mining algorithm on Flink (0.9.0) and encountered a
mysterious IOException (see below for detailed stacktrace). Flink seems to have
problems finding some temporary file even though there is enough free space on
the disk. Do you have an idea why the file cannot
This is your code (it applied the "print" before the aggregation is done)
> ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
> DataSet orders=(DataSet)
> env.readCsvFile("/home/hadoop/Desktop/Dataset/orders.csv")
> .fieldDelimiter('|')
> .includeFields(ma
I did not understand what you mean
--
View this message in context:
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/why-when-use-orders-aggregate-Aggregations-MAX-2-not-return-one-value-but-return-more-value-tp1977p1989.html
Sent from the Apache Flink User Mailing List archi
Thanks Flavio.
Here's my suggestion. If your paths are short and don't vary much, it might
make sense to do this using the DataSet API. The iteration will probably be
too much overhead.
For example, if you want to find the paths for "livesIn.containedIn.name",
then you can create 3 DataSets, one f
In my Implementation every vertex has its own bags of "knowledge" (i.e. it
knows all tuples belonging to it..).
So in Vertex 1 I have a field (an HashMap) containing the following info:
- type=Person
- livesIn=2 (and I know also that 2 is a vertexId)
In Vertex 3 I know:
- type=Place
Hi Flavio,
yes, I think it's possible. I have one question before I try to explain
how:
do you model "Rome", "Berlin", "101.3" etc. in your above example as
separate vertices or as properties of vertices?
On 8 July 2015 at 10:43, Flavio Pompermaier wrote:
> Let's say I have some nodes of some t
Let's say I have some nodes of some type of interest (type1 and type2).
My source data looks like .
For example, I could be interested in sourceNodes having type == Person
to gather the value obtained from the expansion of some paths (for
example livesIn.name and marriedTo.name). Notice that I coul
Hi,
aggregate performs an in-place aggregation, i.e., the input and output type of
aggregate() is identical, but the aggregated fields are updated.
Causion: All fields that are not associated with an aggregation function and
are not not a grouping field have non-deterministic values. That mea
Is it the same message that you propagate or is it different for each
vertex / neighbor? If you have to store a pair for each
neighbor, then you will quickly run out of memory. If it's the same message
you need to send, or you simply need to add the current vertex Id, then you
can probably get rid
hi, you are not printing the aggregation but the input
val result = orders.aggregate(Aggregations.MAX, 2)
result.print
cheers
michele
> Il giorno 08/lug/2015, alle ore 02:00, hagersaleh
> ha scritto:
>
> why when use orders.aggregate(Aggregations.MAX, 2) not return one value but
> retu
The problem is that my nodes have to gather data coming from some path of
interest along the graph (depending on the type of the node), otherwise
they just have to forward the received message adding their id to the
message path (more or less). It's like a postal tracking system.
The problem is th
Hi Flavio!
Are you talking about vertex-centric iterations in gelly?
If yes, you can send messages to a particular vertex with
"sendMessageTo(vertexId, msg)" and
to all neighbors with "sendMessageToAllNeighbors(msg)". These methods are
available inside the MessagingFunction.
Accessing received me
12 matches
Mail list logo