It would help to understand the current issues that you have with this approach? I used a similar approach (not with Flink, but a similar big data technology) some years ago
> Am 04.03.2019 um 11:32 schrieb Wouter Zorgdrager <w.d.zorgdra...@tudelft.nl>: > > Hi all, > > I'm working on a setup to use Apache Flink in an assignment for a Big Data > (bachelor) university course and I'm interested in your view on this. To > sketch the situation: > - > 200 students follow this course > - students have to write some (simple) Flink applications using the > DataStream API; the focus is on writing the transformation code > - students need to write Scala code > - we provide a dataset and a template (Scala class) with function signatures > and detailed description per application. > e.g.: def assignment_one(input: DataStream[Event]): DataStream[(String, Int)] > = ??? > - we provide some setup code like parsing of data and setting up the > streaming environment > - assignments need to be auto-graded, based on correct results > > In last years course edition we approached this by a custom Docker container. > This container first compiled the students code, run all the Flink > applications against a different dataset and then verified the output against > our solutions. This was turned into a grade and reported back to the student. > Although this was a working approach, I think we can do better. > > I'm wondering if any of you have experience with using Apache Flink in a > university course (or have seen this somewhere) as well as assessing Flink > code. > > Thanks a lot! > > Kind regards, > Wouter Zorgdrager