1 / 64

Machine Learning and Open Courseware

Machine Learning and Open Courseware. Colin de la Higuera cdlh@univ-nantes.fr. Outline. What is open courseware ? Is there still research to be done? Recommendation Translation and Transcription Annotation Navigation Traces. Opencourseware =Open Educational Resources.

sovann
Download Presentation

Machine Learning and Open Courseware

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Machine Learning and Open Courseware Colin de la Higueracdlh@univ-nantes.fr

  2. Outline What is open courseware? Is there still research to be done? Recommendation Translation and Transcription Annotation Navigation Traces

  3. Opencourseware=Open Educational Resources • Organization: OCWC • 30 000+ Modules • 280+ Organizations • 40 Countries • 29 Languages

  4. Somereasons (1) By 2025, the global demand for higher education will double to ~200M per year, mostly from emerging economies (NAFSA 2010) 100 M/15 years… 80 000 new students/week! Diana Laurillard. http://www.ucd.ie/teaching/u21conference2013/ AnkaMulder http://ankamulder.weblog.tudelft.nl/

  5. Some reasons (2) 220 M French Speakers in 2010… 700 M in 2050 3% (2012)  more than 7% in 2050 By 2030, more French Speakers than English Speakers (only 5% in 2050) René Marcoux (Université Laval et coordonnateur du RéseauDémographie de l’Agenceuniversitaire de la francophonie)

  6. How does this relate to MOOCs? Historically, appeared before MOOCs Most MOOC players acknowledge inheritance Also, through MOOCs only the upper layer of what we teach is being shared Question is: can we do something with the courseware that has been produced so far? Some believe we can do better

  7. OpenCourseware and Pascal The Pascal network ran 2003-2012 One key asset was videolectures.net PASCAL 2  Knowledge 4 All foundation Goal is to build upon the Pascal legacy

  8. VideoLectures current stats

  9. http://conference.ocwconsortium.org/2014/

  10. Are there really some research questions there?

  11. Research is e-learning? In part. Lots of new scenarios Much more learning data which needs analysing But not only

  12. A simplified picture • The learner and his experience (e-learning) • Knowledge as a merchandise • Sharing open educational resources • Helping the teacher deal with the masses of information (machine learning)

  13. Clue: Think about MOOCs Coursera: Andrew Ng and DaphneKoller Eduacity, Keywords: automaticevaluation, learninganalytics, big data

  14. Clue: Some titles of talks from the Internet of Education conference, Slovenia, November 2013http://www.k4all.org/Internet_of_Education/ TransLectures: cost-effective transcription and translation of video lectures Knowledge Building Through Collaborative Hypervideo Creation Annotations, a key asset for video-based e-learning A MediaMixer for online learning? – making learning materials more valuable for their owner and more useful for their consumer

  15. Clue: OCWC 2014 Global Stats from the accepted papers 4 Tracks

  16. Slidesbased on those by MatjazRihtar <matjaz.rihtar@ijs.si> Machine Learning and Recommendation

  17. Why recommendation? • Simplest question is: after this video, which one should I consider? • Basic system uses history of transactions: most viewers who watched A also watched B • Can we use? • Individual history • Thematic similarity • History over what the users really do

  18. Machine learning approach Obviouslya slide improper for SML Extract many different features Let the learning algorithm put different weights on these depending on their usefulness Think carefully about the validation issues in order to verify what you have learnt

  19. Experience from LaVie • Were used for recommendation • Transcriptions • Graphs of authors • Related pdfs • Other ontologies (Wikipedia, DBLP and Google)

  20. Topic and user modeling • 7 features: Lc = current lecture • Lecture popularity Lp = proposed lecture • Number of visits U = user • Content similarity 1 table has approx. 70 million entries • BoW(Lc) ·BoW(Lp) (for features 2,3,6 and 7) • Category similarity • BoC(Lc) ·BoC(Lp) • User content similarity (computed on the fly) • BoW(Hist(U)) ·BoW(Lp) • User category similarity (computed on the fly) • BoC(Hist(U)) ·BoC(Lp) • Co-visits • Number of times of Lc and Lp viewed in the same browsing session • User similarity • Number of users who have watched both Lc and Lp

  21. Recommendation (1) • Using SVM classifier for training: • Positive samples: two months of clicks using current recommender • Resulting feature weights:

  22. Recommendation (2) • Final recommendation • A linear SVM classifier was used to rank all possible recommendation links: • Lectures with top 10 scores are recommended

  23. Evaluation • Evaluation • Using coin flipping between old and new recommender • Counting the number of clicks • Try http://dev.videolectures.net/ • Our recommendations have /?ref=r00: in links

  24. Transcription and translation

  25. Slides by GonçalGarcésggarces@dsic.upv.esProject coordinator: AlfonsJuan-Ciscarajuan@dsic.upv.es

  26. The transLectures partners

  27. The transLectures approach • Automatic Speech Recognition (ASR) and Machine Translation (MT) • Adaptation: Taking advantage of the characteristics of video lecture repositories • High-quality automatic transcriptions and translations • Interactive postediting: intelligent interaction for reduced effort

  28. Goals • Massive adaptation • Intelligent interaction • Implementation • Case studies: Videolectures.NET & Polimedia • Real-life evaluation • Integration into Opencast Matterhorn http://opencast.org/matterhorn/

  29. Languages • Transcription (ASR) • EN • SL • ES • Translation (MT) • EN>SL , SL>EN • EN>ES , ES>EN • EN>FR • EN>DE

  30. transLectures: video demo

  31. Massive adaptation Characteristics of video lectures

  32. Scientific evaluations Worse • Transcription results • WER: Word Error Rate (%) • Goal: WER < 25% • EN, SL, ES Better

  33. Scientific evaluations Better • Translation results • BLEU • Goal: BLEU > 30 • EN>SL , SL>EN • EN>ES , ES>EN • EN>FR • EN>DE Worse

  34. Y1 results and comparison

  35. Y1 results and comparison

  36. Intelligent interaction • Post-editing automatic transcriptions/translations • The user invests the least possible effort • The system learns the most from it • Confidence measures • Fast constrained search

  37. Intelligent interaction

  38. User evaluations Intelligent interaction

  39. User evaluations User evaluations at UPV: results

  40. transLectures: Open source tools • The tLplayer (& editor) • Coming soon (www.translectures.eu) • The transLectures-UPV Toolkit (TLK) for ASR • www.translectures.eu/tlk • RWTH Aachen: rASR, Jane (MT) • http://www-i6.informatik.rwth-aachen.de/web/Software/

  41. Multilingualism • Multilingualism is a crucial issue • If English is the unique language we have many deprived users • But also we are missing a lot of excellent material • It is a highly sensitive issue • Answers • Being able to navigate between languages • Being able to translate: translectures project

  42. Olivier Aubert - @Olivier_Aubert Yannick Prié - @yprie Annotation issues

  43. Key questions Obtaining the participation of the learner can be vital (at least if the main purpose of the resource is to allow a learner to learn) Examining the MOOC’s success helps Why would someone want to annotate our videos? What tools can we provide in order for quality annotation to take place ?

  44. Video-based e-learning activities • Different activities based on • the nature of the video document • the status of the annotator • the status of the recipient

  45. Assimilation - example Advene: an example of an offline anotator

  46. Collaborative assimilation example VideoNot.es

  47. Feedback - example PolemicTweet

  48. App. of an analysis grid - example Matterhorn Engage player

  49. Analysis - example MediaThread

  50. Reflexivity/Feedback - example Visu

More Related