This was my first visit to IMVC which just celebrated 10 years of activity. I found it to be a very interesting combination of academia and industry work which provides a great platform for disseminating knowledge and fostering future collaborations. Personally, I presented a poster on our recent point cloud classification method.
I would like to thank the committee, IMVC organizers and Nexar for the graduate student award.
The day started with an interesting talk by Prof. Shmuel Peleg on the benefits of combining sight and sounds. He showed a great example of how our mind can fool us to think we are hearing a different sound when a video is manipulated or swapped. He then presented their method for combining video and sound modalities for the application of sound separation and sound enhancement.
Next, Dr. Aya Soffer from IBM talked about moving from perception to comprehension and how they use video and additional modalities for scene detection and semantics. She talked about their experience with applying image models on videos, then moved to detail some “few shots” learning methods (RepMet, LaSO, delta encoder). Finally, she presented their “Moments in Time” dataset.
At this point, they announced the graduate and undergraduate award winners (me!!!). I had the privilege to give a 5-minute “pitch” talk of our point cloud classification work. It was a lot of fun to see how many people found it interesting and relevant for their own work.
Next, Prof. Naftali Tishby talked about the information theory of deep learning. He was followed by Dr. Laurence Keselbrener from Medtronics who talked about AI for capsule endoscopy, specifically for colorectal cancer using PillCam. They use AI for several tasks, including reducing image reading times, image enhancement, detection, and localization. As a person who suffers from Crohn’s disease (Type of chronic inflammatory bowel disease), I think this work will be life changing for many people.
Dr. Yael Pritch Knaan talked about Computational photography on googles smartphone. She presented the tech behind one of my favorite features in Android devices the “Portrait mode“. Here, they achieved shallow depth of field with a single camera using dual pixel data. She also shortly presented “Night Sight” (read more in Google AI blog post), a cool new feature on pixel devices that produces great images in low lighting conditions.
To conclude the session Dr. Andrey Boisman talked about the mission of the ministry of science and technology.
At this point, there was a coffee break after which parallel sessions started. I will detail below a few talks that I attended and particularly liked, though because of my poster presentation I missed a lot of them.
There was also a great exhibition floor that I did not get to fully explore. I only had a chance to visit the booth of “Coral Detection Systems” which was co-founded by an old colleague of mine (Tammar Avraham). They created an amazing product for drowning detection called MANTA.
There was a very mind-opening panel discussion on the future, ethics and social aspects of AI (List of speakers in the image below).
Ilan Kadar from Nexar talked about continuous deep learning at the edge and presented their MUNET – simultaneous multi-tasks for driving on a mobile device. He discussed some corner cases (when something highly improbable happens) – continuous learning pipeline. He also presented their BDD100K dataset for road understanding research.
Elad Richardson from the Israeli defense community presented their work on text detection which finds a common scale for all text elements in an image and mitigates the advantages of multi-scale approach (accurate but slow) and single scale approach (fast but not very accurate)
Finally, Prof. Gal Chechik presented several works on extracting meaningful concepts without explicit supervision.