Tom Hirshberg

When AI Meets Fashion

Data Scientist – Microsoft

Tom Hirshberg

When AI Meets Fashion

Data Scientist – Microsoft

Bio

Tom is a data scientist at Microsoft in the Azure Video Indexer group. She holds BSc and MSc in Computer Science from the Technion at Israel Institute of Technology. Previously, she was a research intern at Microsoft in Redmond, where she worked on new optimization methods in the field of autonomous drones. During she BSc she was part of the algorithm team working on the first autonomous formula race car at  the Technion, and her MSc thesis focused on acoustic-based indoor localization of drones.

Bio

Tom is a data scientist at Microsoft in the Azure Video Indexer group. She holds BSc and MSc in Computer Science from the Technion at Israel Institute of Technology. Previously, she was a research intern at Microsoft in Redmond, where she worked on new optimization methods in the field of autonomous drones. During her BSc she was part of the algorithm team working on the first autonomous formula race car at the Technion, and her MSc thesis focused on acoustic-based indoor localization of drones.

Abstract

Have you ever watched a video and seen an amazing outfit worn by a character, wanting to buy the same for yourself? Recommendations of items based on what we see is a well-known field. But what if these recommendations could be made dynamically based on the events happening in the video, recommending items of featured clothing while still maintaining the privacy of the user? This is our brand new featured clothing model that is part of Azure Video Indexer, a solution that generates insights from videos to help understand and search videos in media content and archives.

In this talk, I will explain how we can detect the main clothing items that are relevant to the viewer based on the changing content of a video, without the need of any personal data. Our model relies on multiple AI systems using both audio and visual domains, such as people detection, key moment detection, celebrity identification, and many more.

Abstract

Have you ever watched a video and seen an amazing outfit worn by a character, wanting to buy the same for yourself? Recommendations of items based on what we see is a well-known field. But what if these recommendations could be made dynamically based on the events happening in the video, recommending items of featured clothing while still maintaining the privacy of the user? This is our brand new featured clothing model that is part of Azure Video Indexer, a solution that generates insights from videos to help understand and search videos in media content and archives.
In this talk, I will explain how we can detect the main clothing items that are relevant to the viewer based on the changing content of a video, without the need of any personal data. Our model relies on multiple AI systems using both audio and visual domains, such as people detection, key moment detection, celebrity identification, and many more.

Planned Agenda

8:45 Reception
9:30 Opening words by WiDS TLV ambassador Nitzan Gado and by Lily Ben Ami, CEO of the Michal Sela Forum
9:50 Prof. Bracha Shapira – Data Challenges in Recommender Systems Research: Insights from Bundle Recommendation
10:20 Juan Liu – Accounting Automation: Making Accounting Easier So That People Can Forget About It
10:50 Break
11:00 Lightning talks
12:20 Lunch & poster session
13:20 Roundtable session & poster session
14:05 Roundtable closure
14:20 Break
14:30 Merav Mofaz – “Every Breath You Take and Every Move You Make…I'll Be Watching You:” The Sensitive Side of Smartwatches
14:50 Reut Yaniv – Ad Serving in the Online Geo Space Along Routes
15:10 Rachel Wities - It’s Not Just the Doctor’s Handwriting: Challenges and Opportunities in Healthcare NLP
15:30 Closing remarks
15:40 End