Skip to the content.

Final Submission

Majorly I worked on finding anchors, shows, channels probable and confident both or added in the final .txt file. The algorithm segmented shows and created a .txt file for each segmented show containing all the show information that you can see below in the Example of final .txt files section. Implemented the final production pipeline and developed python scripts and slurm script to process and handle all the jobs submissions and output .txt files. I have run and tested my algorithm on HPC with extensive data of 7 months. It took 5 continuous days/night to process 10k+ videos and generate .txt file and .mp4(segmented video).

Path of Output files: /mnt..gallina/Singularity/Show-Segmentation-2021/TvSplit_2

Project Code: Show-Segmentation-2021

Weekly Updated Blogs: Blogs

Here is the final production pipeline flow:

Supervisor@2x

Results & Output

I spent a lot of time running algorithms and getting final .txt files at CaseHPC; this was very time-consuming.

The processing of a single video takes 40 to 50 min on average.

I spent 5 days/night to achieve 7 months output.

The estimation for the whole collection is approx 180 days.

Examples of final .txt files:

Example 1 : 2015-12-05_France-2_2-2_unknown-anchor.txt

TOP|20151205000000|2015-12-05France-2_2-2_unknown-anchor
COL|Communication Studies Archive, UCLA
UID|11dc0427-f4b5-4997-816f-d8a01c729aa1
SRC|Rosenthal Collection, UCLA
TTL|
PID|
CMT|
DUR|22:43:43
VID|640x480
LAN|ENG
LBT|
OVD|20151205000000|2015-12-05_0600_FR_France-2_Télématin_speciale_Telethon_2015
OID|
TMS|01:16:17-00:00:00
INF|probable_host1:Grant-Tinker_Orson-Bean_Joe-E.-Tata_Al-Hunt_Pat-Boone
Confident Anchors|Al-Hunt

Network|CBS_NBC_CNN_PBS_
Probable Shows|
{}
Shows|Charlie_Rose

Example 2 : 2015-12-05_KCET_1-1_Sue_Herera.txt

TOP|20151205000000|2015-12-05KCET_1-1_Sue_Herera
COL|Communication Studies Archive, UCLA
UID|a799196c-b03d-4162-8e4d-006b5be19583
SRC|Rosenthal Collection, UCLA
TTL|
PID|
CMT|
DUR|00:00:00
VID|640x426
LAN|ENG
LBT|
OVD|20151205000000|2015-12-05_0130_US_KCET_Nightly_Business_Report
OID|
TMS|00:00:00-00:00:00
INF|probable_host1:Sue-Herera_Elizabeth-Ashley_Crown-Prince-Dan_Emma-Cleasby_Christina-Strmer
Confident Anchors|Sue-Herera

Network|CBS_NBC_
Probable Shows|
{}
Shows|Nightly_Business_Report_with_Sue_Herera_and_Bill_Griffeth

Example 3 : 2015-12-05_SIC_1-1_Cau_Reymond.txt

TOP|20151205000000|2015-12-05SIC_1-1_Cau_Reymond
COL|Communication Studies Archive, UCLA
UID|f39b2c74-1471-48fe-9e51-ffd0fd3cf1b2
SRC|Rosenthal Collection, UCLA
TTL|
PID|
CMT|
DUR|23:55:45
VID|640x512
LAN|ENG
LBT|
OVD|20151205000000|2015-12-05_2300_BR_SIC_A_Regra_do_Jogo
OID|
TMS|00:04:15-00:00:00
INF|probable_host1:Cau-Reymond_Seth-Tobias_Al-B.-Sure!_David-Guterson_Cha-Seung-won
Confident Anchors|Al-B.-Sure!

Network|ABC_
Probable Shows|
{}
Shows|Black_Men_Revealed

Show analytics

These plots give an overall view of shows like No. of episodes of the show on the x-axis and the count of those episodes on the y-axis.

Suppose: This is the dictionary that contains the name of shows and their occurrences in the whole resultant dataset.

{‘Charlie_Rose’: 6, ‘Worldwide_Exchange’: 2, ‘The_Hot_List’: 1, ‘The_Ricki_Lake_Show’: 1, ‘Access_Daily’: 14, ‘48_Hours’: 9, ‘Extra_with_Billy_Bush’: 16, ‘Good_Day_L.A.’: 8, ‘Newsnight’: 10}

I’m putting the same data in the table, only values, not the name, because we have count on the y-axis that represents how many single or few episodes are present in results or that can be improved in the future.

x-axis y-axis(count)
1 2
2 1
8 1
9 1
10 1
14 1
16 1

So now we can say if the show/episode occurring more than 3 or 4 times like Charlie_Rose, Access_Daily, 48_Hours, Extra_with_Billy_Bush, Good_Day_L.A., Newsnight shows have more accuracy; on the other hand, if the show is present only 1 or 2 times like The_Hot_List, The_Ricki_Lake_Show, Worldwide_Exchange then we can’t say that it is more accurate show or not.

Show analysis of 6 month data

download (4)

Future work

If possible, replace the current celeb detection method with Azure’s Computer Vision service.

Currently the most time consuming process in the program is that of going frame by frame and extracting faces. This can be speed up using multi-threading or any other means possible.

Explore the dataset that contains news shows with anchors and include it in the pipeline.

Shows with single or a few episodes can just be dropped to improve precision at the expense of recall.

Acknowledgments

I want to thank my mentors Frankie Robertson, Francis Steen, and Anna Wilson. Special thanks to Frankie. He was very responsive and helpful all along. We talked through email and meet on zoom, google meet. He gave valuable insights whenever I was stuck or something that I could have done better.

Lastly, I would like to thank Mark Turner and all other Red Hen Lab members for their support.

And most importantly, I express my gratitude to the people behind Google Summer of Code. I had a delightful experience working in both GSoC 2020 and 2021. I hope I keep contributing.