Data preprocessing
We began our analysis with the CMU Movie Summary Corpus as our foundational dataset. To incorporate additional films, such as sequel collections, sequels, book adaptations, comic adaptations, and remakes, we utilized the TMDB dataset. Since the TMDB data lacked direct Wikipedia IDs, we scraped Wikipedia to obtain them for each TMDB movie. Armed with these IDs, we merged the TMDB data into the CMU dataset, linking existing entries and adding new films to expand our analysis. The following graph illustrates the varying sizes of the TMDB dataset, the CMU dataset, and the final filtered dataset after these integrations.
We see that the TMDB dataset is much more exhaustive than our database. We lose quite a lot of data points, but we were still able to recover a lot of movies. However, in the rest of the analysis, we will have to be careful when looking at data before 2010, which will be the data coming from the CMU dataset, and the data after, which comes from TMDB. In absolute value, we should be losing about half of the TMDB dataset if CMU was updated until 2023.
To complement the previous one, the following graph shows how many movies were lost during data cleaning. The scale isn’t too large, but it’s always disappointing to lose some data. Most of the movies lost were from made in the 1980s. From a look at the data, it is mostly niche movies that were lost, but some important movies were missing, so this can explain gaps of movies that you are looking for. The extrapolation of movies that should be lost if CMU was updated until 2023 could be estimated at around a fifth of the TMDB dataset.