Kelley Blue Book
NOTE: This had its own repository and was moved here to make my projects a little bit cleaner.
Goal: Predict car price based on the car's specs.
BeautifulSoupwas used to scrape Kelley Blue Book for vehicle data within a 75 mile radius of the area that I live in. Information that was scraped includes vehicle title, price, mileage, whether the vehicle was in an accident, whether it had a previous owner, interior and exterior color, etc.
Cleaning: The data was cleaned as best as I could at the time and at first I thought that I would be able to do more with it than what I actually did. This is sometihng that I could come back to and finish the analysis a litte bit more than what I did.
2019 Movie Scraping from Wikipedia
BeautifulSoupwas used to scrape United States films made in 2019 from Wikipedia. All of the data located in the box under the movie title on Wikipeia was scraped.
At the end of the file I attempted to perform a linear regression with
pytorchwhere the budget was being used to predict the box office perforance. The R2 score was 64%, however the loss was extremely high. Because of this, I tried to turn the problem into a classification problem where I performed a logistic regression.
youtube-transcript-apiwere used to get YouTube video information. The idea was to compare liberal vs conservative YouTube channels' video titles and transcripts.
This is another project that would need to be picked back up later as it was left like this because the time it was taking to scrape some of these channels was taking an extremely long amount of time. For instance, TheYoungTurks had around 40,000 videos to scrape data from.
Goal: Predict community rating of the distribution based on the operating system's specs.
pandaswas used to scrape the top 275 Linux distributions on Distrowatch. Each operating system has a table containing its specs (e.g., distribution the operating system is based on, its architecture, type of desktop, etc.) and the community rating (popularity).
Cleaning: There is a csv containing a generic overview of each operating system, as well as a csv that contains a lot more information about every release of every distribution. The last csv that may be useful was cleaned and put in a one-hot-encoded format.
Analysis: A logistic algorithm coded from scratch was used to analyze the data. The model did better than chance, however; the data that was used may not have been the best to perform such a task--or, perhaps there is no correlation between the operating system's specs and its' community rating.
Trulia House Prices
Goal: Predict house price based on the house's attributes.
BeautifulSoupwas used to scrape 2500 house listings on Trulia across five cities in Missouri (Columbia, Kansas City, St. Louis, Springfield, and Lee's Summit). Data on each listing include things such as crime rate, schools in the area, listing history, tax history, etc.
Cleaning: The csv that is the result of the cleaning was formatted in such a way that the categorical variables are in a one-hot-encoded format (except for the region) and numeric variables were left alone.
Analysis: The explanation of the variables can be found in the file linked to 'analysis' at the beginning of this block. House price, price per square foot, total square footage, and lot size are just one example of what was compared across the five cities in Missouri.
Model: A linear regression algorithm that was coded from scratch was originally used but the results were all
sklearnwas used instead. The R2 score from the model was 0.72, however; the MSE was extremely high. Perhaps a larger selection of the data (or better cleaning of the results) needs to be done to work better with the linear regression models.
BeautifulSoupis being used to scrape posts on slashdot to eventually perform a sentiment analysis on the posts' titles, and the comments on the posts.