This is a fun project where I applied Gatys' artistic neural style transfer algorithm with silhouettes and doodles to generate novel patterns and intricate art. The silhouettes facilitated the algorithm to give a neat finish to the style. Documentation and code are provided here.
This is a synthetic dataset for ridesharing that shows the details of driver sign-up data. The data consists of when the drivers signed up, the city, sign-up OS, sign-up channel (paid, referred, etc.). Once they signed up for the driving, they are requested to be available for background check, add their vehicle and once those stages are cleared, they are then allowed to start picking up passengers.
Here, the company wants to maximize the trips taken by the drivers. There may be several hurdles before first trip is taken, such as, the background check may be negative, or the drivers may fail to add their vehicle, and so on. This data analysis will try to find the important factors that is involved in the first trip taken by the drivers. And consequently, some recommendations are suggested based on the analysis.
Click here for the code and dataset.
I developed an agent-based modeling that represents a synthetic population in a neighborhood. A micro network is created for each household based on the locations people typically travel every day (work, day care center, grocery stores, etc.). People interact with the infrastructure in the neighborhood. Presence of electric chargers at home, work or public as well as hydrogen stations influence their purchase behavior.
The model predicts the purchase probability based on a multinomial logit approach in real-time based on the daily driving pattern of the neighborhood. The model runs for a year.
Click here for more details. The model file is also available to download from the Github link.
I created this Shiny App in 2014 based on my research on consumer behavior and infrastructure analysis. The backend of this model is based on a multinomial logistic regression algorithm, which takes in consumer characteristics, vehicle attributes, and predicts the probability of the consumer purchasing a vehicle technology. The App gives the top 3 optimal choices for the consumer.
A spatial tool was developed using ArcGIS and ArcPy framework to identify optimal locations of renewable energy power plants.
Click here to learn more.
A bi-objective algorithm that optimizes time and money is developed for the modal choices in the San Francisco Bay Area region. Here, the users choose between driving and BART during the peak and non-peak hours. The model also estimates the cost and emissions of their choices.
Click here to learn more.
This was done as part of the Coursera Data Science Specialization project. Based on the corpus of text data given, n-grams were created. The tool predicts the next word in real-time based on the word chains given in the input. Model details are given under the 'Model' tab in the app.
Click here to play with the next word prediction tool.
Kaggle introduced Draper satellite image chronology challenge, where five sets of images were given representing each day. The challenge was to arrange the images in the test data which were given in random order.
I did a simple exploratory analysis on the a set of training data images, using basic computer vision techniques (Sobel edge detection, RGB bands, Hough transform and SURF point detections).
The Github link to the IPython code is here.
This was developed as part of Udacity's Artificial Intelligence nanodegree program. I implemented the 'naked twin' strategy and diagonal sudoku solver for a given sudoku grid.
Here is the link to the Github repository.
This was developed as part of the Udacity's Artificial Intelligence nanodegree program. Here, I implement Minimax algorithm with and without alpha-beta pruning to identify the next move for the computer player.
Click here to see more details on the code.
A "tournament" was given as the part of the exercise to test the computer player with different types of adversaries. I developed three kinds of heuristics for the computer player. See here for the analysis of them and how they can differ to predict the end-game.
This is a simple random forest classification done in R, to identify the conversion of a website.