Various unique "real-world" datasets specifically for deep learning purpose. The files are provided in .msgpack format and can be also used separately from Autonomio for example with Pandas:
import pandas as pd
pd.read_msgpack('https://github.com/autonomio/datasets/raw/master/autonomio-datasets/election_in_twitter')
Dataset consisting of 10 minute samples of 80 million tweets from the beginning of November 2016 to end of December 2016. The keywords used to capture tweets are 'Trump' and 'Hillary'.
Dataset with tweet text classified for sentiment using NLTK Vader including word2vec word vectors for each tweet using spaCy.
4,000 sites with word vectors and 5 categories.
Data from both buy and sell side and over 10 other sources.
9 years of monthly poll and unemployment numbers.
20,000 tweets with various data colums related with tweet quality, including if the tweet is from a bot or not.
The train dataset provided as part of the hugely popular Kaggle Titanic Survitor prediction challenge.
20,000 sites with word vectors based on the landing page content.