New York City (NYC) taxi rides are probably the most commonly used benchmark in the area of data analytics.
It started with Todd W. Schneider deciding to prepare the collection first in 2015 to analyze 1.1 billion NYC Taxi and Uber Trips. Then Mark Litwintschik continued by testing lots of databases and search engines using the data collection.
Now we at DB Benchmarks have dockerized preparation of the data collection to make it easier to use and made it available as a part of the most transparent and open source database benchmarks suite.
The data collection constitutes 1.7 billion taxi and for-hire vehicle (Uber, Lyft, etc.) trips originating in New York City since 2009. Most of the raw data comes from the NYC Taxi & Limousine Commission.
The data collection record includes a lot of different attributes of a taxi ride:
It can be used mostly for testing analytical queries, but it also includes a couple of full-text fields that can be used to test free text capabilities of databases.
The whole list of fields and their data types is:
"properties": {
"vendor_id": {"type": "keyword"},
"pickup_datetime": {"type": "date", "format": "epoch_second"},
"dropoff_datetime": {"type": "date", "format": "epoch_second"},
"store_and_fwd_flag": {"type": "keyword"},
"rate_code_id": {"type": "integer"},
"pickup_longitude": {"type": "float"},
"pickup_latitude": {"type": "float"},
"dropoff_longitude": {"type": "float"},
"dropoff_latitude": {"type": "float"},
"passenger_count": {"type": "integer"},
"trip_distance": {"type": "float"},
"fare_amount": {"type": "float"},
"extra": {"type": "float"},
"mta_tax": {"type": "float"},
"tip_amount": {"type": "float"},
"tolls_amount": {"type": "float"},
"ehail_fee": {"type": "float"},
"improvement_surcharge": {"type": "float"},
"total_amount": {"type": "float"},
"payment_type": {"type": "keyword"},
"trip_type": {"type": "byte"},
"pickup": {"type": "keyword"},
"dropoff": {"type": "keyword"},
"cab_type": {"type": "keyword"},
"rain": {"type": "float"},
"snow_depth": {"type": "float"},
"snowfall": {"type": "float"},
"max_temp": {"type": "byte"},
"min_temp": {"type": "byte"},
"wind": {"type": "float"},
"pickup_nyct2010_gid": {"type": "integer"},
"pickup_ctlabel": {"type": "keyword"},
"pickup_borocode": {"type": "byte"},
"pickup_boroname": {"type": "keyword"},
"pickup_ct2010": {"type": "keyword"},
"pickup_boroct2010": {"type": "keyword"},
"pickup_cdeligibil": {"type": "keyword"},
"pickup_ntacode": {"type": "keyword"},
"pickup_ntaname": {"type": "text", "fields": {"raw": {"type":"keyword"}}},
"pickup_puma": {"type": "keyword"},
"dropoff_nyct2010_gid": {"type": "integer"},
"dropoff_ctlabel": {"type": "keyword"},
"dropoff_borocode": {"type": "byte"},
"dropoff_boroname": {"type": "keyword"},
"dropoff_ct2010": {"type": "keyword"},
"dropoff_boroct2010": {"type": "keyword"},
"dropoff_cdeligibil": {"type": "keyword"},
"dropoff_ntacode": {"type": "keyword"},
"dropoff_ntaname": {"type": "text", "fields": {"raw": {"type":"keyword"}}},
"dropoff_puma": {"type": "keyword"}
}
So far we have made this test available for 3 databases
In this test we make as little changes to database default settings as possible to not give either of them an unfair advantage. Testing at max tuning is no less important, but it's a subject for another benchmark. Here we want to understand what latency a regular non-experienced user can get after just installing a database and running it with its default settings. But to make it fair to compare one with another we still had to change a few settings:
Clickhouse: no tuning, just CREATE TABLE ... ENGINE = MergeTree() ORDER BY id
and standard clickhouse-server docker image.
Elasticsearch: here to make it fair to compare with the other databases we had to help Elasticsearch by:
letting it make 32 shards: ("number_of_shards": 32
), otherwise it couldn’t utilize the CPU which has 32 cores on the server, since as said in Elasticsearch official guide “Each shard runs the search on a single CPU thread”.
bootstrap.memory_lock=true
since as said on https://www.elastic.co/guide/en/elasticsearch/reference/current/docker.html#_disable_swapping it needs to be done for performance.
the docker image is standard
Manticore Search is also used in a form of their own docker image + the columnar library they provide:
as well as with Elasticsearch we also use 32 shards in a form of 32 plain indexes
and we use Manticore columnar storage since comparing Manticore’s default row-wise storage vs Clickhouse’s and Elasticsearch’s columnar storages would be not fair on such a large data collection.
We've also configured the databases to not use any internal caches. Why this is important:
👌 those that just cache raw data stored on disk. For example many databases use mmap()
to map the data stored on disk to memory, access it easily and let the operating system take care about the rest (reading it from disk when there's free memory, removing it from memory when it's needed for something more important etc). This is ok in terms of performance testing, because we let each database leverage the benefit of using the OS page cache (or its internal similar cache that just reads data from disk)
That's exactly what we do in this benchmark.
❗ those that are used to save results of previous calculations. And it's fine in many cases, but in terms of this benchmark letting database enable such a cache is a bad idea, because:
So we do everything to make sure none of the database does this kind of caching.
What exactly we do to achieve that:
SYSTEM DROP MARK CACHE
, SYSTEM DROP UNCOMPRESSED CACHE
, SYSTEM DROP COMPILED EXPRESSION CACHE
before testing each new query (not each attempt of the same query).
"index.queries.cache.enabled": false
in its configuration
/_cache/clear?request=true&query=true&fielddata=true
before testing each new query (not each attempt of the same query).
qcache_max_bytes = 0
docstore_cache_size = 0
echo 3 > /proc/sys/vm/drop_caches; sync
before each NEW query (NOT each attempt). I.e. for each new query we:
The queries are mostly analytical queries that do filtering, sorting and grouping. We’ve also included one full-text query:
[
"SELECT count(*) FROM taxi where pickup_ntaname = '0'",
"SELECT pickup_ntaname, count(*) c FROM taxi GROUP BY pickup_ntaname ORDER BY c desc limit 20",
"SELECT cab_type, count(*) c FROM taxi GROUP BY cab_type order by c desc LIMIT 20",
"SELECT passenger_count, avg(total_amount) a FROM taxi GROUP BY passenger_count order by a desc LIMIT 20",
"SELECT count(*) FROM taxi WHERE tip_amount > 1.5",
"SELECT avg(tip_amount) FROM taxi WHERE tip_amount > 1.5 AND tip_amount < 5",
"SELECT rain, avg(trip_distance) a FROM taxi GROUP BY rain order by a desc LIMIT 20",
{
"manticoresearch": "SELECT * FROM taxi where match('harlem east') LIMIT 20",
"clickhouse": "SELECT * FROM taxi where match(dropoff_ntaname, '(?i)\\WHarlem\\WEast\\W') or match(pickup_ntaname, '(?i)\\WHarlem\\WEast\\W') LIMIT 20",
"elasticsearch": "SELECT * FROM taxi where query('harlem east') LIMIT 20"
},
"SELECT avg(total_amount) FROM taxi WHERE trip_distance = 5",
"SELECT avg(total_amount), count(*) FROM taxi WHERE trip_distance > 0 AND trip_distance < 5",
"SELECT count(*) FROM taxi where pickup_ntaname != '0'",
"select passenger_count, count(*) c from taxi group by passenger_count order by c desc limit 20",
"select rain, count(*) c from taxi group by rain order by c desc limit 20",
"SELECT count(*) from taxi where pickup_ntaname='Upper West Side'",
"SELECT * from taxi limit 5",
"SELECT count(*) FROM taxi WHERE tip_amount = 5",
"SELECT avg(total_amount) FROM taxi"
]
You can find all the results on the results page by selecting “Test: taxi”.
Remember that the only high-quality metric is “Fast avg” since it guarantees a low coefficient of variation and a high query count conducted for each query. The other 2 (“Fastest” and “Slowest”) are provided with no guarantee since:
Slowest - is a single attempt result, in most cases the very first coldest query. Even though we purge OS cache before each cold query it can’t be considered stable. So it can be used for informational purposes only (even though many benchmark authors publish such results without any disclaimer).
Fastest - just the very fastest result, it should be in most cases similar to the “Fast avg” metric, but can be more volatile from run to run.
Remember the tests including the results are 100% transparent as well as everything in this project, so:
you can use the test framework to learn how they were made
and find raw test results in the results directory.
Unlike other less transparent and less objective benchmarks we are not making any conclusions, we are just leaving screenshots of the results here:
The author of this test and the test framework is a member of Manticore Search core team and the test was initially made to compare Manticore Search with Elasticsearch, but as shown above and can be verified in the open source code and by running the same test yourself Manticore Search wasn’t given any unfair advantage, so the test can be considered unprejudiced. However, if something is missing or wrong (i.e. non-objective) in the test feel free to make a pull request or an issue on Github. Your take is appreciated! Thank you for spending your time reading this!
Also published here.