Stay organized with collections
Save and categorize content based on your preferences.
Data Dumps are a downloadable version of the data in Freebase. They constitute a snapshot of the data stored in Freebase and the Schema that structures it, and are provided under the same CC-BY license. The Freebase/Wikidata mappings are provided under the CC0 license.
If you're writing your own code to parse the RDF dumps its often more efficient to read directly from GZip file rather than extracting the data first and then processing the uncompressed data.
<subject> <predicate> <object> .
Note: In Freebase, objects have MIDs that look like /m/012rkqx. In RDF those MIDs become m.012rkqx. Likewise, Freebase schema like /common/topic are written as common.topic.
The subject is the ID of a Freebase object. It can be a Freebase MID (ex. m.012rkqx) for topics and CVTs or a human-readable ID (ex. common.topic) for schema.
The predicate is always a human-readable ID for a Freebase property or a property from a standard RDF vocabulary like RDFS. Freebase foreign key namespaces are also used as predicates to make it easier to look up keys by namespace.
The object field may contain a Freebase MID for an object or a human-readable ID for schema from Freebase or other RDF vocabularies. It may also include literal values like strings, booleans and numeric values.
Topic descriptions often contain newlines. In order to make each triple fit on one line, we have escaped newlines with "\n".
Freebase Deleted Triples
We also provide a dump of triples that have been deleted from Freebase over time. This is a one-time dump through March 2013. In the future, we might consider providing periodic updates of recently deleted triples, but at the moment we have no specific timeframe for doing so, and are only providing this one-time dump.
The dump is distributed as a .tar.gz file (2.1Gb compressed, 7.7Gb uncompressed). It contains 63,036,271 deleted triples in 20 files (there is no particular meaning to the individual files, it is just easier to manipulate several smaller files than one huge file).
Thanks to Chun How Tan and John Giannandrea for making this data release possible.
The data format is essentially CSV with one important caveat. The object field may contain any characters, including commas (as well as any other reasonable delimiters you could think of). However, all the other fields are guaranteed not to contain commas, so the data can still be parsed unambiguously.
The columns in the dataset are defined as:
creation_timestamp (Unix epoch time in milliseconds)
creator
deletion_timestamp (Unix epoch time in milliseconds)
deletor
subject (MID)
predicate (MID)
object (MID/Literal)
language_code
CSV
1352854086000,/user/mwcl_wikipedia_en,1352855856000,/user/mwcl_wikipedia_en,/m/03r90,/type/object/key,/wikipedia/en/$B816,en
1355171076000,/user/mwcl_musicbrainz,1364258198000,/user/turtlewax_bot,/m/0nncp9z,/music/recording/artist,/m/01vbfm4,en
1176630380000,/user/mwcl_images,1335928144000,/user/gardening_bot,/m/029w57m,/common/image/size,/m/0kly56,en
1292854917000,/user/mwcl_musicbrainz,1364823418001,/user/mbz_pipeline_merge_bot,/m/0fv1vl8,/type/object/type,/common/topic,en
1205530905000,/user/mwcl_images,1336022041000,/user/gardening_bot,/m/01x5scz,/common/licensed_object/license,/m/02x6b,en
1302391361000,/user/content_administrator,1336190973000,/user/gardening_bot,/m/0gkb45y,/type/object/type,/type/content,en
1176728962002,/user/mwcl_images,1335954186000,/user/gardening_bot,/m/08430h,/common/topic/image,/m/02cs147,en
1172002568007,/user/mwcl_chefmoz,1283588560000,/user/delete_bot,/m/01z4c1z,/type/object/name,La Casa Rosa Mexican Restaurant,en
Freebase/Wikidata Mappings
The data has been created based on the Wikidata-Dump of October 28, 2013, and contains only those links that have at least two common Wikipedia-Links and not a single disagreeing Wikipedia-Link. Furthermore, the lines are sorted by the number of common Wikipedia-Links (although in Turtle this does not really matter).
Freebase Data Dumps are provided free of charge for any purpose with regular updates by Google. They are distributed, like Freebase itself, under the Creative Commons Attribution (aka CC-BY) and use is subject to the Terms of Service.
The Freebase/Wikidata ID mappings are provided under CC0 and can be used without restrictions.
Citing
If you'd like to cite these data dumps in a publication, you may use:
Google, Freebase Data Dumps, https://developers.google.com/freebase/data, <month> <day>, <year>
Or as BibTeX:
BibTex
@misc{freebase:datadumps,title="Freebase Data Dumps"author="Google",howpublished="\url{https://developers.google.com/freebase/data}",edition="<month> <day>, <year>",year="<year>"}
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2023-11-20 UTC."],[[["\u003cp\u003eThe Freebase API has been shut down, but this page provides access to the last available data dump for research purposes.\u003c/p\u003e\n"],["\u003cp\u003eThe data dumps, including Freebase Triples, Deleted Triples, and Wikidata Mappings, are substantially out of date but are available for download.\u003c/p\u003e\n"],["\u003cp\u003eThe datasets are provided under the CC-BY license, except for the Freebase/Wikidata mappings, which are provided under the CC0 license.\u003c/p\u003e\n"],["\u003cp\u003eThese datasets are not maintained, and Google provides no support regarding their contents or format.\u003c/p\u003e\n"]]],[],null,["| The Freebase API has been shut down. Freebase was a data-sharing project that ran from 2007 to 2015. You can read more about it at [Wikipedia](https://en.wikipedia.org/wiki/Freebase_(database)), or the [announcement of its closure](https://web.archive.org/web/20170729170416/https://plus.google.com/109936836907132434202/posts/bu3z2wVqcQc?cfem=1).\n|\n| \u003cbr /\u003e\n|\n| This page provides access to the last available data dump. The historical data dumps below are\n| substantially out of date, and may contain information that is not now (or never was) accurate.\n| We continue to host these files in acknowledgement that they have been cited in\n| [numerous academic and scientific studies](https://scholar.google.com/scholar?as_sdt=0%2C5&q=freebase+rdf&btnG=),\n| and in the hope that they may contribute to ongoing efforts around open data and graph databases.\n| These datasets aren't maintained and no support can be provided regarding their\n| contents, format, or structure.\n\nData Dumps are a downloadable version of the data in Freebase. They constitute a snapshot of the data stored in Freebase and the Schema that structures it, and are provided under the same CC-BY license. The Freebase/Wikidata mappings are provided under the CC0 license.\n\n1. [Freebase Triples](#freebase-rdf-dumps)\n2. [Freebase Deleted Triples](#freebase-deleted-triples)\n3. [Freebase/Wikidata Mappings](#freebase-wikidata-mappings)\n4. [License](#license)\n5. [Citing](#citing)\n\nFreebase Triples\n\n|---------------------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------|----------|\n| This dataset contains every fact currently in Freebase. | - **Total triples:** 1.9 billion - **Updated:** Weekly - **Data Format:** [N-Triples RDF](https://www.w3.org/TR/rdf-testcases/#ntriples) - **License:** [CC-BY](https://creativecommons.org/licenses/by/2.5/) | **22 GB** gzip **250 GB** uncompressed | Download |\n\nThe RDF data is serialized using the N-Triples format, encoded as [UTF-8](https://en.wikipedia.org/wiki/Utf8) text and compressed with Gzip.\nRDF \n\n```\n\u003chttp://rdf.freebase.com/ns/g.11vjz1ynm\u003e \u003chttp://rdf.freebase.com/ns/measurement_unit.dated_percentage.date\u003e \"2001-02\"^^\u003chttp://www.w3.org/2001/XMLSchema#gYearMonth\u003e .\n\u003chttp://rdf.freebase.com/ns/g.11vjz1ynm\u003e \u003chttp://rdf.freebase.com/ns/measurement_unit.dated_percentage.source\u003e \u003chttp://rdf.freebase.com/ns/g.11x1gf2m6\u003e .\n\u003chttp://rdf.freebase.com/ns/g.11vjz1ynm\u003e \u003chttp://rdf.freebase.com/ns/type.object.type\u003e \u003chttp://rdf.freebase.com/ns/measurement_unit.dated_percentage\u003e .\n\u003chttp://rdf.freebase.com/ns/g.11vjz1ynm\u003e \u003chttp://rdf.freebase.com/ns/measurement_unit.dated_percentage.rate\u003e 4.5 .\n\u003chttp://rdf.freebase.com/ns/g.11vjz1ynm\u003e \u003chttp://www.w3.org/1999/02/22-rdf-syntax-ns#type\u003e \u003chttp://rdf.freebase.com/ns/measurement_unit.dated_percentage\u003e .\n```\n\nIf you're writing your own code to parse the RDF dumps its often more efficient to read directly from GZip file rather than extracting the data first and then processing the uncompressed data. \n\n \u003csubject\u003e \u003cpredicate\u003e \u003cobject\u003e .\n\nNote: In Freebase, objects have MIDs that look like `/m/012rkqx`. In RDF those MIDs become m.012rkqx. Likewise, Freebase schema like `/common/topic` are written as `common.topic`.\n\nThe *subject* is the ID of a Freebase object. It can be a Freebase MID (ex. m.012rkqx) for topics and CVTs or a human-readable ID (ex. `common.topic`) for schema.\n\nThe *predicate* is always a human-readable ID for a Freebase property or a property from a standard RDF vocabulary like [RDFS](https://en.wikipedia.org/wiki/RDFS). Freebase foreign key namespaces are also used as predicates to make it easier to look up keys by namespace.\n\nThe object field may contain a Freebase MID for an object or a human-readable ID for schema from Freebase or other RDF vocabularies. It may also include literal values like strings, booleans and numeric values.\n\nTopic descriptions often contain newlines. In order to make each triple fit on one line, we have escaped newlines with \"\\\\n\".\n\nFreebase Deleted Triples\n\nWe also provide a dump of triples that have been deleted from Freebase over time. This is a one-time dump through March 2013. In the future, we might consider providing periodic updates of recently deleted triples, but at the moment we have no specific timeframe for doing so, and are only providing this one-time dump.\n\nThe dump is distributed as a .tar.gz file (2.1Gb compressed, 7.7Gb uncompressed). It contains 63,036,271 deleted triples in 20 files (there is no particular meaning to the individual files, it is just easier to manipulate several smaller files than one huge file).\n\n\nThanks to Chun How Tan and John Giannandrea for making this data release possible.\n\n|-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------------------------------|----------|\n| - **Total triples:** 63 million - **Updated:** June 9, 2013 - **Data Format:** [CSV](https://en.wikipedia.org/wiki/Comma-separated_values) - **License:** [CC-BY](https://creativecommons.org/licenses/by/2.5/) | **2 GB** gzip **8 GB** uncompressed | Download |\n\nThe data format is essentially CSV with one important caveat. The object field may contain any characters, including commas (as well as any other reasonable delimiters you could think of). However, all the other fields are guaranteed not to contain commas, so the data can still be parsed unambiguously.\n\nThe columns in the dataset are defined as:\n\n- creation_timestamp (Unix epoch time in milliseconds)\n- creator\n- deletion_timestamp (Unix epoch time in milliseconds)\n- deletor\n- subject (MID)\n- predicate (MID)\n- object (MID/Literal)\n- language_code\n\nCSV \n\n```\n1352854086000,/user/mwcl_wikipedia_en,1352855856000,/user/mwcl_wikipedia_en,/m/03r90,/type/object/key,/wikipedia/en/$B816,en\n1355171076000,/user/mwcl_musicbrainz,1364258198000,/user/turtlewax_bot,/m/0nncp9z,/music/recording/artist,/m/01vbfm4,en\n1176630380000,/user/mwcl_images,1335928144000,/user/gardening_bot,/m/029w57m,/common/image/size,/m/0kly56,en\n1292854917000,/user/mwcl_musicbrainz,1364823418001,/user/mbz_pipeline_merge_bot,/m/0fv1vl8,/type/object/type,/common/topic,en\n1205530905000,/user/mwcl_images,1336022041000,/user/gardening_bot,/m/01x5scz,/common/licensed_object/license,/m/02x6b,en\n1302391361000,/user/content_administrator,1336190973000,/user/gardening_bot,/m/0gkb45y,/type/object/type,/type/content,en\n1176728962002,/user/mwcl_images,1335954186000,/user/gardening_bot,/m/08430h,/common/topic/image,/m/02cs147,en\n1172002568007,/user/mwcl_chefmoz,1283588560000,/user/delete_bot,/m/01z4c1z,/type/object/name,La Casa Rosa Mexican Restaurant,en\n```\n\nFreebase/Wikidata Mappings\n\n|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|--------------------------------------------|----------|\n| The data has been created based on the Wikidata-Dump of October 28, 2013, and contains only those links that have at least two common Wikipedia-Links and not a single disagreeing Wikipedia-Link. Furthermore, the lines are sorted by the number of common Wikipedia-Links (although in Turtle this does not really matter). | - **Total triples:** 2.1M - **Updated:** October 28, 2013 - **Data Format:** [N-Triples RDF](https://www.w3.org/TR/rdf-testcases/#ntriples) - **License:** [CC0](https://creativecommons.org/publicdomain/zero/1.0/) | **21.2 MB** gzip **242.9 MB** uncompressed | Download |\n\nThe RDF data is serialized using the N-Triples format, encoded as [UTF-8](https://en.wikipedia.org/wiki/Utf8) text and compressed with Gzip.\nRDF \n\n```\n\u003chttp://rdf.freebase.com/ns/m.0695j\u003e \u003chttp://www.w3.org/2002/07/owl#sameAs\u003e \u003chttp://www.wikidata.org/entity/Q6718\u003e .\n\u003chttp://rdf.freebase.com/ns/m.05nrg\u003e \u003chttp://www.w3.org/2002/07/owl#sameAs7\u003e \u003chttp://www.wikidata.org/entity/Q538\u003e .\n\u003chttp://rdf.freebase.com/ns/m.0jgd\u003e \u003chttp://www.w3.org/2002/07/owl#sameAs\u003e \u003chttp://www.wikidata.org/entity/Q414\u003e .\n\u003chttp://rdf.freebase.com/ns/m.0d_23\u003e \u003chttp://www.w3.org/2002/07/owl#sameAs\u003e \u003chttp://www.wikidata.org/entity/Q2537\u003e .\n\u003chttp://rdf.freebase.com/ns/m.04g7d\u003e \u003chttp://www.w3.org/2002/07/owl#sameAs\u003e \u003chttp://www.wikidata.org/entity/Q315\u003e .\n```\n\nLicense\n\nFreebase Data Dumps are provided free of charge for any purpose with regular updates by Google. They are distributed, like Freebase itself, under the [Creative Commons Attribution (aka CC-BY)](https://creativecommons.org/licenses/by/2.5/) and use is subject to the [Terms of Service](/freebase/terms).\nThe Freebase/Wikidata ID mappings are provided under [CC0](https://creativecommons.org/publicdomain/zero/1.0/) and can be used without restrictions.\n\nCiting\n\nIf you'd like to cite these data dumps in a publication, you may use: \n\n```\nGoogle, Freebase Data Dumps, https://developers.google.com/freebase/data, \u003cmonth\u003e \u003cday\u003e, \u003cyear\u003e\n```\n\nOr as BibTeX:\nBibTex \n\n```transact-sql\n@misc{freebase:datadumps,\n title = \"Freebase Data Dumps\"\n author = \"Google\",\n howpublished = \"\\url{https://developers.google.com/freebase/data}\",\n edition = \"\u003cmonth\u003e \u003cday\u003e, \u003cyear\u003e\",\n year = \"\u003cyear\u003e\"\n}\n```"]]