摘要: The usage of RDF to expose semantic data has increased dramatically over the recent years. Nowadays, datasets are so big and interconnected their management have significant scalability problems. HDT is a compact representation aiming minimize space consumption while providing retrieval features. Nonetheless, generation from traditional formats expensive in terms resources processing time. This work introduces MapReduce, framework for distributed large quantities, serialize huge into HDT, analyzes improvements both time against prior mono-node processes.