Date and Location: 7/28/2016 Thursday, 2:00pm. CS Conference Room, 206 Speaker: Md Raihan Majumder Title: Data Deduplication in a large scale server data Abstract: One of the most key concern of efficient computing is increasing storage costs and low level of backup speed. Data Deduplication essentially addresses that issue. Data Deduplication is a process of removing redundant data from a large data set. In that process, we just keep one instance of a data and delete any extra copies. When we are dealing with large amount of data, the amount of duplicate data can be significant. Removing the duplicate data can thus reduce the cost of storage at a significant scale. And back-up speed can increase essentially. In this work, we will discuss about how data deduplication works, what are the current approaches for Data Deduplication, where are we planning to implement data deduplication and future plan on implementing the Data Deduplication on Data Centre at Texas Tech University.