I come from a lot of SQL servers so it can be a bit difficult to picture exactly what happens to data when it goes into hadoop.
My understanding is that if you have a book in a text format that could be around 200k or so... you simply copy the data into hadoop and it becomes searchable. However does this data become part of a block so that HDFS can be more optimal or does it remain a 200k file in HDFS hurting performance?
Also is a Block what is often called a Tablet in Bigtable?
Thanks a lot for your help. FlyMario